{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 7.0, "eval_steps": 3000, "global_step": 1160894, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00030149178133404085, "grad_norm": 24.07154083251953, "learning_rate": 4.899999999999999e-06, "loss": 16.1284, "step": 50 }, { "epoch": 0.0006029835626680817, "grad_norm": 3.4426422119140625, "learning_rate": 9.9e-06, "loss": 11.9566, "step": 100 }, { "epoch": 0.0009044753440021225, "grad_norm": 10.80983829498291, "learning_rate": 1.4899999999999998e-05, "loss": 10.9522, "step": 150 }, { "epoch": 0.0012059671253361634, "grad_norm": 23.529926300048828, "learning_rate": 1.9899999999999996e-05, "loss": 8.5739, "step": 200 }, { "epoch": 0.0015074589066702041, "grad_norm": 94.45085144042969, "learning_rate": 2.49e-05, "loss": 8.1857, "step": 250 }, { "epoch": 0.001808950688004245, "grad_norm": 6.613641262054443, "learning_rate": 2.99e-05, "loss": 8.3156, "step": 300 }, { "epoch": 0.002110442469338286, "grad_norm": 6.473529815673828, "learning_rate": 3.4899999999999995e-05, "loss": 8.0627, "step": 350 }, { "epoch": 0.0024119342506723268, "grad_norm": 15.137494087219238, "learning_rate": 3.99e-05, "loss": 7.786, "step": 400 }, { "epoch": 0.0027134260320063673, "grad_norm": 3.99068021774292, "learning_rate": 4.49e-05, "loss": 7.3888, "step": 450 }, { "epoch": 0.0030149178133404082, "grad_norm": 3.815237283706665, "learning_rate": 4.989999999999999e-05, "loss": 7.5968, "step": 500 }, { "epoch": 0.003316409594674449, "grad_norm": 2.585122585296631, "learning_rate": 5.489999999999999e-05, "loss": 6.757, "step": 550 }, { "epoch": 0.00361790137600849, "grad_norm": 3.9703378677368164, "learning_rate": 5.989999999999999e-05, "loss": 6.7425, "step": 600 }, { "epoch": 0.003919393157342531, "grad_norm": 3.1499037742614746, "learning_rate": 6.489999999999999e-05, "loss": 7.3689, "step": 650 }, { "epoch": 0.004220884938676572, "grad_norm": 2.5951905250549316, "learning_rate": 6.989999999999999e-05, "loss": 7.3448, "step": 700 }, { "epoch": 0.004522376720010613, "grad_norm": 2.053732395172119, "learning_rate": 7.489999999999999e-05, "loss": 7.2732, "step": 750 }, { "epoch": 0.0048238685013446535, "grad_norm": 4.982893466949463, "learning_rate": 7.989999999999999e-05, "loss": 7.1032, "step": 800 }, { "epoch": 0.0051253602826786945, "grad_norm": 1.9755618572235107, "learning_rate": 8.489999999999999e-05, "loss": 7.4603, "step": 850 }, { "epoch": 0.005426852064012735, "grad_norm": 4.539783954620361, "learning_rate": 8.99e-05, "loss": 6.723, "step": 900 }, { "epoch": 0.0057283438453467755, "grad_norm": 2.8894948959350586, "learning_rate": 9.49e-05, "loss": 6.7829, "step": 950 }, { "epoch": 0.0060298356266808165, "grad_norm": 2.9384207725524902, "learning_rate": 9.99e-05, "loss": 7.1602, "step": 1000 }, { "epoch": 0.0063313274080148575, "grad_norm": 1.4089173078536987, "learning_rate": 0.0001049, "loss": 7.2547, "step": 1050 }, { "epoch": 0.006632819189348898, "grad_norm": 1.6676280498504639, "learning_rate": 0.00010989999999999999, "loss": 7.2161, "step": 1100 }, { "epoch": 0.006934310970682939, "grad_norm": 7.774955749511719, "learning_rate": 0.00011489999999999999, "loss": 6.895, "step": 1150 }, { "epoch": 0.00723580275201698, "grad_norm": 2.5557634830474854, "learning_rate": 0.00011989999999999999, "loss": 6.5208, "step": 1200 }, { "epoch": 0.007537294533351021, "grad_norm": 2.967393636703491, "learning_rate": 0.0001249, "loss": 7.4313, "step": 1250 }, { "epoch": 0.007838786314685061, "grad_norm": 3.4550092220306396, "learning_rate": 0.00012989999999999999, "loss": 7.0111, "step": 1300 }, { "epoch": 0.008140278096019103, "grad_norm": 1.4505691528320312, "learning_rate": 0.0001349, "loss": 7.5859, "step": 1350 }, { "epoch": 0.008441769877353143, "grad_norm": 1.9983046054840088, "learning_rate": 0.00013989999999999999, "loss": 7.0676, "step": 1400 }, { "epoch": 0.008743261658687183, "grad_norm": 1.7132686376571655, "learning_rate": 0.00014489999999999997, "loss": 7.2457, "step": 1450 }, { "epoch": 0.009044753440021225, "grad_norm": 1.987458348274231, "learning_rate": 0.00014989999999999998, "loss": 7.057, "step": 1500 }, { "epoch": 0.009346245221355265, "grad_norm": 1.499489426612854, "learning_rate": 0.00015489999999999997, "loss": 6.438, "step": 1550 }, { "epoch": 0.009647737002689307, "grad_norm": 3.262808322906494, "learning_rate": 0.00015989999999999998, "loss": 6.4895, "step": 1600 }, { "epoch": 0.009949228784023347, "grad_norm": 1.5130480527877808, "learning_rate": 0.00016489999999999997, "loss": 7.2896, "step": 1650 }, { "epoch": 0.010250720565357389, "grad_norm": 2.7138993740081787, "learning_rate": 0.00016989999999999998, "loss": 6.4, "step": 1700 }, { "epoch": 0.010552212346691429, "grad_norm": 1.6543924808502197, "learning_rate": 0.00017489999999999997, "loss": 6.9251, "step": 1750 }, { "epoch": 0.01085370412802547, "grad_norm": 1.3985124826431274, "learning_rate": 0.00017989999999999998, "loss": 7.0632, "step": 1800 }, { "epoch": 0.011155195909359511, "grad_norm": 1.5578947067260742, "learning_rate": 0.00018489999999999997, "loss": 7.396, "step": 1850 }, { "epoch": 0.011456687690693551, "grad_norm": 2.2867844104766846, "learning_rate": 0.00018989999999999998, "loss": 7.1139, "step": 1900 }, { "epoch": 0.011758179472027593, "grad_norm": 2.5316085815429688, "learning_rate": 0.00019489999999999997, "loss": 6.409, "step": 1950 }, { "epoch": 0.012059671253361633, "grad_norm": 1.5559688806533813, "learning_rate": 0.00019989999999999998, "loss": 7.5437, "step": 2000 }, { "epoch": 0.012361163034695675, "grad_norm": 0.6869194507598877, "learning_rate": 0.0002049, "loss": 6.6363, "step": 2050 }, { "epoch": 0.012662654816029715, "grad_norm": 1.3657041788101196, "learning_rate": 0.00020989999999999998, "loss": 7.2027, "step": 2100 }, { "epoch": 0.012964146597363757, "grad_norm": 0.7406147718429565, "learning_rate": 0.0002149, "loss": 7.3653, "step": 2150 }, { "epoch": 0.013265638378697797, "grad_norm": 1.1688302755355835, "learning_rate": 0.00021989999999999998, "loss": 7.2613, "step": 2200 }, { "epoch": 0.013567130160031837, "grad_norm": 1.2446898221969604, "learning_rate": 0.0002249, "loss": 6.875, "step": 2250 }, { "epoch": 0.013868621941365879, "grad_norm": 1.0190248489379883, "learning_rate": 0.00022989999999999998, "loss": 5.8176, "step": 2300 }, { "epoch": 0.014170113722699919, "grad_norm": 1.240607500076294, "learning_rate": 0.0002349, "loss": 6.8662, "step": 2350 }, { "epoch": 0.01447160550403396, "grad_norm": 3.916651487350464, "learning_rate": 0.00023989999999999998, "loss": 7.3535, "step": 2400 }, { "epoch": 0.014773097285368, "grad_norm": 1.3996976613998413, "learning_rate": 0.0002449, "loss": 6.9708, "step": 2450 }, { "epoch": 0.015074589066702043, "grad_norm": 2.462759494781494, "learning_rate": 0.00024989999999999995, "loss": 6.7047, "step": 2500 }, { "epoch": 0.015376080848036083, "grad_norm": 1.4236079454421997, "learning_rate": 0.00025489999999999996, "loss": 6.8466, "step": 2550 }, { "epoch": 0.015677572629370123, "grad_norm": 1.842383861541748, "learning_rate": 0.0002599, "loss": 6.8982, "step": 2600 }, { "epoch": 0.015979064410704163, "grad_norm": 3.818260431289673, "learning_rate": 0.0002649, "loss": 6.9569, "step": 2650 }, { "epoch": 0.016280556192038206, "grad_norm": 1.919683814048767, "learning_rate": 0.00026989999999999995, "loss": 6.7685, "step": 2700 }, { "epoch": 0.016582047973372246, "grad_norm": 1.633453369140625, "learning_rate": 0.00027489999999999996, "loss": 6.8701, "step": 2750 }, { "epoch": 0.016883539754706287, "grad_norm": 1.5087050199508667, "learning_rate": 0.0002799, "loss": 6.9135, "step": 2800 }, { "epoch": 0.017185031536040327, "grad_norm": 1.309898853302002, "learning_rate": 0.0002849, "loss": 7.1631, "step": 2850 }, { "epoch": 0.017486523317374367, "grad_norm": 2.1227364540100098, "learning_rate": 0.0002899, "loss": 7.5174, "step": 2900 }, { "epoch": 0.01778801509870841, "grad_norm": 1.3713490962982178, "learning_rate": 0.00029489999999999996, "loss": 7.1179, "step": 2950 }, { "epoch": 0.01808950688004245, "grad_norm": 1.7487388849258423, "learning_rate": 0.00029989999999999997, "loss": 6.5973, "step": 3000 }, { "epoch": 0.01808950688004245, "eval_loss": 7.558620452880859, "eval_runtime": 39.1223, "eval_samples_per_second": 13.087, "eval_steps_per_second": 6.544, "eval_tts_loss": 6.7164397030532, "step": 3000 }, { "epoch": 0.01839099866137649, "grad_norm": 1.3124167919158936, "learning_rate": 0.0002999999986743906, "loss": 6.9514, "step": 3050 }, { "epoch": 0.01869249044271053, "grad_norm": 0.8561250567436218, "learning_rate": 0.00029999999458879745, "loss": 6.9838, "step": 3100 }, { "epoch": 0.018993982224044574, "grad_norm": 1.2769933938980103, "learning_rate": 0.00029999998774266843, "loss": 6.4498, "step": 3150 }, { "epoch": 0.019295474005378614, "grad_norm": 6.013032913208008, "learning_rate": 0.00029999997813600374, "loss": 6.7783, "step": 3200 }, { "epoch": 0.019596965786712654, "grad_norm": 1.572847604751587, "learning_rate": 0.0002999999657688035, "loss": 7.0035, "step": 3250 }, { "epoch": 0.019898457568046694, "grad_norm": 0.6928724646568298, "learning_rate": 0.0002999999506410679, "loss": 7.0563, "step": 3300 }, { "epoch": 0.020199949349380734, "grad_norm": 2.120283365249634, "learning_rate": 0.0002999999327527973, "loss": 6.8947, "step": 3350 }, { "epoch": 0.020501441130714778, "grad_norm": 0.8892444372177124, "learning_rate": 0.000299999912103992, "loss": 6.533, "step": 3400 }, { "epoch": 0.020802932912048818, "grad_norm": 0.7623141407966614, "learning_rate": 0.00029999988869465244, "loss": 6.4411, "step": 3450 }, { "epoch": 0.021104424693382858, "grad_norm": 1.5981277227401733, "learning_rate": 0.0002999998625247789, "loss": 6.525, "step": 3500 }, { "epoch": 0.021405916474716898, "grad_norm": 1.2722316980361938, "learning_rate": 0.00029999983359437204, "loss": 6.9016, "step": 3550 }, { "epoch": 0.02170740825605094, "grad_norm": 1.624096393585205, "learning_rate": 0.00029999980190343223, "loss": 6.6612, "step": 3600 }, { "epoch": 0.022008900037384982, "grad_norm": 2.7896175384521484, "learning_rate": 0.0002999997674519602, "loss": 6.3364, "step": 3650 }, { "epoch": 0.022310391818719022, "grad_norm": 0.9449705481529236, "learning_rate": 0.0002999997302399564, "loss": 6.6145, "step": 3700 }, { "epoch": 0.022611883600053062, "grad_norm": 0.48087042570114136, "learning_rate": 0.0002999996902674217, "loss": 6.727, "step": 3750 }, { "epoch": 0.022913375381387102, "grad_norm": 1.4540601968765259, "learning_rate": 0.0002999996475343567, "loss": 6.5251, "step": 3800 }, { "epoch": 0.023214867162721146, "grad_norm": 1.0572201013565063, "learning_rate": 0.00029999960204076236, "loss": 7.0064, "step": 3850 }, { "epoch": 0.023516358944055186, "grad_norm": 1.410529375076294, "learning_rate": 0.00029999955378663933, "loss": 6.6876, "step": 3900 }, { "epoch": 0.023817850725389226, "grad_norm": 0.9826208353042603, "learning_rate": 0.00029999950277198855, "loss": 7.0781, "step": 3950 }, { "epoch": 0.024119342506723266, "grad_norm": 0.8987851142883301, "learning_rate": 0.00029999944899681104, "loss": 6.3925, "step": 4000 }, { "epoch": 0.024420834288057306, "grad_norm": 0.8632957935333252, "learning_rate": 0.00029999939246110773, "loss": 7.0482, "step": 4050 }, { "epoch": 0.02472232606939135, "grad_norm": 1.0685280561447144, "learning_rate": 0.0002999993331648796, "loss": 7.1819, "step": 4100 }, { "epoch": 0.02502381785072539, "grad_norm": 0.6388174891471863, "learning_rate": 0.0002999992711081278, "loss": 7.0396, "step": 4150 }, { "epoch": 0.02532530963205943, "grad_norm": 1.2864134311676025, "learning_rate": 0.00029999920629085355, "loss": 7.1433, "step": 4200 }, { "epoch": 0.02562680141339347, "grad_norm": 0.28794628381729126, "learning_rate": 0.000299999138713058, "loss": 5.2993, "step": 4250 }, { "epoch": 0.025928293194727513, "grad_norm": 0.8163861632347107, "learning_rate": 0.0002999990683747423, "loss": 6.7418, "step": 4300 }, { "epoch": 0.026229784976061554, "grad_norm": 1.174522042274475, "learning_rate": 0.0002999989952759078, "loss": 6.4138, "step": 4350 }, { "epoch": 0.026531276757395594, "grad_norm": 0.9004911780357361, "learning_rate": 0.00029999891941655585, "loss": 7.0695, "step": 4400 }, { "epoch": 0.026832768538729634, "grad_norm": 0.5787389874458313, "learning_rate": 0.00029999884079668786, "loss": 6.0368, "step": 4450 }, { "epoch": 0.027134260320063674, "grad_norm": 0.5247835516929626, "learning_rate": 0.0002999987594163053, "loss": 6.8442, "step": 4500 }, { "epoch": 0.027435752101397717, "grad_norm": 1.0003234148025513, "learning_rate": 0.00029999867527540953, "loss": 6.9786, "step": 4550 }, { "epoch": 0.027737243882731757, "grad_norm": 1.5438580513000488, "learning_rate": 0.0002999985883740023, "loss": 6.9171, "step": 4600 }, { "epoch": 0.028038735664065798, "grad_norm": 0.5654528141021729, "learning_rate": 0.0002999984987120851, "loss": 6.5263, "step": 4650 }, { "epoch": 0.028340227445399838, "grad_norm": 0.8113811016082764, "learning_rate": 0.0002999984062896595, "loss": 7.262, "step": 4700 }, { "epoch": 0.028641719226733878, "grad_norm": 0.8245108127593994, "learning_rate": 0.0002999983111067274, "loss": 6.7066, "step": 4750 }, { "epoch": 0.02894321100806792, "grad_norm": 1.6271929740905762, "learning_rate": 0.0002999982131632904, "loss": 6.7183, "step": 4800 }, { "epoch": 0.02924470278940196, "grad_norm": 0.6488800644874573, "learning_rate": 0.0002999981124593503, "loss": 6.2973, "step": 4850 }, { "epoch": 0.029546194570736, "grad_norm": 0.8199513554573059, "learning_rate": 0.0002999980089949091, "loss": 6.2805, "step": 4900 }, { "epoch": 0.02984768635207004, "grad_norm": 0.7314168810844421, "learning_rate": 0.0002999979027699685, "loss": 6.9033, "step": 4950 }, { "epoch": 0.030149178133404085, "grad_norm": 1.8117356300354004, "learning_rate": 0.00029999779378453065, "loss": 7.2588, "step": 5000 }, { "epoch": 0.030450669914738125, "grad_norm": 1.695137858390808, "learning_rate": 0.0002999976820385974, "loss": 6.8973, "step": 5050 }, { "epoch": 0.030752161696072165, "grad_norm": 1.979278564453125, "learning_rate": 0.0002999975675321709, "loss": 6.656, "step": 5100 }, { "epoch": 0.031053653477406205, "grad_norm": 0.9660855531692505, "learning_rate": 0.00029999745026525323, "loss": 6.6088, "step": 5150 }, { "epoch": 0.031355145258740245, "grad_norm": 2.088982343673706, "learning_rate": 0.00029999733023784654, "loss": 7.1544, "step": 5200 }, { "epoch": 0.031656637040074286, "grad_norm": 1.3137118816375732, "learning_rate": 0.00029999720744995303, "loss": 6.4744, "step": 5250 }, { "epoch": 0.031958128821408326, "grad_norm": 0.9743380546569824, "learning_rate": 0.000299997081901575, "loss": 6.358, "step": 5300 }, { "epoch": 0.03225962060274237, "grad_norm": 1.0119413137435913, "learning_rate": 0.00029999695359271474, "loss": 6.2418, "step": 5350 }, { "epoch": 0.03256111238407641, "grad_norm": 1.463389277458191, "learning_rate": 0.00029999682252337457, "loss": 6.8938, "step": 5400 }, { "epoch": 0.03286260416541045, "grad_norm": 0.7909618616104126, "learning_rate": 0.00029999668869355695, "loss": 6.8575, "step": 5450 }, { "epoch": 0.03316409594674449, "grad_norm": 0.4970797300338745, "learning_rate": 0.00029999655210326433, "loss": 6.3322, "step": 5500 }, { "epoch": 0.03346558772807853, "grad_norm": 0.6798031330108643, "learning_rate": 0.00029999641275249924, "loss": 5.8525, "step": 5550 }, { "epoch": 0.03376707950941257, "grad_norm": 1.0178723335266113, "learning_rate": 0.0002999962706412642, "loss": 6.4097, "step": 5600 }, { "epoch": 0.03406857129074661, "grad_norm": 0.8497694134712219, "learning_rate": 0.00029999612576956194, "loss": 6.734, "step": 5650 }, { "epoch": 0.03437006307208065, "grad_norm": 0.8777180314064026, "learning_rate": 0.000299995978137395, "loss": 6.495, "step": 5700 }, { "epoch": 0.03467155485341469, "grad_norm": 1.183712124824524, "learning_rate": 0.00029999582774476607, "loss": 6.5079, "step": 5750 }, { "epoch": 0.03497304663474873, "grad_norm": 1.560831904411316, "learning_rate": 0.000299995674591678, "loss": 6.3847, "step": 5800 }, { "epoch": 0.03527453841608278, "grad_norm": 1.382006049156189, "learning_rate": 0.00029999551867813365, "loss": 6.3504, "step": 5850 }, { "epoch": 0.03557603019741682, "grad_norm": 1.2293498516082764, "learning_rate": 0.00029999536000413577, "loss": 6.4648, "step": 5900 }, { "epoch": 0.03587752197875086, "grad_norm": 0.9750383496284485, "learning_rate": 0.00029999519856968744, "loss": 6.4407, "step": 5950 }, { "epoch": 0.0361790137600849, "grad_norm": 1.0200170278549194, "learning_rate": 0.00029999503437479143, "loss": 5.7454, "step": 6000 }, { "epoch": 0.0361790137600849, "eval_loss": 7.148951530456543, "eval_runtime": 39.0342, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 6.298241136232405, "step": 6000 }, { "epoch": 0.03648050554141894, "grad_norm": 0.7105050683021545, "learning_rate": 0.0002999948674194509, "loss": 6.9374, "step": 6050 }, { "epoch": 0.03678199732275298, "grad_norm": 1.632693886756897, "learning_rate": 0.0002999946977036689, "loss": 6.3988, "step": 6100 }, { "epoch": 0.03708348910408702, "grad_norm": 0.549126148223877, "learning_rate": 0.0002999945252274486, "loss": 6.6679, "step": 6150 }, { "epoch": 0.03738498088542106, "grad_norm": 1.750744342803955, "learning_rate": 0.00029999434999079306, "loss": 6.511, "step": 6200 }, { "epoch": 0.0376864726667551, "grad_norm": 1.3266537189483643, "learning_rate": 0.00029999417199370556, "loss": 6.2503, "step": 6250 }, { "epoch": 0.03798796444808915, "grad_norm": 1.6928045749664307, "learning_rate": 0.0002999939912361894, "loss": 6.7535, "step": 6300 }, { "epoch": 0.03828945622942319, "grad_norm": 0.41295602917671204, "learning_rate": 0.00029999380771824784, "loss": 6.1588, "step": 6350 }, { "epoch": 0.03859094801075723, "grad_norm": 1.16289222240448, "learning_rate": 0.00029999362143988433, "loss": 6.7316, "step": 6400 }, { "epoch": 0.03889243979209127, "grad_norm": 0.8607932925224304, "learning_rate": 0.0002999934324011023, "loss": 6.507, "step": 6450 }, { "epoch": 0.03919393157342531, "grad_norm": 1.2949159145355225, "learning_rate": 0.0002999932406019051, "loss": 7.1178, "step": 6500 }, { "epoch": 0.03949542335475935, "grad_norm": 0.6569135785102844, "learning_rate": 0.00029999304604229647, "loss": 6.8451, "step": 6550 }, { "epoch": 0.03979691513609339, "grad_norm": 1.3544552326202393, "learning_rate": 0.00029999284872227986, "loss": 6.1347, "step": 6600 }, { "epoch": 0.04009840691742743, "grad_norm": 0.5609135627746582, "learning_rate": 0.0002999926486418589, "loss": 5.9826, "step": 6650 }, { "epoch": 0.04039989869876147, "grad_norm": 1.1646627187728882, "learning_rate": 0.00029999244580103726, "loss": 6.0193, "step": 6700 }, { "epoch": 0.040701390480095516, "grad_norm": 1.1053508520126343, "learning_rate": 0.00029999224019981876, "loss": 6.4591, "step": 6750 }, { "epoch": 0.041002882261429556, "grad_norm": 1.1980043649673462, "learning_rate": 0.0002999920318382071, "loss": 6.0548, "step": 6800 }, { "epoch": 0.041304374042763596, "grad_norm": 1.3872520923614502, "learning_rate": 0.0002999918207162061, "loss": 6.4814, "step": 6850 }, { "epoch": 0.041605865824097636, "grad_norm": 1.1756304502487183, "learning_rate": 0.0002999916068338198, "loss": 6.2171, "step": 6900 }, { "epoch": 0.041907357605431676, "grad_norm": 1.4248311519622803, "learning_rate": 0.00029999139019105194, "loss": 6.8876, "step": 6950 }, { "epoch": 0.042208849386765716, "grad_norm": 2.8325917720794678, "learning_rate": 0.0002999911707879066, "loss": 6.8428, "step": 7000 }, { "epoch": 0.042510341168099756, "grad_norm": 0.2809693217277527, "learning_rate": 0.00029999094862438784, "loss": 6.3567, "step": 7050 }, { "epoch": 0.042811832949433797, "grad_norm": 2.3078622817993164, "learning_rate": 0.0002999907237004997, "loss": 6.1103, "step": 7100 }, { "epoch": 0.04311332473076784, "grad_norm": 1.4853559732437134, "learning_rate": 0.00029999049601624634, "loss": 6.454, "step": 7150 }, { "epoch": 0.04341481651210188, "grad_norm": 0.9374569058418274, "learning_rate": 0.00029999026557163194, "loss": 6.4174, "step": 7200 }, { "epoch": 0.043716308293435924, "grad_norm": 1.3244030475616455, "learning_rate": 0.0002999900323666608, "loss": 5.8845, "step": 7250 }, { "epoch": 0.044017800074769964, "grad_norm": 0.7449643611907959, "learning_rate": 0.0002999897964013371, "loss": 6.2426, "step": 7300 }, { "epoch": 0.044319291856104004, "grad_norm": 0.9419520497322083, "learning_rate": 0.00029998955767566527, "loss": 6.5926, "step": 7350 }, { "epoch": 0.044620783637438044, "grad_norm": 1.5739175081253052, "learning_rate": 0.0002999893161896497, "loss": 6.4324, "step": 7400 }, { "epoch": 0.044922275418772084, "grad_norm": 0.9271004796028137, "learning_rate": 0.00029998907194329476, "loss": 6.0301, "step": 7450 }, { "epoch": 0.045223767200106124, "grad_norm": 0.961740255355835, "learning_rate": 0.00029998882493660503, "loss": 6.0883, "step": 7500 }, { "epoch": 0.045525258981440164, "grad_norm": 1.1774910688400269, "learning_rate": 0.00029998857516958503, "loss": 6.3235, "step": 7550 }, { "epoch": 0.045826750762774204, "grad_norm": 0.40739724040031433, "learning_rate": 0.00029998832264223937, "loss": 6.5365, "step": 7600 }, { "epoch": 0.046128242544108244, "grad_norm": 0.9402585625648499, "learning_rate": 0.0002999880673545726, "loss": 6.448, "step": 7650 }, { "epoch": 0.04642973432544229, "grad_norm": 1.7140198945999146, "learning_rate": 0.00029998780930658956, "loss": 6.2629, "step": 7700 }, { "epoch": 0.04673122610677633, "grad_norm": 0.8772280216217041, "learning_rate": 0.0002999875484982949, "loss": 6.1578, "step": 7750 }, { "epoch": 0.04703271788811037, "grad_norm": 2.0432791709899902, "learning_rate": 0.0002999872849296935, "loss": 6.1681, "step": 7800 }, { "epoch": 0.04733420966944441, "grad_norm": 0.7553979754447937, "learning_rate": 0.0002999870186007901, "loss": 6.2743, "step": 7850 }, { "epoch": 0.04763570145077845, "grad_norm": 1.3929811716079712, "learning_rate": 0.0002999867495115897, "loss": 6.3658, "step": 7900 }, { "epoch": 0.04793719323211249, "grad_norm": 1.021152138710022, "learning_rate": 0.00029998647766209724, "loss": 6.0279, "step": 7950 }, { "epoch": 0.04823868501344653, "grad_norm": 2.2280824184417725, "learning_rate": 0.00029998620305231766, "loss": 6.5676, "step": 8000 }, { "epoch": 0.04854017679478057, "grad_norm": 0.571535050868988, "learning_rate": 0.00029998592568225606, "loss": 6.1899, "step": 8050 }, { "epoch": 0.04884166857611461, "grad_norm": 1.4695807695388794, "learning_rate": 0.00029998564555191754, "loss": 6.5012, "step": 8100 }, { "epoch": 0.04914316035744866, "grad_norm": 0.9992523789405823, "learning_rate": 0.00029998536266130724, "loss": 6.7557, "step": 8150 }, { "epoch": 0.0494446521387827, "grad_norm": 1.3762989044189453, "learning_rate": 0.00029998507701043043, "loss": 6.6603, "step": 8200 }, { "epoch": 0.04974614392011674, "grad_norm": 0.7077290415763855, "learning_rate": 0.00029998478859929225, "loss": 6.4314, "step": 8250 }, { "epoch": 0.05004763570145078, "grad_norm": 1.286186933517456, "learning_rate": 0.0002999844974278981, "loss": 6.6426, "step": 8300 }, { "epoch": 0.05034912748278482, "grad_norm": 1.08108651638031, "learning_rate": 0.00029998420349625337, "loss": 6.471, "step": 8350 }, { "epoch": 0.05065061926411886, "grad_norm": 1.362344741821289, "learning_rate": 0.0002999839068043634, "loss": 6.0766, "step": 8400 }, { "epoch": 0.0509521110454529, "grad_norm": 0.7339256405830383, "learning_rate": 0.00029998360735223364, "loss": 6.1943, "step": 8450 }, { "epoch": 0.05125360282678694, "grad_norm": 0.4033621549606323, "learning_rate": 0.0002999833051398696, "loss": 6.5324, "step": 8500 }, { "epoch": 0.05155509460812098, "grad_norm": 0.7146540880203247, "learning_rate": 0.00029998300016727687, "loss": 6.2183, "step": 8550 }, { "epoch": 0.05185658638945503, "grad_norm": 0.8660885095596313, "learning_rate": 0.0002999826924344611, "loss": 6.1279, "step": 8600 }, { "epoch": 0.05215807817078907, "grad_norm": 1.3407227993011475, "learning_rate": 0.0002999823819414279, "loss": 5.8754, "step": 8650 }, { "epoch": 0.05245956995212311, "grad_norm": 2.424730062484741, "learning_rate": 0.000299982068688183, "loss": 6.7546, "step": 8700 }, { "epoch": 0.05276106173345715, "grad_norm": 0.9553194046020508, "learning_rate": 0.0002999817526747321, "loss": 6.0028, "step": 8750 }, { "epoch": 0.05306255351479119, "grad_norm": 0.9534602761268616, "learning_rate": 0.0002999814339010812, "loss": 6.2666, "step": 8800 }, { "epoch": 0.05336404529612523, "grad_norm": 0.6412819027900696, "learning_rate": 0.00029998111236723596, "loss": 6.7577, "step": 8850 }, { "epoch": 0.05366553707745927, "grad_norm": 1.1979879140853882, "learning_rate": 0.0002999807880732024, "loss": 6.0309, "step": 8900 }, { "epoch": 0.05396702885879331, "grad_norm": 0.7457671761512756, "learning_rate": 0.0002999804610189864, "loss": 6.3695, "step": 8950 }, { "epoch": 0.05426852064012735, "grad_norm": 1.1704164743423462, "learning_rate": 0.00029998013120459416, "loss": 6.0096, "step": 9000 }, { "epoch": 0.05426852064012735, "eval_loss": 6.853858947753906, "eval_runtime": 38.8909, "eval_samples_per_second": 13.165, "eval_steps_per_second": 6.583, "eval_tts_loss": 6.374538193644816, "step": 9000 }, { "epoch": 0.05457001242146139, "grad_norm": 1.7051907777786255, "learning_rate": 0.00029997979863003154, "loss": 5.756, "step": 9050 }, { "epoch": 0.054871504202795435, "grad_norm": 1.1388027667999268, "learning_rate": 0.0002999794632953048, "loss": 6.5721, "step": 9100 }, { "epoch": 0.055172995984129475, "grad_norm": 0.7564111948013306, "learning_rate": 0.00029997912520042015, "loss": 6.1961, "step": 9150 }, { "epoch": 0.055474487765463515, "grad_norm": 0.7273915410041809, "learning_rate": 0.0002999787843453836, "loss": 5.7454, "step": 9200 }, { "epoch": 0.055775979546797555, "grad_norm": 0.96399986743927, "learning_rate": 0.00029997844073020164, "loss": 6.1643, "step": 9250 }, { "epoch": 0.056077471328131595, "grad_norm": 0.7099197506904602, "learning_rate": 0.00029997809435488047, "loss": 6.4389, "step": 9300 }, { "epoch": 0.056378963109465635, "grad_norm": 1.24149489402771, "learning_rate": 0.0002999777452194265, "loss": 5.9782, "step": 9350 }, { "epoch": 0.056680454890799675, "grad_norm": 0.9256930947303772, "learning_rate": 0.0002999773933238462, "loss": 5.9739, "step": 9400 }, { "epoch": 0.056981946672133715, "grad_norm": 1.6185530424118042, "learning_rate": 0.000299977038668146, "loss": 6.26, "step": 9450 }, { "epoch": 0.057283438453467755, "grad_norm": 0.9521205425262451, "learning_rate": 0.0002999766812523324, "loss": 6.0022, "step": 9500 }, { "epoch": 0.0575849302348018, "grad_norm": 0.5591525435447693, "learning_rate": 0.000299976321076412, "loss": 6.1725, "step": 9550 }, { "epoch": 0.05788642201613584, "grad_norm": 1.880816102027893, "learning_rate": 0.0002999759581403915, "loss": 6.6351, "step": 9600 }, { "epoch": 0.05818791379746988, "grad_norm": 0.873722493648529, "learning_rate": 0.00029997559244427746, "loss": 5.9925, "step": 9650 }, { "epoch": 0.05848940557880392, "grad_norm": 1.0638855695724487, "learning_rate": 0.0002999752239880767, "loss": 6.3206, "step": 9700 }, { "epoch": 0.05879089736013796, "grad_norm": 1.13161039352417, "learning_rate": 0.000299974852771796, "loss": 6.0952, "step": 9750 }, { "epoch": 0.059092389141472, "grad_norm": 0.7114644646644592, "learning_rate": 0.0002999744787954421, "loss": 5.8422, "step": 9800 }, { "epoch": 0.05939388092280604, "grad_norm": 1.0790412425994873, "learning_rate": 0.00029997410205902193, "loss": 6.1069, "step": 9850 }, { "epoch": 0.05969537270414008, "grad_norm": 0.8494043946266174, "learning_rate": 0.00029997372256254244, "loss": 6.3125, "step": 9900 }, { "epoch": 0.05999686448547412, "grad_norm": 0.39023908972740173, "learning_rate": 0.0002999733403060107, "loss": 5.6981, "step": 9950 }, { "epoch": 0.06029835626680817, "grad_norm": 1.0198321342468262, "learning_rate": 0.00029997295528943357, "loss": 6.3574, "step": 10000 }, { "epoch": 0.06059984804814221, "grad_norm": 0.5844268202781677, "learning_rate": 0.0002999725675128183, "loss": 6.4807, "step": 10050 }, { "epoch": 0.06090133982947625, "grad_norm": 0.6224924325942993, "learning_rate": 0.00029997217697617186, "loss": 6.2056, "step": 10100 }, { "epoch": 0.06120283161081029, "grad_norm": 0.8489497303962708, "learning_rate": 0.00029997178367950156, "loss": 6.1161, "step": 10150 }, { "epoch": 0.06150432339214433, "grad_norm": 1.190521240234375, "learning_rate": 0.00029997138762281455, "loss": 5.885, "step": 10200 }, { "epoch": 0.06180581517347837, "grad_norm": 0.26921728253364563, "learning_rate": 0.00029997098880611825, "loss": 6.0028, "step": 10250 }, { "epoch": 0.06210730695481241, "grad_norm": 0.9677279591560364, "learning_rate": 0.00029997058722941986, "loss": 5.7377, "step": 10300 }, { "epoch": 0.06240879873614645, "grad_norm": 1.8228585720062256, "learning_rate": 0.0002999701828927269, "loss": 6.3412, "step": 10350 }, { "epoch": 0.06271029051748049, "grad_norm": 0.8896975517272949, "learning_rate": 0.0002999697757960467, "loss": 6.1218, "step": 10400 }, { "epoch": 0.06301178229881453, "grad_norm": 0.4236070215702057, "learning_rate": 0.0002999693659393868, "loss": 5.936, "step": 10450 }, { "epoch": 0.06331327408014857, "grad_norm": 1.5041698217391968, "learning_rate": 0.0002999689533227547, "loss": 5.9462, "step": 10500 }, { "epoch": 0.06361476586148261, "grad_norm": 1.4361202716827393, "learning_rate": 0.00029996853794615806, "loss": 5.9707, "step": 10550 }, { "epoch": 0.06391625764281665, "grad_norm": 0.4873582124710083, "learning_rate": 0.0002999681198096045, "loss": 5.8282, "step": 10600 }, { "epoch": 0.06421774942415069, "grad_norm": 0.7817939519882202, "learning_rate": 0.00029996769891310176, "loss": 5.678, "step": 10650 }, { "epoch": 0.06451924120548475, "grad_norm": 0.7149667739868164, "learning_rate": 0.00029996727525665747, "loss": 6.0165, "step": 10700 }, { "epoch": 0.06482073298681879, "grad_norm": 0.8508471846580505, "learning_rate": 0.00029996684884027946, "loss": 6.0269, "step": 10750 }, { "epoch": 0.06512222476815283, "grad_norm": 0.43526604771614075, "learning_rate": 0.0002999664196639757, "loss": 6.317, "step": 10800 }, { "epoch": 0.06542371654948687, "grad_norm": 1.2091552019119263, "learning_rate": 0.000299965987727754, "loss": 6.077, "step": 10850 }, { "epoch": 0.0657252083308209, "grad_norm": 1.4578585624694824, "learning_rate": 0.00029996555303162225, "loss": 5.7256, "step": 10900 }, { "epoch": 0.06602670011215495, "grad_norm": 0.6858163475990295, "learning_rate": 0.00029996511557558856, "loss": 6.0328, "step": 10950 }, { "epoch": 0.06632819189348899, "grad_norm": 1.397073745727539, "learning_rate": 0.0002999646753596609, "loss": 6.2904, "step": 11000 }, { "epoch": 0.06662968367482303, "grad_norm": 1.113183617591858, "learning_rate": 0.00029996423238384736, "loss": 6.3548, "step": 11050 }, { "epoch": 0.06693117545615707, "grad_norm": 1.3642055988311768, "learning_rate": 0.0002999637866481562, "loss": 5.9942, "step": 11100 }, { "epoch": 0.0672326672374911, "grad_norm": 0.49496331810951233, "learning_rate": 0.00029996333815259554, "loss": 6.0619, "step": 11150 }, { "epoch": 0.06753415901882515, "grad_norm": 0.46001115441322327, "learning_rate": 0.0002999628868971736, "loss": 6.2041, "step": 11200 }, { "epoch": 0.06783565080015919, "grad_norm": 1.0289374589920044, "learning_rate": 0.0002999624328818988, "loss": 6.6582, "step": 11250 }, { "epoch": 0.06813714258149323, "grad_norm": 1.51503324508667, "learning_rate": 0.0002999619761067794, "loss": 5.9715, "step": 11300 }, { "epoch": 0.06843863436282727, "grad_norm": 1.2017478942871094, "learning_rate": 0.0002999615165718239, "loss": 6.0784, "step": 11350 }, { "epoch": 0.0687401261441613, "grad_norm": 0.7249144315719604, "learning_rate": 0.0002999610542770406, "loss": 6.4587, "step": 11400 }, { "epoch": 0.06904161792549535, "grad_norm": 0.8152539134025574, "learning_rate": 0.00029996058922243814, "loss": 6.5666, "step": 11450 }, { "epoch": 0.06934310970682939, "grad_norm": 0.7658820748329163, "learning_rate": 0.00029996012140802507, "loss": 5.9792, "step": 11500 }, { "epoch": 0.06964460148816343, "grad_norm": 1.1073342561721802, "learning_rate": 0.00029995965083380995, "loss": 6.3654, "step": 11550 }, { "epoch": 0.06994609326949747, "grad_norm": 0.732800304889679, "learning_rate": 0.0002999591774998014, "loss": 6.4374, "step": 11600 }, { "epoch": 0.07024758505083152, "grad_norm": 1.4218056201934814, "learning_rate": 0.0002999587014060083, "loss": 6.212, "step": 11650 }, { "epoch": 0.07054907683216556, "grad_norm": 0.8243114352226257, "learning_rate": 0.0002999582225524392, "loss": 6.0599, "step": 11700 }, { "epoch": 0.0708505686134996, "grad_norm": 1.1666407585144043, "learning_rate": 0.0002999577409391031, "loss": 5.9497, "step": 11750 }, { "epoch": 0.07115206039483364, "grad_norm": 0.6501251459121704, "learning_rate": 0.00029995725656600873, "loss": 6.2061, "step": 11800 }, { "epoch": 0.07145355217616768, "grad_norm": 1.2147225141525269, "learning_rate": 0.00029995676943316506, "loss": 6.4408, "step": 11850 }, { "epoch": 0.07175504395750172, "grad_norm": 0.7991725206375122, "learning_rate": 0.0002999562795405811, "loss": 6.1724, "step": 11900 }, { "epoch": 0.07205653573883576, "grad_norm": 1.1153818368911743, "learning_rate": 0.0002999557868882658, "loss": 5.9426, "step": 11950 }, { "epoch": 0.0723580275201698, "grad_norm": 0.9130564332008362, "learning_rate": 0.00029995529147622816, "loss": 6.6836, "step": 12000 }, { "epoch": 0.0723580275201698, "eval_loss": 6.653695106506348, "eval_runtime": 39.3097, "eval_samples_per_second": 13.025, "eval_steps_per_second": 6.512, "eval_tts_loss": 6.368440416910866, "step": 12000 }, { "epoch": 0.07265951930150384, "grad_norm": 0.6069143414497375, "learning_rate": 0.00029995479330447746, "loss": 5.4037, "step": 12050 }, { "epoch": 0.07296101108283788, "grad_norm": 0.6745807528495789, "learning_rate": 0.0002999542923730227, "loss": 6.0836, "step": 12100 }, { "epoch": 0.07326250286417192, "grad_norm": 0.9333676099777222, "learning_rate": 0.0002999537886818733, "loss": 5.9016, "step": 12150 }, { "epoch": 0.07356399464550596, "grad_norm": 1.0785313844680786, "learning_rate": 0.00029995328223103834, "loss": 6.235, "step": 12200 }, { "epoch": 0.07386548642684, "grad_norm": 1.0318669080734253, "learning_rate": 0.00029995277302052723, "loss": 6.0035, "step": 12250 }, { "epoch": 0.07416697820817404, "grad_norm": 0.7586814165115356, "learning_rate": 0.0002999522610503493, "loss": 6.6649, "step": 12300 }, { "epoch": 0.07446846998950808, "grad_norm": 0.9731693863868713, "learning_rate": 0.00029995174632051403, "loss": 6.2807, "step": 12350 }, { "epoch": 0.07476996177084212, "grad_norm": 1.1046092510223389, "learning_rate": 0.00029995122883103085, "loss": 5.7898, "step": 12400 }, { "epoch": 0.07507145355217616, "grad_norm": 1.4071550369262695, "learning_rate": 0.00029995070858190935, "loss": 6.1444, "step": 12450 }, { "epoch": 0.0753729453335102, "grad_norm": 0.7278856039047241, "learning_rate": 0.00029995018557315894, "loss": 6.2796, "step": 12500 }, { "epoch": 0.07567443711484426, "grad_norm": 0.7889416813850403, "learning_rate": 0.00029994965980478946, "loss": 6.1502, "step": 12550 }, { "epoch": 0.0759759288961783, "grad_norm": 0.43993058800697327, "learning_rate": 0.00029994913127681037, "loss": 5.8375, "step": 12600 }, { "epoch": 0.07627742067751234, "grad_norm": 1.2705730199813843, "learning_rate": 0.0002999485999892316, "loss": 6.2228, "step": 12650 }, { "epoch": 0.07657891245884638, "grad_norm": 0.7181907892227173, "learning_rate": 0.00029994806594206287, "loss": 5.9357, "step": 12700 }, { "epoch": 0.07688040424018042, "grad_norm": 1.0752971172332764, "learning_rate": 0.00029994752913531386, "loss": 6.4622, "step": 12750 }, { "epoch": 0.07718189602151446, "grad_norm": 1.022194743156433, "learning_rate": 0.0002999469895689946, "loss": 5.9878, "step": 12800 }, { "epoch": 0.0774833878028485, "grad_norm": 1.052477478981018, "learning_rate": 0.00029994644724311504, "loss": 6.0335, "step": 12850 }, { "epoch": 0.07778487958418254, "grad_norm": 1.4742577075958252, "learning_rate": 0.000299945902157685, "loss": 6.0874, "step": 12900 }, { "epoch": 0.07808637136551658, "grad_norm": 1.2429776191711426, "learning_rate": 0.0002999453543127147, "loss": 6.2863, "step": 12950 }, { "epoch": 0.07838786314685062, "grad_norm": 0.9748464822769165, "learning_rate": 0.00029994480370821414, "loss": 5.8786, "step": 13000 }, { "epoch": 0.07868935492818466, "grad_norm": 0.8923277258872986, "learning_rate": 0.00029994425034419347, "loss": 6.4427, "step": 13050 }, { "epoch": 0.0789908467095187, "grad_norm": 0.6761963963508606, "learning_rate": 0.0002999436942206628, "loss": 6.0833, "step": 13100 }, { "epoch": 0.07929233849085274, "grad_norm": 0.9927817583084106, "learning_rate": 0.00029994313533763243, "loss": 5.8391, "step": 13150 }, { "epoch": 0.07959383027218678, "grad_norm": 0.67519211769104, "learning_rate": 0.0002999425736951126, "loss": 5.7471, "step": 13200 }, { "epoch": 0.07989532205352082, "grad_norm": 0.7345342636108398, "learning_rate": 0.00029994200929311373, "loss": 6.0902, "step": 13250 }, { "epoch": 0.08019681383485486, "grad_norm": 0.9733516573905945, "learning_rate": 0.00029994144213164615, "loss": 6.3584, "step": 13300 }, { "epoch": 0.0804983056161889, "grad_norm": 1.3821094036102295, "learning_rate": 0.0002999408722107203, "loss": 6.3408, "step": 13350 }, { "epoch": 0.08079979739752294, "grad_norm": 0.7694543600082397, "learning_rate": 0.00029994029953034666, "loss": 5.2777, "step": 13400 }, { "epoch": 0.08110128917885698, "grad_norm": 0.5615427494049072, "learning_rate": 0.0002999397240905358, "loss": 6.0617, "step": 13450 }, { "epoch": 0.08140278096019103, "grad_norm": 1.0411791801452637, "learning_rate": 0.00029993914589129827, "loss": 6.1343, "step": 13500 }, { "epoch": 0.08170427274152507, "grad_norm": 1.0363430976867676, "learning_rate": 0.00029993856493264475, "loss": 6.5783, "step": 13550 }, { "epoch": 0.08200576452285911, "grad_norm": 0.8533387780189514, "learning_rate": 0.00029993798121458594, "loss": 6.3625, "step": 13600 }, { "epoch": 0.08230725630419315, "grad_norm": 0.468665212392807, "learning_rate": 0.00029993739473713253, "loss": 5.841, "step": 13650 }, { "epoch": 0.08260874808552719, "grad_norm": 1.1873750686645508, "learning_rate": 0.00029993680550029537, "loss": 6.082, "step": 13700 }, { "epoch": 0.08291023986686123, "grad_norm": 0.8951630592346191, "learning_rate": 0.00029993621350408523, "loss": 5.803, "step": 13750 }, { "epoch": 0.08321173164819527, "grad_norm": 0.8430988192558289, "learning_rate": 0.0002999356187485131, "loss": 6.1387, "step": 13800 }, { "epoch": 0.08351322342952931, "grad_norm": 0.48554691672325134, "learning_rate": 0.00029993502123358984, "loss": 6.5084, "step": 13850 }, { "epoch": 0.08381471521086335, "grad_norm": 0.7373800873756409, "learning_rate": 0.00029993442095932655, "loss": 6.3099, "step": 13900 }, { "epoch": 0.08411620699219739, "grad_norm": 0.9098138809204102, "learning_rate": 0.00029993381792573414, "loss": 6.1392, "step": 13950 }, { "epoch": 0.08441769877353143, "grad_norm": 1.2371550798416138, "learning_rate": 0.0002999332121328238, "loss": 5.6586, "step": 14000 }, { "epoch": 0.08471919055486547, "grad_norm": 1.1697723865509033, "learning_rate": 0.0002999326035806066, "loss": 6.0574, "step": 14050 }, { "epoch": 0.08502068233619951, "grad_norm": 0.7116578817367554, "learning_rate": 0.0002999319922690939, "loss": 5.4289, "step": 14100 }, { "epoch": 0.08532217411753355, "grad_norm": 1.0169576406478882, "learning_rate": 0.00029993137819829676, "loss": 6.5147, "step": 14150 }, { "epoch": 0.08562366589886759, "grad_norm": 1.0505529642105103, "learning_rate": 0.0002999307613682266, "loss": 5.948, "step": 14200 }, { "epoch": 0.08592515768020163, "grad_norm": 1.1250602006912231, "learning_rate": 0.00029993014177889473, "loss": 6.0217, "step": 14250 }, { "epoch": 0.08622664946153567, "grad_norm": 0.7322179675102234, "learning_rate": 0.0002999295194303126, "loss": 6.355, "step": 14300 }, { "epoch": 0.08652814124286971, "grad_norm": 0.9960523247718811, "learning_rate": 0.00029992889432249164, "loss": 6.1413, "step": 14350 }, { "epoch": 0.08682963302420375, "grad_norm": 1.0365965366363525, "learning_rate": 0.0002999282664554433, "loss": 6.1192, "step": 14400 }, { "epoch": 0.08713112480553781, "grad_norm": 0.6457193493843079, "learning_rate": 0.00029992763582917915, "loss": 6.4375, "step": 14450 }, { "epoch": 0.08743261658687185, "grad_norm": 0.8957722187042236, "learning_rate": 0.0002999270024437109, "loss": 6.0362, "step": 14500 }, { "epoch": 0.08773410836820589, "grad_norm": 0.7410829663276672, "learning_rate": 0.00029992636629905, "loss": 6.3745, "step": 14550 }, { "epoch": 0.08803560014953993, "grad_norm": 1.0071444511413574, "learning_rate": 0.0002999257273952084, "loss": 5.9626, "step": 14600 }, { "epoch": 0.08833709193087397, "grad_norm": 0.9684539437294006, "learning_rate": 0.0002999250857321977, "loss": 5.7156, "step": 14650 }, { "epoch": 0.08863858371220801, "grad_norm": 1.4066418409347534, "learning_rate": 0.0002999244413100297, "loss": 6.4049, "step": 14700 }, { "epoch": 0.08894007549354205, "grad_norm": 1.0870819091796875, "learning_rate": 0.0002999237941287164, "loss": 6.0252, "step": 14750 }, { "epoch": 0.08924156727487609, "grad_norm": 0.7366995215415955, "learning_rate": 0.00029992314418826957, "loss": 5.7008, "step": 14800 }, { "epoch": 0.08954305905621013, "grad_norm": 1.6406645774841309, "learning_rate": 0.0002999224914887013, "loss": 5.6115, "step": 14850 }, { "epoch": 0.08984455083754417, "grad_norm": 1.005753755569458, "learning_rate": 0.0002999218360300234, "loss": 5.9878, "step": 14900 }, { "epoch": 0.09014604261887821, "grad_norm": 1.8024574518203735, "learning_rate": 0.00029992117781224814, "loss": 6.4214, "step": 14950 }, { "epoch": 0.09044753440021225, "grad_norm": 1.3003273010253906, "learning_rate": 0.00029992051683538755, "loss": 5.7791, "step": 15000 }, { "epoch": 0.09044753440021225, "eval_loss": 6.540810585021973, "eval_runtime": 38.9647, "eval_samples_per_second": 13.14, "eval_steps_per_second": 6.57, "eval_tts_loss": 6.4847882920121585, "step": 15000 }, { "epoch": 0.09074902618154629, "grad_norm": 1.1977522373199463, "learning_rate": 0.00029991985309945376, "loss": 5.9245, "step": 15050 }, { "epoch": 0.09105051796288033, "grad_norm": 0.763748288154602, "learning_rate": 0.00029991918660445904, "loss": 6.146, "step": 15100 }, { "epoch": 0.09135200974421437, "grad_norm": 0.898868203163147, "learning_rate": 0.00029991851735041564, "loss": 6.3328, "step": 15150 }, { "epoch": 0.09165350152554841, "grad_norm": 1.298453688621521, "learning_rate": 0.0002999178453373359, "loss": 6.1307, "step": 15200 }, { "epoch": 0.09195499330688245, "grad_norm": 0.6493868827819824, "learning_rate": 0.0002999171705652321, "loss": 5.7518, "step": 15250 }, { "epoch": 0.09225648508821649, "grad_norm": 0.6809651851654053, "learning_rate": 0.00029991649303411674, "loss": 5.4732, "step": 15300 }, { "epoch": 0.09255797686955054, "grad_norm": 0.9689059853553772, "learning_rate": 0.0002999158127440023, "loss": 5.6883, "step": 15350 }, { "epoch": 0.09285946865088458, "grad_norm": 0.7053125500679016, "learning_rate": 0.00029991512969490124, "loss": 5.7061, "step": 15400 }, { "epoch": 0.09316096043221862, "grad_norm": 1.2019368410110474, "learning_rate": 0.0002999144438868262, "loss": 6.0384, "step": 15450 }, { "epoch": 0.09346245221355266, "grad_norm": 1.012933373451233, "learning_rate": 0.0002999137553197897, "loss": 5.6916, "step": 15500 }, { "epoch": 0.0937639439948867, "grad_norm": 0.16412775218486786, "learning_rate": 0.0002999130639938045, "loss": 6.191, "step": 15550 }, { "epoch": 0.09406543577622074, "grad_norm": 0.8868213891983032, "learning_rate": 0.0002999123699088833, "loss": 6.0374, "step": 15600 }, { "epoch": 0.09436692755755478, "grad_norm": 0.24545547366142273, "learning_rate": 0.00029991167306503885, "loss": 5.5041, "step": 15650 }, { "epoch": 0.09466841933888882, "grad_norm": 0.9122506976127625, "learning_rate": 0.000299910973462284, "loss": 5.8648, "step": 15700 }, { "epoch": 0.09496991112022286, "grad_norm": 1.0678976774215698, "learning_rate": 0.0002999102711006316, "loss": 6.4097, "step": 15750 }, { "epoch": 0.0952714029015569, "grad_norm": 0.6544720530509949, "learning_rate": 0.0002999095659800946, "loss": 6.2402, "step": 15800 }, { "epoch": 0.09557289468289094, "grad_norm": 1.782633900642395, "learning_rate": 0.00029990885810068604, "loss": 6.4527, "step": 15850 }, { "epoch": 0.09587438646422498, "grad_norm": 0.4637916386127472, "learning_rate": 0.0002999081474624188, "loss": 5.709, "step": 15900 }, { "epoch": 0.09617587824555902, "grad_norm": 1.2355046272277832, "learning_rate": 0.0002999074340653061, "loss": 5.9062, "step": 15950 }, { "epoch": 0.09647737002689306, "grad_norm": 0.6199498176574707, "learning_rate": 0.00029990671790936093, "loss": 5.5796, "step": 16000 }, { "epoch": 0.0967788618082271, "grad_norm": 0.8208577036857605, "learning_rate": 0.00029990599899459657, "loss": 6.3443, "step": 16050 }, { "epoch": 0.09708035358956114, "grad_norm": 0.8263229131698608, "learning_rate": 0.0002999052773210262, "loss": 6.0808, "step": 16100 }, { "epoch": 0.09738184537089518, "grad_norm": 1.1035619974136353, "learning_rate": 0.0002999045528886632, "loss": 6.284, "step": 16150 }, { "epoch": 0.09768333715222922, "grad_norm": 1.051700234413147, "learning_rate": 0.0002999038256975208, "loss": 6.2136, "step": 16200 }, { "epoch": 0.09798482893356326, "grad_norm": 0.8564934134483337, "learning_rate": 0.0002999030957476124, "loss": 6.3353, "step": 16250 }, { "epoch": 0.09828632071489732, "grad_norm": 1.0746164321899414, "learning_rate": 0.00029990236303895144, "loss": 5.811, "step": 16300 }, { "epoch": 0.09858781249623136, "grad_norm": 1.01360285282135, "learning_rate": 0.0002999016275715514, "loss": 5.1593, "step": 16350 }, { "epoch": 0.0988893042775654, "grad_norm": 0.8538985252380371, "learning_rate": 0.0002999008893454259, "loss": 5.0552, "step": 16400 }, { "epoch": 0.09919079605889944, "grad_norm": 1.296225905418396, "learning_rate": 0.0002999001483605884, "loss": 5.6403, "step": 16450 }, { "epoch": 0.09949228784023348, "grad_norm": 0.3570355772972107, "learning_rate": 0.0002998994046170526, "loss": 5.8805, "step": 16500 }, { "epoch": 0.09979377962156752, "grad_norm": 0.8173230886459351, "learning_rate": 0.00029989865811483216, "loss": 5.6731, "step": 16550 }, { "epoch": 0.10009527140290156, "grad_norm": 0.7958263158798218, "learning_rate": 0.0002998979088539408, "loss": 6.1758, "step": 16600 }, { "epoch": 0.1003967631842356, "grad_norm": 0.5890770554542542, "learning_rate": 0.00029989715683439235, "loss": 6.029, "step": 16650 }, { "epoch": 0.10069825496556964, "grad_norm": 0.794362485408783, "learning_rate": 0.0002998964020562007, "loss": 5.791, "step": 16700 }, { "epoch": 0.10099974674690368, "grad_norm": 0.5799646973609924, "learning_rate": 0.00029989564451937965, "loss": 5.9618, "step": 16750 }, { "epoch": 0.10130123852823772, "grad_norm": 0.7651001811027527, "learning_rate": 0.0002998948842239432, "loss": 6.1882, "step": 16800 }, { "epoch": 0.10160273030957176, "grad_norm": 0.9858241081237793, "learning_rate": 0.00029989412116990526, "loss": 5.8342, "step": 16850 }, { "epoch": 0.1019042220909058, "grad_norm": 0.6572415232658386, "learning_rate": 0.00029989335535728003, "loss": 5.861, "step": 16900 }, { "epoch": 0.10220571387223984, "grad_norm": 1.026075839996338, "learning_rate": 0.0002998925867860814, "loss": 6.3786, "step": 16950 }, { "epoch": 0.10250720565357388, "grad_norm": 0.8431625366210938, "learning_rate": 0.00029989181545632366, "loss": 6.1546, "step": 17000 }, { "epoch": 0.10280869743490792, "grad_norm": 1.169025182723999, "learning_rate": 0.000299891041368021, "loss": 6.4129, "step": 17050 }, { "epoch": 0.10311018921624196, "grad_norm": 1.1446847915649414, "learning_rate": 0.00029989026452118754, "loss": 6.1516, "step": 17100 }, { "epoch": 0.103411680997576, "grad_norm": 0.8701067566871643, "learning_rate": 0.00029988948491583775, "loss": 6.0253, "step": 17150 }, { "epoch": 0.10371317277891005, "grad_norm": 0.9107640385627747, "learning_rate": 0.0002998887025519859, "loss": 6.2971, "step": 17200 }, { "epoch": 0.1040146645602441, "grad_norm": 0.8390336632728577, "learning_rate": 0.0002998879174296463, "loss": 5.8322, "step": 17250 }, { "epoch": 0.10431615634157813, "grad_norm": 0.23110517859458923, "learning_rate": 0.00029988712954883356, "loss": 5.9841, "step": 17300 }, { "epoch": 0.10461764812291217, "grad_norm": 0.7787916660308838, "learning_rate": 0.00029988633890956206, "loss": 5.3362, "step": 17350 }, { "epoch": 0.10491913990424621, "grad_norm": 1.2890371084213257, "learning_rate": 0.00029988554551184637, "loss": 5.6252, "step": 17400 }, { "epoch": 0.10522063168558025, "grad_norm": 0.599693775177002, "learning_rate": 0.0002998847493557012, "loss": 6.2068, "step": 17450 }, { "epoch": 0.1055221234669143, "grad_norm": 0.8028955459594727, "learning_rate": 0.000299883950441141, "loss": 5.7205, "step": 17500 }, { "epoch": 0.10582361524824833, "grad_norm": 1.6344531774520874, "learning_rate": 0.00029988314876818064, "loss": 6.0143, "step": 17550 }, { "epoch": 0.10612510702958237, "grad_norm": 2.0237650871276855, "learning_rate": 0.00029988234433683483, "loss": 5.2938, "step": 17600 }, { "epoch": 0.10642659881091641, "grad_norm": 0.8227550387382507, "learning_rate": 0.0002998815371471184, "loss": 5.9316, "step": 17650 }, { "epoch": 0.10672809059225045, "grad_norm": 1.1820669174194336, "learning_rate": 0.00029988072719904613, "loss": 5.9262, "step": 17700 }, { "epoch": 0.1070295823735845, "grad_norm": 0.8800268769264221, "learning_rate": 0.000299879914492633, "loss": 6.1893, "step": 17750 }, { "epoch": 0.10733107415491853, "grad_norm": 0.7236452698707581, "learning_rate": 0.00029987909902789387, "loss": 6.416, "step": 17800 }, { "epoch": 0.10763256593625257, "grad_norm": 0.7106224298477173, "learning_rate": 0.0002998782808048439, "loss": 6.1444, "step": 17850 }, { "epoch": 0.10793405771758662, "grad_norm": 1.2282185554504395, "learning_rate": 0.00029987745982349795, "loss": 5.7979, "step": 17900 }, { "epoch": 0.10823554949892066, "grad_norm": 1.1029248237609863, "learning_rate": 0.0002998766360838713, "loss": 5.4135, "step": 17950 }, { "epoch": 0.1085370412802547, "grad_norm": 0.6998524069786072, "learning_rate": 0.00029987580958597905, "loss": 6.1549, "step": 18000 }, { "epoch": 0.1085370412802547, "eval_loss": 6.450806617736816, "eval_runtime": 39.1318, "eval_samples_per_second": 13.084, "eval_steps_per_second": 6.542, "eval_tts_loss": 6.497903316684779, "step": 18000 }, { "epoch": 0.10883853306158874, "grad_norm": 1.4595850706100464, "learning_rate": 0.0002998749803298364, "loss": 5.96, "step": 18050 }, { "epoch": 0.10914002484292278, "grad_norm": 0.9234338402748108, "learning_rate": 0.00029987414831545865, "loss": 5.3808, "step": 18100 }, { "epoch": 0.10944151662425683, "grad_norm": 0.9661875367164612, "learning_rate": 0.00029987331354286103, "loss": 5.5385, "step": 18150 }, { "epoch": 0.10974300840559087, "grad_norm": 1.079479694366455, "learning_rate": 0.000299872476012059, "loss": 5.7928, "step": 18200 }, { "epoch": 0.11004450018692491, "grad_norm": 0.6925485730171204, "learning_rate": 0.00029987163572306794, "loss": 5.7686, "step": 18250 }, { "epoch": 0.11034599196825895, "grad_norm": 0.7660449147224426, "learning_rate": 0.00029987079267590326, "loss": 6.0931, "step": 18300 }, { "epoch": 0.11064748374959299, "grad_norm": 0.7390903830528259, "learning_rate": 0.0002998699468705805, "loss": 5.7706, "step": 18350 }, { "epoch": 0.11094897553092703, "grad_norm": 0.9840453863143921, "learning_rate": 0.0002998690983071153, "loss": 6.0136, "step": 18400 }, { "epoch": 0.11125046731226107, "grad_norm": 0.6170691847801208, "learning_rate": 0.0002998682469855232, "loss": 5.9653, "step": 18450 }, { "epoch": 0.11155195909359511, "grad_norm": 1.106951117515564, "learning_rate": 0.00029986739290581985, "loss": 5.9854, "step": 18500 }, { "epoch": 0.11185345087492915, "grad_norm": 0.8451488614082336, "learning_rate": 0.00029986653606802106, "loss": 5.605, "step": 18550 }, { "epoch": 0.11215494265626319, "grad_norm": 1.134221076965332, "learning_rate": 0.00029986567647214253, "loss": 6.2793, "step": 18600 }, { "epoch": 0.11245643443759723, "grad_norm": 0.6529774069786072, "learning_rate": 0.00029986481411820005, "loss": 5.6264, "step": 18650 }, { "epoch": 0.11275792621893127, "grad_norm": 1.107420563697815, "learning_rate": 0.0002998639490062096, "loss": 5.6989, "step": 18700 }, { "epoch": 0.11305941800026531, "grad_norm": 0.6795862317085266, "learning_rate": 0.000299863081136187, "loss": 5.7349, "step": 18750 }, { "epoch": 0.11336090978159935, "grad_norm": 0.6698465347290039, "learning_rate": 0.00029986221050814826, "loss": 6.2136, "step": 18800 }, { "epoch": 0.11366240156293339, "grad_norm": 0.7908414006233215, "learning_rate": 0.0002998613371221094, "loss": 5.9588, "step": 18850 }, { "epoch": 0.11396389334426743, "grad_norm": 0.8591198325157166, "learning_rate": 0.00029986046097808653, "loss": 5.7297, "step": 18900 }, { "epoch": 0.11426538512560147, "grad_norm": 1.2069357633590698, "learning_rate": 0.0002998595820760957, "loss": 5.2936, "step": 18950 }, { "epoch": 0.11456687690693551, "grad_norm": 0.9974901676177979, "learning_rate": 0.0002998587004161531, "loss": 5.2076, "step": 19000 }, { "epoch": 0.11486836868826955, "grad_norm": 0.6980935335159302, "learning_rate": 0.0002998578159982751, "loss": 5.4657, "step": 19050 }, { "epoch": 0.1151698604696036, "grad_norm": 1.1752214431762695, "learning_rate": 0.0002998569288224777, "loss": 5.9529, "step": 19100 }, { "epoch": 0.11547135225093764, "grad_norm": 0.7149506211280823, "learning_rate": 0.00029985603888877744, "loss": 4.9746, "step": 19150 }, { "epoch": 0.11577284403227169, "grad_norm": 0.6598160862922668, "learning_rate": 0.00029985514619719066, "loss": 5.6392, "step": 19200 }, { "epoch": 0.11607433581360573, "grad_norm": 0.9422267079353333, "learning_rate": 0.0002998542507477338, "loss": 6.1296, "step": 19250 }, { "epoch": 0.11637582759493977, "grad_norm": 0.652341902256012, "learning_rate": 0.0002998533525404232, "loss": 5.9108, "step": 19300 }, { "epoch": 0.1166773193762738, "grad_norm": 0.8928499221801758, "learning_rate": 0.0002998524515752756, "loss": 5.8089, "step": 19350 }, { "epoch": 0.11697881115760785, "grad_norm": 0.684892475605011, "learning_rate": 0.0002998515478523074, "loss": 5.8931, "step": 19400 }, { "epoch": 0.11728030293894189, "grad_norm": 0.7053046226501465, "learning_rate": 0.0002998506413715354, "loss": 5.6323, "step": 19450 }, { "epoch": 0.11758179472027593, "grad_norm": 1.3431217670440674, "learning_rate": 0.00029984973213297613, "loss": 6.0854, "step": 19500 }, { "epoch": 0.11788328650160997, "grad_norm": 0.8500294089317322, "learning_rate": 0.0002998488201366464, "loss": 5.8533, "step": 19550 }, { "epoch": 0.118184778282944, "grad_norm": 1.1020880937576294, "learning_rate": 0.000299847905382563, "loss": 5.4396, "step": 19600 }, { "epoch": 0.11848627006427805, "grad_norm": 0.9641672372817993, "learning_rate": 0.0002998469878707427, "loss": 5.7444, "step": 19650 }, { "epoch": 0.11878776184561209, "grad_norm": 1.4753443002700806, "learning_rate": 0.0002998460676012025, "loss": 5.917, "step": 19700 }, { "epoch": 0.11908925362694613, "grad_norm": 1.049117922782898, "learning_rate": 0.0002998451445739592, "loss": 6.1829, "step": 19750 }, { "epoch": 0.11939074540828017, "grad_norm": 0.6328831911087036, "learning_rate": 0.00029984421878902993, "loss": 5.7303, "step": 19800 }, { "epoch": 0.1196922371896142, "grad_norm": 1.6566139459609985, "learning_rate": 0.0002998432902464316, "loss": 5.8457, "step": 19850 }, { "epoch": 0.11999372897094825, "grad_norm": 0.7753614783287048, "learning_rate": 0.00029984235894618135, "loss": 6.2431, "step": 19900 }, { "epoch": 0.12029522075228229, "grad_norm": 1.409170389175415, "learning_rate": 0.00029984142488829633, "loss": 5.6822, "step": 19950 }, { "epoch": 0.12059671253361634, "grad_norm": 1.5813260078430176, "learning_rate": 0.0002998404880727937, "loss": 5.7415, "step": 20000 }, { "epoch": 0.12089820431495038, "grad_norm": 0.5167080163955688, "learning_rate": 0.0002998395484996908, "loss": 5.852, "step": 20050 }, { "epoch": 0.12119969609628442, "grad_norm": 0.8582403063774109, "learning_rate": 0.0002998386061690047, "loss": 5.5904, "step": 20100 }, { "epoch": 0.12150118787761846, "grad_norm": 0.7270346879959106, "learning_rate": 0.00029983766108075303, "loss": 6.0091, "step": 20150 }, { "epoch": 0.1218026796589525, "grad_norm": 1.3469833135604858, "learning_rate": 0.00029983671323495297, "loss": 5.5286, "step": 20200 }, { "epoch": 0.12210417144028654, "grad_norm": 1.151395559310913, "learning_rate": 0.000299835762631622, "loss": 5.4648, "step": 20250 }, { "epoch": 0.12240566322162058, "grad_norm": 0.7673705220222473, "learning_rate": 0.0002998348092707777, "loss": 6.0689, "step": 20300 }, { "epoch": 0.12270715500295462, "grad_norm": 0.8488967418670654, "learning_rate": 0.0002998338531524376, "loss": 6.1047, "step": 20350 }, { "epoch": 0.12300864678428866, "grad_norm": 0.8969296216964722, "learning_rate": 0.00029983289427661916, "loss": 5.7827, "step": 20400 }, { "epoch": 0.1233101385656227, "grad_norm": 0.5451242327690125, "learning_rate": 0.00029983193264334013, "loss": 6.0431, "step": 20450 }, { "epoch": 0.12361163034695674, "grad_norm": 0.7929562330245972, "learning_rate": 0.0002998309682526183, "loss": 6.2164, "step": 20500 }, { "epoch": 0.12391312212829078, "grad_norm": 1.0044353008270264, "learning_rate": 0.00029983000110447123, "loss": 6.2159, "step": 20550 }, { "epoch": 0.12421461390962482, "grad_norm": 0.7870169878005981, "learning_rate": 0.0002998290311989168, "loss": 5.9454, "step": 20600 }, { "epoch": 0.12451610569095886, "grad_norm": 0.7006690502166748, "learning_rate": 0.0002998280585359729, "loss": 6.0108, "step": 20650 }, { "epoch": 0.1248175974722929, "grad_norm": 1.1039345264434814, "learning_rate": 0.00029982708311565744, "loss": 6.1166, "step": 20700 }, { "epoch": 0.12511908925362694, "grad_norm": 1.1750527620315552, "learning_rate": 0.00029982610493798826, "loss": 6.1375, "step": 20750 }, { "epoch": 0.12542058103496098, "grad_norm": 0.745803952217102, "learning_rate": 0.0002998251240029835, "loss": 5.752, "step": 20800 }, { "epoch": 0.12572207281629502, "grad_norm": 0.8570643067359924, "learning_rate": 0.0002998241403106611, "loss": 6.127, "step": 20850 }, { "epoch": 0.12602356459762906, "grad_norm": 2.124727964401245, "learning_rate": 0.0002998231538610392, "loss": 6.2222, "step": 20900 }, { "epoch": 0.1263250563789631, "grad_norm": 0.30511021614074707, "learning_rate": 0.000299822164654136, "loss": 6.0493, "step": 20950 }, { "epoch": 0.12662654816029714, "grad_norm": 0.4885910749435425, "learning_rate": 0.0002998211726899696, "loss": 6.0309, "step": 21000 }, { "epoch": 0.12662654816029714, "eval_loss": 6.38372802734375, "eval_runtime": 38.8646, "eval_samples_per_second": 13.174, "eval_steps_per_second": 6.587, "eval_tts_loss": 6.590201215275383, "step": 21000 }, { "epoch": 0.12692803994163118, "grad_norm": 1.0846093893051147, "learning_rate": 0.0002998201779685584, "loss": 6.2362, "step": 21050 }, { "epoch": 0.12722953172296522, "grad_norm": 0.9149523377418518, "learning_rate": 0.0002998191804899206, "loss": 5.5917, "step": 21100 }, { "epoch": 0.12753102350429926, "grad_norm": 1.4513638019561768, "learning_rate": 0.00029981818025407456, "loss": 5.4137, "step": 21150 }, { "epoch": 0.1278325152856333, "grad_norm": 0.8786852955818176, "learning_rate": 0.00029981717726103877, "loss": 5.6276, "step": 21200 }, { "epoch": 0.12813400706696734, "grad_norm": 0.8669352531433105, "learning_rate": 0.0002998161715108316, "loss": 5.7677, "step": 21250 }, { "epoch": 0.12843549884830138, "grad_norm": 0.7132697105407715, "learning_rate": 0.0002998151630034716, "loss": 5.4312, "step": 21300 }, { "epoch": 0.12873699062963542, "grad_norm": 1.2484006881713867, "learning_rate": 0.0002998141517389773, "loss": 5.5117, "step": 21350 }, { "epoch": 0.1290384824109695, "grad_norm": 1.1156971454620361, "learning_rate": 0.00029981313771736736, "loss": 6.2253, "step": 21400 }, { "epoch": 0.12933997419230353, "grad_norm": 1.1848152875900269, "learning_rate": 0.00029981212093866045, "loss": 5.5256, "step": 21450 }, { "epoch": 0.12964146597363757, "grad_norm": 1.1360630989074707, "learning_rate": 0.00029981110140287517, "loss": 5.4092, "step": 21500 }, { "epoch": 0.1299429577549716, "grad_norm": 1.1173902750015259, "learning_rate": 0.0002998100791100304, "loss": 6.1738, "step": 21550 }, { "epoch": 0.13024444953630565, "grad_norm": 1.1543892621994019, "learning_rate": 0.00029980905406014487, "loss": 6.1824, "step": 21600 }, { "epoch": 0.1305459413176397, "grad_norm": 0.6018022298812866, "learning_rate": 0.0002998080262532376, "loss": 6.0884, "step": 21650 }, { "epoch": 0.13084743309897373, "grad_norm": 1.0597225427627563, "learning_rate": 0.0002998069956893273, "loss": 5.7031, "step": 21700 }, { "epoch": 0.13114892488030777, "grad_norm": 1.212719440460205, "learning_rate": 0.00029980596236843304, "loss": 5.8175, "step": 21750 }, { "epoch": 0.1314504166616418, "grad_norm": 0.8754852414131165, "learning_rate": 0.0002998049262905738, "loss": 5.9523, "step": 21800 }, { "epoch": 0.13175190844297585, "grad_norm": 0.6216838359832764, "learning_rate": 0.0002998038874557688, "loss": 5.7006, "step": 21850 }, { "epoch": 0.1320534002243099, "grad_norm": 0.8056514859199524, "learning_rate": 0.0002998028458640369, "loss": 5.6733, "step": 21900 }, { "epoch": 0.13235489200564393, "grad_norm": 0.9873939752578735, "learning_rate": 0.00029980180151539744, "loss": 6.1347, "step": 21950 }, { "epoch": 0.13265638378697797, "grad_norm": 1.046749472618103, "learning_rate": 0.0002998007544098696, "loss": 5.934, "step": 22000 }, { "epoch": 0.132957875568312, "grad_norm": 0.26720646023750305, "learning_rate": 0.00029979970454747267, "loss": 5.5718, "step": 22050 }, { "epoch": 0.13325936734964605, "grad_norm": 1.280653953552246, "learning_rate": 0.0002997986519282259, "loss": 5.5545, "step": 22100 }, { "epoch": 0.1335608591309801, "grad_norm": 0.9807652235031128, "learning_rate": 0.0002997975965521488, "loss": 6.0188, "step": 22150 }, { "epoch": 0.13386235091231413, "grad_norm": 0.6851532459259033, "learning_rate": 0.0002997965384192607, "loss": 6.1163, "step": 22200 }, { "epoch": 0.13416384269364817, "grad_norm": 0.8494829535484314, "learning_rate": 0.00029979547752958103, "loss": 5.7473, "step": 22250 }, { "epoch": 0.1344653344749822, "grad_norm": 0.6508592367172241, "learning_rate": 0.0002997944138831294, "loss": 6.6296, "step": 22300 }, { "epoch": 0.13476682625631625, "grad_norm": 0.952364444732666, "learning_rate": 0.00029979334747992533, "loss": 5.9322, "step": 22350 }, { "epoch": 0.1350683180376503, "grad_norm": 0.8203451037406921, "learning_rate": 0.0002997922783199885, "loss": 5.9377, "step": 22400 }, { "epoch": 0.13536980981898433, "grad_norm": 1.0202277898788452, "learning_rate": 0.0002997912064033385, "loss": 5.5101, "step": 22450 }, { "epoch": 0.13567130160031837, "grad_norm": 1.078608751296997, "learning_rate": 0.0002997901317299951, "loss": 6.0079, "step": 22500 }, { "epoch": 0.1359727933816524, "grad_norm": 1.170854926109314, "learning_rate": 0.00029978905429997815, "loss": 5.4454, "step": 22550 }, { "epoch": 0.13627428516298645, "grad_norm": 0.7983636856079102, "learning_rate": 0.0002997879741133074, "loss": 6.1405, "step": 22600 }, { "epoch": 0.1365757769443205, "grad_norm": 1.070150375366211, "learning_rate": 0.0002997868911700027, "loss": 5.9011, "step": 22650 }, { "epoch": 0.13687726872565453, "grad_norm": 0.9191415309906006, "learning_rate": 0.0002997858054700841, "loss": 6.0426, "step": 22700 }, { "epoch": 0.13717876050698857, "grad_norm": 0.20462839305400848, "learning_rate": 0.0002997847170135715, "loss": 6.1219, "step": 22750 }, { "epoch": 0.1374802522883226, "grad_norm": 1.2746565341949463, "learning_rate": 0.00029978362580048486, "loss": 6.0365, "step": 22800 }, { "epoch": 0.13778174406965665, "grad_norm": 1.393287181854248, "learning_rate": 0.00029978253183084435, "loss": 5.848, "step": 22850 }, { "epoch": 0.1380832358509907, "grad_norm": 0.8269675374031067, "learning_rate": 0.00029978143510467017, "loss": 5.8229, "step": 22900 }, { "epoch": 0.13838472763232473, "grad_norm": 1.3555891513824463, "learning_rate": 0.0002997803356219824, "loss": 6.2754, "step": 22950 }, { "epoch": 0.13868621941365877, "grad_norm": 0.8588695526123047, "learning_rate": 0.0002997792333828013, "loss": 5.8246, "step": 23000 }, { "epoch": 0.1389877111949928, "grad_norm": 0.5260128378868103, "learning_rate": 0.0002997781283871471, "loss": 5.9337, "step": 23050 }, { "epoch": 0.13928920297632685, "grad_norm": 0.5688589215278625, "learning_rate": 0.00029977702063504027, "loss": 5.2028, "step": 23100 }, { "epoch": 0.1395906947576609, "grad_norm": 0.7482021450996399, "learning_rate": 0.0002997759101265011, "loss": 5.9121, "step": 23150 }, { "epoch": 0.13989218653899493, "grad_norm": 0.7916771173477173, "learning_rate": 0.00029977479686155, "loss": 6.3605, "step": 23200 }, { "epoch": 0.140193678320329, "grad_norm": 0.6448206305503845, "learning_rate": 0.0002997736808402075, "loss": 6.1306, "step": 23250 }, { "epoch": 0.14049517010166304, "grad_norm": 0.5073733925819397, "learning_rate": 0.00029977256206249417, "loss": 5.7472, "step": 23300 }, { "epoch": 0.14079666188299708, "grad_norm": 0.8803818821907043, "learning_rate": 0.0002997714405284306, "loss": 5.6239, "step": 23350 }, { "epoch": 0.14109815366433112, "grad_norm": 1.062461018562317, "learning_rate": 0.00029977031623803737, "loss": 5.574, "step": 23400 }, { "epoch": 0.14139964544566516, "grad_norm": 0.7356359958648682, "learning_rate": 0.0002997691891913352, "loss": 5.6249, "step": 23450 }, { "epoch": 0.1417011372269992, "grad_norm": 1.1652953624725342, "learning_rate": 0.00029976805938834483, "loss": 6.2627, "step": 23500 }, { "epoch": 0.14200262900833324, "grad_norm": 1.143570899963379, "learning_rate": 0.0002997669268290871, "loss": 5.8699, "step": 23550 }, { "epoch": 0.14230412078966728, "grad_norm": 1.0183955430984497, "learning_rate": 0.00029976579151358277, "loss": 5.8043, "step": 23600 }, { "epoch": 0.14260561257100132, "grad_norm": 0.6530879735946655, "learning_rate": 0.00029976465344185276, "loss": 5.9405, "step": 23650 }, { "epoch": 0.14290710435233536, "grad_norm": 0.7351316809654236, "learning_rate": 0.0002997635126139181, "loss": 6.1868, "step": 23700 }, { "epoch": 0.1432085961336694, "grad_norm": 1.30414617061615, "learning_rate": 0.0002997623690297997, "loss": 5.0963, "step": 23750 }, { "epoch": 0.14351008791500344, "grad_norm": 0.6204509139060974, "learning_rate": 0.0002997612226895186, "loss": 6.0259, "step": 23800 }, { "epoch": 0.14381157969633748, "grad_norm": 0.23674315214157104, "learning_rate": 0.0002997600735930959, "loss": 5.8131, "step": 23850 }, { "epoch": 0.14411307147767152, "grad_norm": 1.1874409914016724, "learning_rate": 0.0002997589217405527, "loss": 5.5075, "step": 23900 }, { "epoch": 0.14441456325900556, "grad_norm": 0.43627938628196716, "learning_rate": 0.0002997577671319104, "loss": 5.7873, "step": 23950 }, { "epoch": 0.1447160550403396, "grad_norm": 0.9005704522132874, "learning_rate": 0.00029975660976719, "loss": 6.1418, "step": 24000 }, { "epoch": 0.1447160550403396, "eval_loss": 6.325107574462891, "eval_runtime": 38.9741, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.568, "eval_tts_loss": 6.53273311156827, "step": 24000 }, { "epoch": 0.14501754682167364, "grad_norm": 0.8167188167572021, "learning_rate": 0.000299755449646413, "loss": 5.7993, "step": 24050 }, { "epoch": 0.14531903860300768, "grad_norm": 0.7282929420471191, "learning_rate": 0.00029975428676960064, "loss": 5.9784, "step": 24100 }, { "epoch": 0.14562053038434172, "grad_norm": 0.2753675878047943, "learning_rate": 0.0002997531211367743, "loss": 5.3149, "step": 24150 }, { "epoch": 0.14592202216567576, "grad_norm": 1.079245686531067, "learning_rate": 0.00029975195274795546, "loss": 5.595, "step": 24200 }, { "epoch": 0.1462235139470098, "grad_norm": 0.9694415926933289, "learning_rate": 0.0002997507816031657, "loss": 5.8323, "step": 24250 }, { "epoch": 0.14652500572834384, "grad_norm": 1.0887646675109863, "learning_rate": 0.0002997496077024265, "loss": 5.3635, "step": 24300 }, { "epoch": 0.14682649750967788, "grad_norm": 1.009871006011963, "learning_rate": 0.0002997484310457595, "loss": 5.9426, "step": 24350 }, { "epoch": 0.14712798929101192, "grad_norm": 0.8857830166816711, "learning_rate": 0.0002997472516331863, "loss": 5.493, "step": 24400 }, { "epoch": 0.14742948107234596, "grad_norm": 1.3425889015197754, "learning_rate": 0.00029974606946472855, "loss": 5.6304, "step": 24450 }, { "epoch": 0.14773097285368, "grad_norm": 0.6240967512130737, "learning_rate": 0.00029974488454040816, "loss": 5.8833, "step": 24500 }, { "epoch": 0.14803246463501404, "grad_norm": 1.3265914916992188, "learning_rate": 0.0002997436968602469, "loss": 5.6735, "step": 24550 }, { "epoch": 0.14833395641634808, "grad_norm": 1.3449559211730957, "learning_rate": 0.0002997425064242666, "loss": 6.1009, "step": 24600 }, { "epoch": 0.14863544819768212, "grad_norm": 1.0099847316741943, "learning_rate": 0.0002997413132324891, "loss": 5.5654, "step": 24650 }, { "epoch": 0.14893693997901616, "grad_norm": 1.454803705215454, "learning_rate": 0.00029974011728493643, "loss": 6.1597, "step": 24700 }, { "epoch": 0.1492384317603502, "grad_norm": 0.8489559888839722, "learning_rate": 0.0002997389185816306, "loss": 5.5617, "step": 24750 }, { "epoch": 0.14953992354168424, "grad_norm": 0.6369065642356873, "learning_rate": 0.00029973771712259374, "loss": 5.7873, "step": 24800 }, { "epoch": 0.14984141532301828, "grad_norm": 0.8052045702934265, "learning_rate": 0.0002997365129078478, "loss": 6.1726, "step": 24850 }, { "epoch": 0.15014290710435232, "grad_norm": 0.9027529954910278, "learning_rate": 0.000299735305937415, "loss": 5.6321, "step": 24900 }, { "epoch": 0.15044439888568636, "grad_norm": 1.1388731002807617, "learning_rate": 0.0002997340962113177, "loss": 5.9957, "step": 24950 }, { "epoch": 0.1507458906670204, "grad_norm": 1.3063418865203857, "learning_rate": 0.00029973288372957794, "loss": 5.7603, "step": 25000 }, { "epoch": 0.15104738244835444, "grad_norm": 0.7326682806015015, "learning_rate": 0.00029973166849221814, "loss": 5.5043, "step": 25050 }, { "epoch": 0.1513488742296885, "grad_norm": 0.7688078284263611, "learning_rate": 0.0002997304504992607, "loss": 5.9283, "step": 25100 }, { "epoch": 0.15165036601102255, "grad_norm": 0.3574155867099762, "learning_rate": 0.000299729229750728, "loss": 5.2271, "step": 25150 }, { "epoch": 0.1519518577923566, "grad_norm": 1.89810311794281, "learning_rate": 0.00029972800624664245, "loss": 5.7305, "step": 25200 }, { "epoch": 0.15225334957369063, "grad_norm": 0.887994110584259, "learning_rate": 0.0002997267799870266, "loss": 5.7431, "step": 25250 }, { "epoch": 0.15255484135502467, "grad_norm": 0.7674331068992615, "learning_rate": 0.0002997255509719031, "loss": 5.2437, "step": 25300 }, { "epoch": 0.1528563331363587, "grad_norm": 0.6838112473487854, "learning_rate": 0.00029972431920129454, "loss": 5.599, "step": 25350 }, { "epoch": 0.15315782491769275, "grad_norm": 0.715670108795166, "learning_rate": 0.00029972308467522343, "loss": 5.7585, "step": 25400 }, { "epoch": 0.1534593166990268, "grad_norm": 1.018279790878296, "learning_rate": 0.00029972184739371273, "loss": 5.5317, "step": 25450 }, { "epoch": 0.15376080848036083, "grad_norm": 1.1330264806747437, "learning_rate": 0.00029972060735678505, "loss": 5.9247, "step": 25500 }, { "epoch": 0.15406230026169487, "grad_norm": 0.8631731867790222, "learning_rate": 0.00029971936456446325, "loss": 5.9161, "step": 25550 }, { "epoch": 0.1543637920430289, "grad_norm": 0.6622151136398315, "learning_rate": 0.0002997181190167702, "loss": 6.0616, "step": 25600 }, { "epoch": 0.15466528382436295, "grad_norm": 1.1876293420791626, "learning_rate": 0.0002997168707137288, "loss": 5.7053, "step": 25650 }, { "epoch": 0.154966775605697, "grad_norm": 1.185960054397583, "learning_rate": 0.0002997156196553621, "loss": 6.0796, "step": 25700 }, { "epoch": 0.15526826738703103, "grad_norm": 1.0712347030639648, "learning_rate": 0.0002997143658416931, "loss": 5.6357, "step": 25750 }, { "epoch": 0.15556975916836507, "grad_norm": 0.6954686045646667, "learning_rate": 0.00029971310927274483, "loss": 5.8232, "step": 25800 }, { "epoch": 0.1558712509496991, "grad_norm": 0.7894907593727112, "learning_rate": 0.0002997118499485404, "loss": 6.1667, "step": 25850 }, { "epoch": 0.15617274273103315, "grad_norm": 1.0289521217346191, "learning_rate": 0.000299710587869103, "loss": 5.7948, "step": 25900 }, { "epoch": 0.1564742345123672, "grad_norm": 1.1280009746551514, "learning_rate": 0.000299709323034456, "loss": 5.9437, "step": 25950 }, { "epoch": 0.15677572629370123, "grad_norm": 0.7206826210021973, "learning_rate": 0.0002997080554446224, "loss": 5.3621, "step": 26000 }, { "epoch": 0.15707721807503527, "grad_norm": 0.7407723665237427, "learning_rate": 0.0002997067850996258, "loss": 5.7378, "step": 26050 }, { "epoch": 0.15737870985636931, "grad_norm": 0.6900263428688049, "learning_rate": 0.00029970551199948945, "loss": 5.8784, "step": 26100 }, { "epoch": 0.15768020163770335, "grad_norm": 0.9872252941131592, "learning_rate": 0.00029970423614423674, "loss": 5.9178, "step": 26150 }, { "epoch": 0.1579816934190374, "grad_norm": 0.9071255922317505, "learning_rate": 0.0002997029575338912, "loss": 5.6466, "step": 26200 }, { "epoch": 0.15828318520037143, "grad_norm": 1.178222894668579, "learning_rate": 0.00029970167616847644, "loss": 5.2543, "step": 26250 }, { "epoch": 0.15858467698170547, "grad_norm": 0.5755072236061096, "learning_rate": 0.0002997003920480159, "loss": 5.6067, "step": 26300 }, { "epoch": 0.15888616876303951, "grad_norm": 1.486236333847046, "learning_rate": 0.0002996991051725333, "loss": 5.7845, "step": 26350 }, { "epoch": 0.15918766054437355, "grad_norm": 0.6378349661827087, "learning_rate": 0.0002996978155420523, "loss": 6.0712, "step": 26400 }, { "epoch": 0.1594891523257076, "grad_norm": 1.6551440954208374, "learning_rate": 0.0002996965231565966, "loss": 5.3541, "step": 26450 }, { "epoch": 0.15979064410704164, "grad_norm": 0.8616222143173218, "learning_rate": 0.0002996952280161901, "loss": 5.6155, "step": 26500 }, { "epoch": 0.16009213588837568, "grad_norm": 0.6234305500984192, "learning_rate": 0.0002996939301208565, "loss": 5.6455, "step": 26550 }, { "epoch": 0.16039362766970972, "grad_norm": 1.2547117471694946, "learning_rate": 0.00029969262947061973, "loss": 5.6683, "step": 26600 }, { "epoch": 0.16069511945104376, "grad_norm": 0.348393052816391, "learning_rate": 0.0002996913260655038, "loss": 5.1752, "step": 26650 }, { "epoch": 0.1609966112323778, "grad_norm": 0.7471150755882263, "learning_rate": 0.00029969001990553254, "loss": 5.4982, "step": 26700 }, { "epoch": 0.16129810301371184, "grad_norm": 0.31546828150749207, "learning_rate": 0.00029968871099073015, "loss": 5.2066, "step": 26750 }, { "epoch": 0.16159959479504588, "grad_norm": 1.0939152240753174, "learning_rate": 0.00029968739932112064, "loss": 5.9793, "step": 26800 }, { "epoch": 0.16190108657637992, "grad_norm": 0.7880067229270935, "learning_rate": 0.0002996860848967282, "loss": 5.5901, "step": 26850 }, { "epoch": 0.16220257835771396, "grad_norm": 0.6796835660934448, "learning_rate": 0.0002996847677175769, "loss": 5.596, "step": 26900 }, { "epoch": 0.162504070139048, "grad_norm": 0.7355455756187439, "learning_rate": 0.0002996834477836911, "loss": 5.9722, "step": 26950 }, { "epoch": 0.16280556192038206, "grad_norm": 1.0660473108291626, "learning_rate": 0.0002996821250950951, "loss": 6.1163, "step": 27000 }, { "epoch": 0.16280556192038206, "eval_loss": 6.27783203125, "eval_runtime": 39.1201, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 6.59189299553096, "step": 27000 }, { "epoch": 0.1631070537017161, "grad_norm": 1.2566839456558228, "learning_rate": 0.0002996807996518131, "loss": 6.2788, "step": 27050 }, { "epoch": 0.16340854548305014, "grad_norm": 0.322790265083313, "learning_rate": 0.00029967947145386967, "loss": 5.5652, "step": 27100 }, { "epoch": 0.16371003726438418, "grad_norm": 0.7861691117286682, "learning_rate": 0.00029967814050128915, "loss": 5.8082, "step": 27150 }, { "epoch": 0.16401152904571822, "grad_norm": 0.6268158555030823, "learning_rate": 0.00029967680679409607, "loss": 5.5453, "step": 27200 }, { "epoch": 0.16431302082705226, "grad_norm": 0.9664158821105957, "learning_rate": 0.00029967547033231493, "loss": 5.9332, "step": 27250 }, { "epoch": 0.1646145126083863, "grad_norm": 1.0398691892623901, "learning_rate": 0.00029967413111597034, "loss": 6.0712, "step": 27300 }, { "epoch": 0.16491600438972034, "grad_norm": 0.2071402668952942, "learning_rate": 0.00029967278914508696, "loss": 5.417, "step": 27350 }, { "epoch": 0.16521749617105438, "grad_norm": 1.2137508392333984, "learning_rate": 0.0002996714444196895, "loss": 5.8037, "step": 27400 }, { "epoch": 0.16551898795238842, "grad_norm": 0.7249096632003784, "learning_rate": 0.00029967009693980274, "loss": 5.5948, "step": 27450 }, { "epoch": 0.16582047973372246, "grad_norm": 1.0282317399978638, "learning_rate": 0.0002996687467054514, "loss": 6.1388, "step": 27500 }, { "epoch": 0.1661219715150565, "grad_norm": 0.5799879431724548, "learning_rate": 0.0002996673937166604, "loss": 6.3924, "step": 27550 }, { "epoch": 0.16642346329639054, "grad_norm": 1.239750862121582, "learning_rate": 0.00029966603797345454, "loss": 6.1446, "step": 27600 }, { "epoch": 0.16672495507772458, "grad_norm": 0.9596663117408752, "learning_rate": 0.00029966467947585893, "loss": 5.6571, "step": 27650 }, { "epoch": 0.16702644685905862, "grad_norm": 1.3126919269561768, "learning_rate": 0.00029966331822389844, "loss": 5.7265, "step": 27700 }, { "epoch": 0.16732793864039266, "grad_norm": 0.7126249074935913, "learning_rate": 0.00029966195421759817, "loss": 5.9778, "step": 27750 }, { "epoch": 0.1676294304217267, "grad_norm": 0.3778236210346222, "learning_rate": 0.0002996605874569831, "loss": 5.6637, "step": 27800 }, { "epoch": 0.16793092220306075, "grad_norm": 1.0018073320388794, "learning_rate": 0.00029965921794207864, "loss": 5.8568, "step": 27850 }, { "epoch": 0.16823241398439479, "grad_norm": 0.8710664510726929, "learning_rate": 0.0002996578456729098, "loss": 6.1913, "step": 27900 }, { "epoch": 0.16853390576572883, "grad_norm": 0.6624279022216797, "learning_rate": 0.0002996564706495019, "loss": 6.0248, "step": 27950 }, { "epoch": 0.16883539754706287, "grad_norm": 0.798795223236084, "learning_rate": 0.0002996550928718802, "loss": 6.1014, "step": 28000 }, { "epoch": 0.1691368893283969, "grad_norm": 0.844026505947113, "learning_rate": 0.00029965371234007005, "loss": 5.7302, "step": 28050 }, { "epoch": 0.16943838110973095, "grad_norm": 0.6853495836257935, "learning_rate": 0.00029965232905409693, "loss": 6.1065, "step": 28100 }, { "epoch": 0.16973987289106499, "grad_norm": 0.8556489944458008, "learning_rate": 0.0002996509430139862, "loss": 6.3176, "step": 28150 }, { "epoch": 0.17004136467239903, "grad_norm": 1.5645843744277954, "learning_rate": 0.00029964955421976354, "loss": 5.4234, "step": 28200 }, { "epoch": 0.17034285645373307, "grad_norm": 0.6834583282470703, "learning_rate": 0.0002996481626714544, "loss": 5.6575, "step": 28250 }, { "epoch": 0.1706443482350671, "grad_norm": 0.7330533862113953, "learning_rate": 0.00029964676836908427, "loss": 5.5658, "step": 28300 }, { "epoch": 0.17094584001640115, "grad_norm": 0.6509259939193726, "learning_rate": 0.000299645371312679, "loss": 5.6718, "step": 28350 }, { "epoch": 0.17124733179773519, "grad_norm": 1.0465052127838135, "learning_rate": 0.0002996439715022642, "loss": 6.2163, "step": 28400 }, { "epoch": 0.17154882357906923, "grad_norm": 1.0213632583618164, "learning_rate": 0.00029964256893786565, "loss": 5.4475, "step": 28450 }, { "epoch": 0.17185031536040327, "grad_norm": 0.889758825302124, "learning_rate": 0.00029964116361950923, "loss": 5.6849, "step": 28500 }, { "epoch": 0.1721518071417373, "grad_norm": 0.8377730846405029, "learning_rate": 0.0002996397555472207, "loss": 5.8302, "step": 28550 }, { "epoch": 0.17245329892307135, "grad_norm": 1.079965353012085, "learning_rate": 0.000299638344721026, "loss": 5.4097, "step": 28600 }, { "epoch": 0.1727547907044054, "grad_norm": 1.0394246578216553, "learning_rate": 0.0002996369311409511, "loss": 5.894, "step": 28650 }, { "epoch": 0.17305628248573943, "grad_norm": 1.304404854774475, "learning_rate": 0.00029963551480702207, "loss": 5.4521, "step": 28700 }, { "epoch": 0.17335777426707347, "grad_norm": 0.7137879133224487, "learning_rate": 0.00029963409571926494, "loss": 5.5202, "step": 28750 }, { "epoch": 0.1736592660484075, "grad_norm": 0.7652526497840881, "learning_rate": 0.0002996326738777058, "loss": 5.6688, "step": 28800 }, { "epoch": 0.17396075782974157, "grad_norm": 0.346299409866333, "learning_rate": 0.00029963124928237086, "loss": 5.126, "step": 28850 }, { "epoch": 0.17426224961107561, "grad_norm": 1.3495221138000488, "learning_rate": 0.00029962982193328623, "loss": 5.9346, "step": 28900 }, { "epoch": 0.17456374139240965, "grad_norm": 0.9206358194351196, "learning_rate": 0.00029962839183047835, "loss": 6.052, "step": 28950 }, { "epoch": 0.1748652331737437, "grad_norm": 0.5165162682533264, "learning_rate": 0.0002996269589739734, "loss": 5.0295, "step": 29000 }, { "epoch": 0.17516672495507773, "grad_norm": 1.7423135042190552, "learning_rate": 0.0002996255233637978, "loss": 5.3789, "step": 29050 }, { "epoch": 0.17546821673641178, "grad_norm": 0.8589541912078857, "learning_rate": 0.00029962408499997794, "loss": 5.7241, "step": 29100 }, { "epoch": 0.17576970851774582, "grad_norm": 0.8374860882759094, "learning_rate": 0.00029962264388254037, "loss": 6.0127, "step": 29150 }, { "epoch": 0.17607120029907986, "grad_norm": 0.45102357864379883, "learning_rate": 0.00029962120001151155, "loss": 5.8928, "step": 29200 }, { "epoch": 0.1763726920804139, "grad_norm": 0.690015435218811, "learning_rate": 0.0002996197533869181, "loss": 6.0201, "step": 29250 }, { "epoch": 0.17667418386174794, "grad_norm": 1.1167585849761963, "learning_rate": 0.00029961830400878655, "loss": 5.2536, "step": 29300 }, { "epoch": 0.17697567564308198, "grad_norm": 0.697078287601471, "learning_rate": 0.0002996168518771437, "loss": 5.8866, "step": 29350 }, { "epoch": 0.17727716742441602, "grad_norm": 0.7666856050491333, "learning_rate": 0.0002996153969920161, "loss": 6.0954, "step": 29400 }, { "epoch": 0.17757865920575006, "grad_norm": 0.7764576077461243, "learning_rate": 0.00029961393935343077, "loss": 5.7911, "step": 29450 }, { "epoch": 0.1778801509870841, "grad_norm": 0.9142379760742188, "learning_rate": 0.0002996124789614143, "loss": 6.0536, "step": 29500 }, { "epoch": 0.17818164276841814, "grad_norm": 1.2754111289978027, "learning_rate": 0.00029961101581599367, "loss": 6.1843, "step": 29550 }, { "epoch": 0.17848313454975218, "grad_norm": 0.9652661681175232, "learning_rate": 0.0002996095499171958, "loss": 5.4801, "step": 29600 }, { "epoch": 0.17878462633108622, "grad_norm": 0.9565760493278503, "learning_rate": 0.0002996080812650477, "loss": 6.2817, "step": 29650 }, { "epoch": 0.17908611811242026, "grad_norm": 0.6391888856887817, "learning_rate": 0.0002996066098595763, "loss": 6.1873, "step": 29700 }, { "epoch": 0.1793876098937543, "grad_norm": 0.945609450340271, "learning_rate": 0.0002996051357008088, "loss": 5.9767, "step": 29750 }, { "epoch": 0.17968910167508834, "grad_norm": 0.6454881429672241, "learning_rate": 0.0002996036587887723, "loss": 6.2099, "step": 29800 }, { "epoch": 0.17999059345642238, "grad_norm": 1.3561986684799194, "learning_rate": 0.0002996021791234939, "loss": 5.6228, "step": 29850 }, { "epoch": 0.18029208523775642, "grad_norm": 1.0500048398971558, "learning_rate": 0.0002996006967050009, "loss": 5.9622, "step": 29900 }, { "epoch": 0.18059357701909046, "grad_norm": 0.8863379955291748, "learning_rate": 0.0002995992115333206, "loss": 5.6371, "step": 29950 }, { "epoch": 0.1808950688004245, "grad_norm": 0.6777317523956299, "learning_rate": 0.0002995977236084803, "loss": 5.536, "step": 30000 }, { "epoch": 0.1808950688004245, "eval_loss": 6.244597434997559, "eval_runtime": 39.1124, "eval_samples_per_second": 13.09, "eval_steps_per_second": 6.545, "eval_tts_loss": 6.655323188232187, "step": 30000 }, { "epoch": 0.18119656058175854, "grad_norm": 0.38350504636764526, "learning_rate": 0.00029959623293050736, "loss": 5.8804, "step": 30050 }, { "epoch": 0.18149805236309258, "grad_norm": 1.1050869226455688, "learning_rate": 0.0002995947394994292, "loss": 5.8673, "step": 30100 }, { "epoch": 0.18179954414442662, "grad_norm": 0.9024484157562256, "learning_rate": 0.00029959324331527344, "loss": 5.2314, "step": 30150 }, { "epoch": 0.18210103592576066, "grad_norm": 0.9609419107437134, "learning_rate": 0.00029959174437806747, "loss": 5.6797, "step": 30200 }, { "epoch": 0.1824025277070947, "grad_norm": 0.843263566493988, "learning_rate": 0.0002995902426878389, "loss": 6.198, "step": 30250 }, { "epoch": 0.18270401948842874, "grad_norm": 0.7237895131111145, "learning_rate": 0.0002995887382446154, "loss": 5.8622, "step": 30300 }, { "epoch": 0.18300551126976278, "grad_norm": 0.613498330116272, "learning_rate": 0.00029958723104842467, "loss": 6.0156, "step": 30350 }, { "epoch": 0.18330700305109682, "grad_norm": 1.0621836185455322, "learning_rate": 0.00029958572109929437, "loss": 5.1739, "step": 30400 }, { "epoch": 0.18360849483243086, "grad_norm": 0.6061580181121826, "learning_rate": 0.00029958420839725245, "loss": 5.9737, "step": 30450 }, { "epoch": 0.1839099866137649, "grad_norm": 1.0284167528152466, "learning_rate": 0.00029958269294232657, "loss": 5.7843, "step": 30500 }, { "epoch": 0.18421147839509894, "grad_norm": 0.6920543313026428, "learning_rate": 0.00029958117473454464, "loss": 5.6341, "step": 30550 }, { "epoch": 0.18451297017643298, "grad_norm": 0.8935782313346863, "learning_rate": 0.00029957965377393474, "loss": 5.9737, "step": 30600 }, { "epoch": 0.18481446195776702, "grad_norm": 0.7980189323425293, "learning_rate": 0.00029957813006052474, "loss": 5.7312, "step": 30650 }, { "epoch": 0.18511595373910109, "grad_norm": 0.6610565781593323, "learning_rate": 0.0002995766035943427, "loss": 5.8302, "step": 30700 }, { "epoch": 0.18541744552043513, "grad_norm": 1.0015157461166382, "learning_rate": 0.0002995750743754168, "loss": 5.3896, "step": 30750 }, { "epoch": 0.18571893730176917, "grad_norm": 1.8899494409561157, "learning_rate": 0.0002995735424037751, "loss": 5.4756, "step": 30800 }, { "epoch": 0.1860204290831032, "grad_norm": 0.7442725896835327, "learning_rate": 0.00029957200767944574, "loss": 5.6021, "step": 30850 }, { "epoch": 0.18632192086443725, "grad_norm": 1.0238033533096313, "learning_rate": 0.00029957047020245703, "loss": 5.7532, "step": 30900 }, { "epoch": 0.18662341264577129, "grad_norm": 0.9829300045967102, "learning_rate": 0.0002995689299728373, "loss": 5.8236, "step": 30950 }, { "epoch": 0.18692490442710533, "grad_norm": 0.8700386881828308, "learning_rate": 0.00029956738699061483, "loss": 5.4716, "step": 31000 }, { "epoch": 0.18722639620843937, "grad_norm": 0.8137874007225037, "learning_rate": 0.00029956584125581805, "loss": 6.1327, "step": 31050 }, { "epoch": 0.1875278879897734, "grad_norm": 1.043655276298523, "learning_rate": 0.0002995642927684754, "loss": 6.0843, "step": 31100 }, { "epoch": 0.18782937977110745, "grad_norm": 1.1411972045898438, "learning_rate": 0.00029956274152861544, "loss": 5.4975, "step": 31150 }, { "epoch": 0.1881308715524415, "grad_norm": 0.8675227761268616, "learning_rate": 0.0002995611875362666, "loss": 5.5838, "step": 31200 }, { "epoch": 0.18843236333377553, "grad_norm": 0.7343224287033081, "learning_rate": 0.00029955963079145756, "loss": 5.1225, "step": 31250 }, { "epoch": 0.18873385511510957, "grad_norm": 1.226737380027771, "learning_rate": 0.00029955807129421696, "loss": 6.1746, "step": 31300 }, { "epoch": 0.1890353468964436, "grad_norm": 0.8458810448646545, "learning_rate": 0.0002995565090445735, "loss": 5.5839, "step": 31350 }, { "epoch": 0.18933683867777765, "grad_norm": 1.1334621906280518, "learning_rate": 0.0002995549440425559, "loss": 5.453, "step": 31400 }, { "epoch": 0.1896383304591117, "grad_norm": 0.7378872632980347, "learning_rate": 0.00029955337628819297, "loss": 5.9349, "step": 31450 }, { "epoch": 0.18993982224044573, "grad_norm": 1.6535991430282593, "learning_rate": 0.00029955180578151357, "loss": 5.9964, "step": 31500 }, { "epoch": 0.19024131402177977, "grad_norm": 1.3575032949447632, "learning_rate": 0.0002995502325225466, "loss": 5.3922, "step": 31550 }, { "epoch": 0.1905428058031138, "grad_norm": 1.1317638158798218, "learning_rate": 0.00029954865651132107, "loss": 6.0338, "step": 31600 }, { "epoch": 0.19084429758444785, "grad_norm": 0.7753003835678101, "learning_rate": 0.00029954707774786587, "loss": 5.9103, "step": 31650 }, { "epoch": 0.1911457893657819, "grad_norm": 0.6257813572883606, "learning_rate": 0.0002995454962322102, "loss": 5.5436, "step": 31700 }, { "epoch": 0.19144728114711593, "grad_norm": 0.6421615481376648, "learning_rate": 0.000299543911964383, "loss": 5.8105, "step": 31750 }, { "epoch": 0.19174877292844997, "grad_norm": 0.9760420322418213, "learning_rate": 0.0002995423249444135, "loss": 5.6155, "step": 31800 }, { "epoch": 0.192050264709784, "grad_norm": 2.79966402053833, "learning_rate": 0.00029954073517233105, "loss": 5.9085, "step": 31850 }, { "epoch": 0.19235175649111805, "grad_norm": 0.7176964282989502, "learning_rate": 0.0002995391426481647, "loss": 5.9981, "step": 31900 }, { "epoch": 0.1926532482724521, "grad_norm": 0.7628802061080933, "learning_rate": 0.0002995375473719438, "loss": 6.0613, "step": 31950 }, { "epoch": 0.19295474005378613, "grad_norm": 1.3332328796386719, "learning_rate": 0.00029953594934369775, "loss": 5.9771, "step": 32000 }, { "epoch": 0.19325623183512017, "grad_norm": 0.46233248710632324, "learning_rate": 0.00029953434856345594, "loss": 5.5821, "step": 32050 }, { "epoch": 0.1935577236164542, "grad_norm": 0.9309054613113403, "learning_rate": 0.0002995327450312479, "loss": 5.568, "step": 32100 }, { "epoch": 0.19385921539778825, "grad_norm": 1.221728801727295, "learning_rate": 0.000299531138747103, "loss": 5.8276, "step": 32150 }, { "epoch": 0.1941607071791223, "grad_norm": 0.718416690826416, "learning_rate": 0.000299529529711051, "loss": 5.3501, "step": 32200 }, { "epoch": 0.19446219896045633, "grad_norm": 0.9893563985824585, "learning_rate": 0.0002995279179231213, "loss": 6.032, "step": 32250 }, { "epoch": 0.19476369074179037, "grad_norm": 0.7511883974075317, "learning_rate": 0.0002995263033833437, "loss": 5.4759, "step": 32300 }, { "epoch": 0.1950651825231244, "grad_norm": 0.9407055377960205, "learning_rate": 0.0002995246860917479, "loss": 5.8143, "step": 32350 }, { "epoch": 0.19536667430445845, "grad_norm": 0.5538185238838196, "learning_rate": 0.00029952306604836355, "loss": 5.7495, "step": 32400 }, { "epoch": 0.1956681660857925, "grad_norm": 0.8587387800216675, "learning_rate": 0.0002995214432532206, "loss": 5.5412, "step": 32450 }, { "epoch": 0.19596965786712653, "grad_norm": 0.9932447671890259, "learning_rate": 0.00029951981770634886, "loss": 6.0371, "step": 32500 }, { "epoch": 0.1962711496484606, "grad_norm": 0.4897419214248657, "learning_rate": 0.0002995181894077783, "loss": 5.6386, "step": 32550 }, { "epoch": 0.19657264142979464, "grad_norm": 1.1052745580673218, "learning_rate": 0.00029951655835753875, "loss": 6.0553, "step": 32600 }, { "epoch": 0.19687413321112868, "grad_norm": 0.6619361639022827, "learning_rate": 0.00029951492455566035, "loss": 5.4631, "step": 32650 }, { "epoch": 0.19717562499246272, "grad_norm": 0.6642810702323914, "learning_rate": 0.00029951328800217317, "loss": 5.8034, "step": 32700 }, { "epoch": 0.19747711677379676, "grad_norm": 1.182899832725525, "learning_rate": 0.00029951164869710723, "loss": 5.1894, "step": 32750 }, { "epoch": 0.1977786085551308, "grad_norm": 0.8860312104225159, "learning_rate": 0.0002995100066404928, "loss": 5.8634, "step": 32800 }, { "epoch": 0.19808010033646484, "grad_norm": 0.6597983837127686, "learning_rate": 0.00029950836183236, "loss": 5.6958, "step": 32850 }, { "epoch": 0.19838159211779888, "grad_norm": 0.6668276786804199, "learning_rate": 0.00029950671427273926, "loss": 5.7664, "step": 32900 }, { "epoch": 0.19868308389913292, "grad_norm": 0.8033652305603027, "learning_rate": 0.00029950506396166077, "loss": 5.8901, "step": 32950 }, { "epoch": 0.19898457568046696, "grad_norm": 0.8728138208389282, "learning_rate": 0.0002995034108991549, "loss": 5.7462, "step": 33000 }, { "epoch": 0.19898457568046696, "eval_loss": 6.203675270080566, "eval_runtime": 38.9196, "eval_samples_per_second": 13.155, "eval_steps_per_second": 6.578, "eval_tts_loss": 6.654470839057997, "step": 33000 }, { "epoch": 0.199286067461801, "grad_norm": 1.428969144821167, "learning_rate": 0.0002995017550852521, "loss": 5.5023, "step": 33050 }, { "epoch": 0.19958755924313504, "grad_norm": 0.9233891367912292, "learning_rate": 0.0002995000965199829, "loss": 6.1663, "step": 33100 }, { "epoch": 0.19988905102446908, "grad_norm": 0.6567472219467163, "learning_rate": 0.0002994984352033777, "loss": 5.151, "step": 33150 }, { "epoch": 0.20019054280580312, "grad_norm": 0.8571580052375793, "learning_rate": 0.00029949677113546723, "loss": 5.557, "step": 33200 }, { "epoch": 0.20049203458713716, "grad_norm": 0.6078487038612366, "learning_rate": 0.00029949510431628195, "loss": 5.6409, "step": 33250 }, { "epoch": 0.2007935263684712, "grad_norm": 0.8290396928787231, "learning_rate": 0.00029949343474585267, "loss": 5.8491, "step": 33300 }, { "epoch": 0.20109501814980524, "grad_norm": 0.7085034251213074, "learning_rate": 0.00029949176242421, "loss": 5.3327, "step": 33350 }, { "epoch": 0.20139650993113928, "grad_norm": 1.1288365125656128, "learning_rate": 0.0002994900873513849, "loss": 5.4397, "step": 33400 }, { "epoch": 0.20169800171247332, "grad_norm": 1.1679863929748535, "learning_rate": 0.00029948840952740796, "loss": 5.4475, "step": 33450 }, { "epoch": 0.20199949349380736, "grad_norm": 0.6894846558570862, "learning_rate": 0.0002994867289523102, "loss": 5.6383, "step": 33500 }, { "epoch": 0.2023009852751414, "grad_norm": 0.5424469709396362, "learning_rate": 0.00029948504562612256, "loss": 5.7153, "step": 33550 }, { "epoch": 0.20260247705647544, "grad_norm": 1.046682596206665, "learning_rate": 0.00029948335954887593, "loss": 5.8695, "step": 33600 }, { "epoch": 0.20290396883780948, "grad_norm": 0.28510087728500366, "learning_rate": 0.00029948167072060144, "loss": 5.4503, "step": 33650 }, { "epoch": 0.20320546061914352, "grad_norm": 0.8443584442138672, "learning_rate": 0.0002994799791413301, "loss": 5.8021, "step": 33700 }, { "epoch": 0.20350695240047756, "grad_norm": 0.91343092918396, "learning_rate": 0.00029947828481109305, "loss": 5.7654, "step": 33750 }, { "epoch": 0.2038084441818116, "grad_norm": 1.260368824005127, "learning_rate": 0.00029947658772992155, "loss": 5.9563, "step": 33800 }, { "epoch": 0.20410993596314564, "grad_norm": 0.7925406098365784, "learning_rate": 0.0002994748878978467, "loss": 5.5193, "step": 33850 }, { "epoch": 0.20441142774447968, "grad_norm": 0.7077336311340332, "learning_rate": 0.0002994731853148999, "loss": 5.7978, "step": 33900 }, { "epoch": 0.20471291952581372, "grad_norm": 1.4308980703353882, "learning_rate": 0.00029947147998111236, "loss": 5.3792, "step": 33950 }, { "epoch": 0.20501441130714776, "grad_norm": 1.3471267223358154, "learning_rate": 0.0002994697718965156, "loss": 5.6714, "step": 34000 }, { "epoch": 0.2053159030884818, "grad_norm": 1.026523470878601, "learning_rate": 0.00029946806106114095, "loss": 5.7018, "step": 34050 }, { "epoch": 0.20561739486981584, "grad_norm": 0.9758976697921753, "learning_rate": 0.00029946634747501997, "loss": 5.9553, "step": 34100 }, { "epoch": 0.20591888665114988, "grad_norm": 0.9617396593093872, "learning_rate": 0.00029946463113818413, "loss": 6.2455, "step": 34150 }, { "epoch": 0.20622037843248392, "grad_norm": 0.3525865972042084, "learning_rate": 0.0002994629120506651, "loss": 5.2828, "step": 34200 }, { "epoch": 0.20652187021381796, "grad_norm": 0.7660099267959595, "learning_rate": 0.00029946119021249446, "loss": 5.9228, "step": 34250 }, { "epoch": 0.206823361995152, "grad_norm": 0.8033444285392761, "learning_rate": 0.0002994594656237039, "loss": 5.9055, "step": 34300 }, { "epoch": 0.20712485377648604, "grad_norm": 1.3096176385879517, "learning_rate": 0.00029945773828432516, "loss": 5.563, "step": 34350 }, { "epoch": 0.2074263455578201, "grad_norm": 0.6573106050491333, "learning_rate": 0.00029945600819439, "loss": 6.2855, "step": 34400 }, { "epoch": 0.20772783733915415, "grad_norm": 0.63188236951828, "learning_rate": 0.0002994542753539303, "loss": 5.0015, "step": 34450 }, { "epoch": 0.2080293291204882, "grad_norm": 0.8749607801437378, "learning_rate": 0.00029945253976297804, "loss": 5.7298, "step": 34500 }, { "epoch": 0.20833082090182223, "grad_norm": 1.2316994667053223, "learning_rate": 0.00029945080142156495, "loss": 5.2777, "step": 34550 }, { "epoch": 0.20863231268315627, "grad_norm": 0.8167499303817749, "learning_rate": 0.0002994490603297232, "loss": 5.3778, "step": 34600 }, { "epoch": 0.2089338044644903, "grad_norm": 0.9203357100486755, "learning_rate": 0.00029944731648748476, "loss": 5.847, "step": 34650 }, { "epoch": 0.20923529624582435, "grad_norm": 0.9156278967857361, "learning_rate": 0.00029944556989488174, "loss": 6.1359, "step": 34700 }, { "epoch": 0.2095367880271584, "grad_norm": 0.7734649777412415, "learning_rate": 0.00029944382055194625, "loss": 5.9082, "step": 34750 }, { "epoch": 0.20983827980849243, "grad_norm": 0.6787011027336121, "learning_rate": 0.00029944206845871057, "loss": 5.5344, "step": 34800 }, { "epoch": 0.21013977158982647, "grad_norm": 0.3994613587856293, "learning_rate": 0.00029944031361520677, "loss": 5.7267, "step": 34850 }, { "epoch": 0.2104412633711605, "grad_norm": 1.045607566833496, "learning_rate": 0.0002994385560214674, "loss": 5.7746, "step": 34900 }, { "epoch": 0.21074275515249455, "grad_norm": 0.6796344518661499, "learning_rate": 0.00029943679567752456, "loss": 5.9117, "step": 34950 }, { "epoch": 0.2110442469338286, "grad_norm": 1.7223607301712036, "learning_rate": 0.00029943503258341076, "loss": 5.7298, "step": 35000 }, { "epoch": 0.21134573871516263, "grad_norm": 0.7007985711097717, "learning_rate": 0.0002994332667391585, "loss": 5.9214, "step": 35050 }, { "epoch": 0.21164723049649667, "grad_norm": 0.8825207352638245, "learning_rate": 0.00029943149814480016, "loss": 5.9703, "step": 35100 }, { "epoch": 0.2119487222778307, "grad_norm": 2.357522964477539, "learning_rate": 0.0002994297268003684, "loss": 5.5196, "step": 35150 }, { "epoch": 0.21225021405916475, "grad_norm": 0.51401686668396, "learning_rate": 0.00029942795270589573, "loss": 5.3417, "step": 35200 }, { "epoch": 0.2125517058404988, "grad_norm": 1.408463478088379, "learning_rate": 0.0002994261758614149, "loss": 5.9483, "step": 35250 }, { "epoch": 0.21285319762183283, "grad_norm": 0.9654785990715027, "learning_rate": 0.00029942439626695844, "loss": 5.4187, "step": 35300 }, { "epoch": 0.21315468940316687, "grad_norm": 0.9260451197624207, "learning_rate": 0.0002994226139225593, "loss": 5.5572, "step": 35350 }, { "epoch": 0.2134561811845009, "grad_norm": 0.5950373411178589, "learning_rate": 0.0002994208288282501, "loss": 5.4409, "step": 35400 }, { "epoch": 0.21375767296583495, "grad_norm": 0.9789560437202454, "learning_rate": 0.00029941904098406384, "loss": 5.9287, "step": 35450 }, { "epoch": 0.214059164747169, "grad_norm": 0.26466694474220276, "learning_rate": 0.00029941725039003337, "loss": 5.3726, "step": 35500 }, { "epoch": 0.21436065652850303, "grad_norm": 1.0796654224395752, "learning_rate": 0.0002994154570461916, "loss": 5.4165, "step": 35550 }, { "epoch": 0.21466214830983707, "grad_norm": 0.7381030321121216, "learning_rate": 0.0002994136609525716, "loss": 5.64, "step": 35600 }, { "epoch": 0.2149636400911711, "grad_norm": 1.203645944595337, "learning_rate": 0.0002994118621092064, "loss": 5.8192, "step": 35650 }, { "epoch": 0.21526513187250515, "grad_norm": 0.3052436113357544, "learning_rate": 0.00029941006051612906, "loss": 5.9214, "step": 35700 }, { "epoch": 0.2155666236538392, "grad_norm": 0.7235275506973267, "learning_rate": 0.0002994082561733728, "loss": 5.925, "step": 35750 }, { "epoch": 0.21586811543517323, "grad_norm": 1.2800064086914062, "learning_rate": 0.0002994064490809708, "loss": 5.8127, "step": 35800 }, { "epoch": 0.21616960721650727, "grad_norm": 1.0311309099197388, "learning_rate": 0.0002994046392389564, "loss": 5.826, "step": 35850 }, { "epoch": 0.2164710989978413, "grad_norm": 0.7096371650695801, "learning_rate": 0.0002994028266473627, "loss": 5.949, "step": 35900 }, { "epoch": 0.21677259077917535, "grad_norm": 0.3503604233264923, "learning_rate": 0.00029940101130622324, "loss": 5.7934, "step": 35950 }, { "epoch": 0.2170740825605094, "grad_norm": 0.7133362293243408, "learning_rate": 0.00029939919321557136, "loss": 5.8878, "step": 36000 }, { "epoch": 0.2170740825605094, "eval_loss": 6.177604675292969, "eval_runtime": 38.9639, "eval_samples_per_second": 13.14, "eval_steps_per_second": 6.57, "eval_tts_loss": 6.798329052442516, "step": 36000 }, { "epoch": 0.21737557434184343, "grad_norm": 0.6016349792480469, "learning_rate": 0.00029939737237544056, "loss": 5.3382, "step": 36050 }, { "epoch": 0.21767706612317747, "grad_norm": 0.8493909239768982, "learning_rate": 0.0002993955487858643, "loss": 5.8772, "step": 36100 }, { "epoch": 0.2179785579045115, "grad_norm": 3.267277240753174, "learning_rate": 0.00029939372244687615, "loss": 5.3395, "step": 36150 }, { "epoch": 0.21828004968584555, "grad_norm": 1.1933790445327759, "learning_rate": 0.00029939189335850976, "loss": 5.0514, "step": 36200 }, { "epoch": 0.2185815414671796, "grad_norm": 0.7991106510162354, "learning_rate": 0.0002993900615207988, "loss": 5.6278, "step": 36250 }, { "epoch": 0.21888303324851366, "grad_norm": 0.420701801776886, "learning_rate": 0.0002993882269337769, "loss": 5.3896, "step": 36300 }, { "epoch": 0.2191845250298477, "grad_norm": 0.9231565594673157, "learning_rate": 0.0002993863895974778, "loss": 5.3163, "step": 36350 }, { "epoch": 0.21948601681118174, "grad_norm": 1.1241753101348877, "learning_rate": 0.0002993845495119355, "loss": 5.7169, "step": 36400 }, { "epoch": 0.21978750859251578, "grad_norm": 0.7133686542510986, "learning_rate": 0.0002993827066771837, "loss": 5.1062, "step": 36450 }, { "epoch": 0.22008900037384982, "grad_norm": 0.6079972386360168, "learning_rate": 0.0002993808610932564, "loss": 5.6851, "step": 36500 }, { "epoch": 0.22039049215518386, "grad_norm": 0.8592256307601929, "learning_rate": 0.0002993790127601875, "loss": 5.7501, "step": 36550 }, { "epoch": 0.2206919839365179, "grad_norm": 0.9237623810768127, "learning_rate": 0.000299377161678011, "loss": 5.6914, "step": 36600 }, { "epoch": 0.22099347571785194, "grad_norm": 0.8226586580276489, "learning_rate": 0.0002993753078467611, "loss": 5.6975, "step": 36650 }, { "epoch": 0.22129496749918598, "grad_norm": 0.8427727222442627, "learning_rate": 0.0002993734512664718, "loss": 5.3458, "step": 36700 }, { "epoch": 0.22159645928052002, "grad_norm": 1.2071361541748047, "learning_rate": 0.00029937159193717727, "loss": 5.5163, "step": 36750 }, { "epoch": 0.22189795106185406, "grad_norm": 1.1641355752944946, "learning_rate": 0.0002993697298589118, "loss": 5.4904, "step": 36800 }, { "epoch": 0.2221994428431881, "grad_norm": 0.3434714674949646, "learning_rate": 0.0002993678650317096, "loss": 5.8138, "step": 36850 }, { "epoch": 0.22250093462452214, "grad_norm": 1.1161950826644897, "learning_rate": 0.000299365997455605, "loss": 5.4924, "step": 36900 }, { "epoch": 0.22280242640585618, "grad_norm": 1.0036495923995972, "learning_rate": 0.00029936412713063234, "loss": 5.8668, "step": 36950 }, { "epoch": 0.22310391818719022, "grad_norm": 0.8869696855545044, "learning_rate": 0.0002993622540568261, "loss": 5.5314, "step": 37000 }, { "epoch": 0.22340540996852426, "grad_norm": 0.6879080533981323, "learning_rate": 0.0002993603782342207, "loss": 5.8051, "step": 37050 }, { "epoch": 0.2237069017498583, "grad_norm": 0.8208459615707397, "learning_rate": 0.0002993584996628507, "loss": 5.5058, "step": 37100 }, { "epoch": 0.22400839353119234, "grad_norm": 0.7945531606674194, "learning_rate": 0.00029935661834275067, "loss": 5.3629, "step": 37150 }, { "epoch": 0.22430988531252638, "grad_norm": 0.634579598903656, "learning_rate": 0.0002993547342739552, "loss": 5.5204, "step": 37200 }, { "epoch": 0.22461137709386042, "grad_norm": 0.7066417932510376, "learning_rate": 0.0002993528474564991, "loss": 5.5052, "step": 37250 }, { "epoch": 0.22491286887519446, "grad_norm": 1.2294734716415405, "learning_rate": 0.0002993509578904169, "loss": 5.7526, "step": 37300 }, { "epoch": 0.2252143606565285, "grad_norm": 0.863842248916626, "learning_rate": 0.0002993490655757434, "loss": 5.6839, "step": 37350 }, { "epoch": 0.22551585243786254, "grad_norm": 0.7017498016357422, "learning_rate": 0.00029934717051251355, "loss": 5.6709, "step": 37400 }, { "epoch": 0.22581734421919658, "grad_norm": 0.7737565636634827, "learning_rate": 0.0002993452727007621, "loss": 5.767, "step": 37450 }, { "epoch": 0.22611883600053062, "grad_norm": 1.2495025396347046, "learning_rate": 0.0002993433721405241, "loss": 5.3989, "step": 37500 }, { "epoch": 0.22642032778186466, "grad_norm": 0.9943605661392212, "learning_rate": 0.0002993414688318344, "loss": 5.2253, "step": 37550 }, { "epoch": 0.2267218195631987, "grad_norm": 1.0650808811187744, "learning_rate": 0.0002993395627747281, "loss": 6.0206, "step": 37600 }, { "epoch": 0.22702331134453274, "grad_norm": 0.4302443563938141, "learning_rate": 0.0002993376539692403, "loss": 5.727, "step": 37650 }, { "epoch": 0.22732480312586678, "grad_norm": 0.8434755206108093, "learning_rate": 0.0002993357424154061, "loss": 6.0385, "step": 37700 }, { "epoch": 0.22762629490720082, "grad_norm": 1.636345624923706, "learning_rate": 0.0002993338281132606, "loss": 5.7433, "step": 37750 }, { "epoch": 0.22792778668853486, "grad_norm": 0.6742101311683655, "learning_rate": 0.0002993319110628392, "loss": 5.735, "step": 37800 }, { "epoch": 0.2282292784698689, "grad_norm": 1.2941607236862183, "learning_rate": 0.00029932999126417706, "loss": 5.257, "step": 37850 }, { "epoch": 0.22853077025120294, "grad_norm": 0.8701023459434509, "learning_rate": 0.00029932806871730953, "loss": 5.9228, "step": 37900 }, { "epoch": 0.22883226203253698, "grad_norm": 1.3218936920166016, "learning_rate": 0.00029932614342227193, "loss": 5.8875, "step": 37950 }, { "epoch": 0.22913375381387102, "grad_norm": 0.7481233477592468, "learning_rate": 0.0002993242153790998, "loss": 5.7453, "step": 38000 }, { "epoch": 0.22943524559520506, "grad_norm": 0.8515599370002747, "learning_rate": 0.00029932228458782865, "loss": 5.9547, "step": 38050 }, { "epoch": 0.2297367373765391, "grad_norm": 1.1230281591415405, "learning_rate": 0.0002993203510484939, "loss": 5.8341, "step": 38100 }, { "epoch": 0.23003822915787317, "grad_norm": 0.6871639490127563, "learning_rate": 0.0002993184147611311, "loss": 5.2751, "step": 38150 }, { "epoch": 0.2303397209392072, "grad_norm": 1.1182130575180054, "learning_rate": 0.0002993164757257761, "loss": 5.0585, "step": 38200 }, { "epoch": 0.23064121272054125, "grad_norm": 1.2863048315048218, "learning_rate": 0.00029931453394246437, "loss": 5.4582, "step": 38250 }, { "epoch": 0.2309427045018753, "grad_norm": 0.6852511167526245, "learning_rate": 0.0002993125894112318, "loss": 5.2697, "step": 38300 }, { "epoch": 0.23124419628320933, "grad_norm": 0.6605222225189209, "learning_rate": 0.000299310642132114, "loss": 5.3509, "step": 38350 }, { "epoch": 0.23154568806454337, "grad_norm": 0.8887448310852051, "learning_rate": 0.000299308692105147, "loss": 5.8436, "step": 38400 }, { "epoch": 0.2318471798458774, "grad_norm": 0.868958055973053, "learning_rate": 0.0002993067393303665, "loss": 5.7505, "step": 38450 }, { "epoch": 0.23214867162721145, "grad_norm": 0.4443792402744293, "learning_rate": 0.0002993047838078086, "loss": 5.5584, "step": 38500 }, { "epoch": 0.2324501634085455, "grad_norm": 0.8734575510025024, "learning_rate": 0.0002993028255375092, "loss": 5.4245, "step": 38550 }, { "epoch": 0.23275165518987953, "grad_norm": 1.2025113105773926, "learning_rate": 0.0002993008645195044, "loss": 5.6703, "step": 38600 }, { "epoch": 0.23305314697121357, "grad_norm": 0.6866753101348877, "learning_rate": 0.00029929890075383023, "loss": 5.7907, "step": 38650 }, { "epoch": 0.2333546387525476, "grad_norm": 1.0846598148345947, "learning_rate": 0.00029929693424052284, "loss": 5.7358, "step": 38700 }, { "epoch": 0.23365613053388165, "grad_norm": 1.0715410709381104, "learning_rate": 0.0002992949649796185, "loss": 5.2188, "step": 38750 }, { "epoch": 0.2339576223152157, "grad_norm": 0.9942847490310669, "learning_rate": 0.0002992929929711533, "loss": 5.9139, "step": 38800 }, { "epoch": 0.23425911409654973, "grad_norm": 0.9440348744392395, "learning_rate": 0.0002992910182151636, "loss": 5.929, "step": 38850 }, { "epoch": 0.23456060587788377, "grad_norm": 0.865071713924408, "learning_rate": 0.0002992890407116858, "loss": 6.0968, "step": 38900 }, { "epoch": 0.2348620976592178, "grad_norm": 0.8507125973701477, "learning_rate": 0.0002992870604607563, "loss": 5.632, "step": 38950 }, { "epoch": 0.23516358944055185, "grad_norm": 1.1168873310089111, "learning_rate": 0.0002992850774624114, "loss": 5.9427, "step": 39000 }, { "epoch": 0.23516358944055185, "eval_loss": 6.176456928253174, "eval_runtime": 38.9241, "eval_samples_per_second": 13.154, "eval_steps_per_second": 6.577, "eval_tts_loss": 6.695558188714054, "step": 39000 }, { "epoch": 0.2354650812218859, "grad_norm": 0.7778925895690918, "learning_rate": 0.0002992830917166878, "loss": 5.9696, "step": 39050 }, { "epoch": 0.23576657300321993, "grad_norm": 0.6787109971046448, "learning_rate": 0.00029928110322362185, "loss": 5.6941, "step": 39100 }, { "epoch": 0.23606806478455397, "grad_norm": 0.5592363476753235, "learning_rate": 0.00029927911198325026, "loss": 5.8928, "step": 39150 }, { "epoch": 0.236369556565888, "grad_norm": 0.9353315830230713, "learning_rate": 0.0002992771179956097, "loss": 5.4616, "step": 39200 }, { "epoch": 0.23667104834722205, "grad_norm": 1.534864902496338, "learning_rate": 0.00029927512126073675, "loss": 5.648, "step": 39250 }, { "epoch": 0.2369725401285561, "grad_norm": 1.2831447124481201, "learning_rate": 0.0002992731217786682, "loss": 5.9965, "step": 39300 }, { "epoch": 0.23727403190989013, "grad_norm": 0.8142523169517517, "learning_rate": 0.0002992711195494409, "loss": 5.5896, "step": 39350 }, { "epoch": 0.23757552369122417, "grad_norm": 0.7090405225753784, "learning_rate": 0.00029926911457309164, "loss": 5.7927, "step": 39400 }, { "epoch": 0.2378770154725582, "grad_norm": 1.6040847301483154, "learning_rate": 0.0002992671068496574, "loss": 5.5695, "step": 39450 }, { "epoch": 0.23817850725389225, "grad_norm": 0.6957007646560669, "learning_rate": 0.00029926509637917504, "loss": 5.6164, "step": 39500 }, { "epoch": 0.2384799990352263, "grad_norm": 0.9968054890632629, "learning_rate": 0.0002992630831616816, "loss": 5.5684, "step": 39550 }, { "epoch": 0.23878149081656033, "grad_norm": 0.8350685238838196, "learning_rate": 0.0002992610671972141, "loss": 5.2238, "step": 39600 }, { "epoch": 0.23908298259789437, "grad_norm": 0.8477494120597839, "learning_rate": 0.00029925904848580973, "loss": 5.4138, "step": 39650 }, { "epoch": 0.2393844743792284, "grad_norm": 0.4250677525997162, "learning_rate": 0.00029925702702750554, "loss": 5.7897, "step": 39700 }, { "epoch": 0.23968596616056245, "grad_norm": 0.9327124953269958, "learning_rate": 0.0002992550028223387, "loss": 5.524, "step": 39750 }, { "epoch": 0.2399874579418965, "grad_norm": 0.6679199934005737, "learning_rate": 0.0002992529758703466, "loss": 5.8148, "step": 39800 }, { "epoch": 0.24028894972323053, "grad_norm": 0.3222520649433136, "learning_rate": 0.0002992509461715665, "loss": 5.6626, "step": 39850 }, { "epoch": 0.24059044150456457, "grad_norm": 1.597019076347351, "learning_rate": 0.0002992489137260357, "loss": 5.6782, "step": 39900 }, { "epoch": 0.2408919332858986, "grad_norm": 0.7556604743003845, "learning_rate": 0.0002992468785337916, "loss": 5.4119, "step": 39950 }, { "epoch": 0.24119342506723268, "grad_norm": 0.4057804346084595, "learning_rate": 0.0002992448405948717, "loss": 5.2569, "step": 40000 }, { "epoch": 0.24149491684856672, "grad_norm": 0.7154269814491272, "learning_rate": 0.0002992427999093135, "loss": 5.3925, "step": 40050 }, { "epoch": 0.24179640862990076, "grad_norm": 0.39007192850112915, "learning_rate": 0.0002992407564771546, "loss": 5.2268, "step": 40100 }, { "epoch": 0.2420979004112348, "grad_norm": 0.8535629510879517, "learning_rate": 0.0002992387102984325, "loss": 5.5114, "step": 40150 }, { "epoch": 0.24239939219256884, "grad_norm": 0.5998932123184204, "learning_rate": 0.00029923666137318493, "loss": 5.4424, "step": 40200 }, { "epoch": 0.24270088397390288, "grad_norm": 0.4233469069004059, "learning_rate": 0.0002992346097014496, "loss": 4.7012, "step": 40250 }, { "epoch": 0.24300237575523692, "grad_norm": 0.6151072382926941, "learning_rate": 0.0002992325552832642, "loss": 5.3967, "step": 40300 }, { "epoch": 0.24330386753657096, "grad_norm": 1.208114504814148, "learning_rate": 0.0002992304981186666, "loss": 5.7831, "step": 40350 }, { "epoch": 0.243605359317905, "grad_norm": 1.238381028175354, "learning_rate": 0.00029922843820769465, "loss": 5.8503, "step": 40400 }, { "epoch": 0.24390685109923904, "grad_norm": 0.5472038388252258, "learning_rate": 0.0002992263755503863, "loss": 5.5908, "step": 40450 }, { "epoch": 0.24420834288057308, "grad_norm": 0.6071680784225464, "learning_rate": 0.0002992243101467794, "loss": 5.4095, "step": 40500 }, { "epoch": 0.24450983466190712, "grad_norm": 0.9499772191047668, "learning_rate": 0.00029922224199691207, "loss": 5.5054, "step": 40550 }, { "epoch": 0.24481132644324116, "grad_norm": 0.6782316565513611, "learning_rate": 0.0002992201711008223, "loss": 5.3963, "step": 40600 }, { "epoch": 0.2451128182245752, "grad_norm": 0.6434725522994995, "learning_rate": 0.0002992180974585482, "loss": 6.077, "step": 40650 }, { "epoch": 0.24541431000590924, "grad_norm": 0.9563531875610352, "learning_rate": 0.000299216021070128, "loss": 5.9052, "step": 40700 }, { "epoch": 0.24571580178724328, "grad_norm": 0.7196192145347595, "learning_rate": 0.0002992139419355999, "loss": 5.3831, "step": 40750 }, { "epoch": 0.24601729356857732, "grad_norm": 0.9343360662460327, "learning_rate": 0.0002992118600550021, "loss": 5.6441, "step": 40800 }, { "epoch": 0.24631878534991136, "grad_norm": 1.3392139673233032, "learning_rate": 0.000299209775428373, "loss": 5.2413, "step": 40850 }, { "epoch": 0.2466202771312454, "grad_norm": 0.9116024374961853, "learning_rate": 0.0002992076880557509, "loss": 5.3501, "step": 40900 }, { "epoch": 0.24692176891257944, "grad_norm": 0.5895945429801941, "learning_rate": 0.00029920559793717414, "loss": 5.2109, "step": 40950 }, { "epoch": 0.24722326069391348, "grad_norm": 0.7687879204750061, "learning_rate": 0.0002992035050726814, "loss": 5.5101, "step": 41000 }, { "epoch": 0.24752475247524752, "grad_norm": 0.6981392502784729, "learning_rate": 0.000299201409462311, "loss": 5.6718, "step": 41050 }, { "epoch": 0.24782624425658156, "grad_norm": 1.0416169166564941, "learning_rate": 0.0002991993111061016, "loss": 5.4736, "step": 41100 }, { "epoch": 0.2481277360379156, "grad_norm": 0.7597665786743164, "learning_rate": 0.0002991972100040918, "loss": 5.5339, "step": 41150 }, { "epoch": 0.24842922781924964, "grad_norm": 0.8918690085411072, "learning_rate": 0.0002991951061563203, "loss": 5.5845, "step": 41200 }, { "epoch": 0.24873071960058368, "grad_norm": 1.1933698654174805, "learning_rate": 0.00029919299956282573, "loss": 5.436, "step": 41250 }, { "epoch": 0.24903221138191772, "grad_norm": 0.8983391523361206, "learning_rate": 0.000299190890223647, "loss": 5.9103, "step": 41300 }, { "epoch": 0.24933370316325176, "grad_norm": 0.6299484372138977, "learning_rate": 0.0002991887781388228, "loss": 5.5255, "step": 41350 }, { "epoch": 0.2496351949445858, "grad_norm": 0.5601974129676819, "learning_rate": 0.00029918666330839194, "loss": 5.2604, "step": 41400 }, { "epoch": 0.24993668672591984, "grad_norm": 0.6841612458229065, "learning_rate": 0.00029918454573239355, "loss": 5.9307, "step": 41450 }, { "epoch": 0.2502381785072539, "grad_norm": 0.8626320362091064, "learning_rate": 0.00029918242541086647, "loss": 5.8494, "step": 41500 }, { "epoch": 0.2505396702885879, "grad_norm": 0.9264465570449829, "learning_rate": 0.00029918030234384974, "loss": 5.772, "step": 41550 }, { "epoch": 0.25084116206992196, "grad_norm": 1.0838125944137573, "learning_rate": 0.00029917817653138244, "loss": 5.7249, "step": 41600 }, { "epoch": 0.251142653851256, "grad_norm": 0.950257420539856, "learning_rate": 0.00029917604797350365, "loss": 6.05, "step": 41650 }, { "epoch": 0.25144414563259004, "grad_norm": 1.208414077758789, "learning_rate": 0.00029917391667025263, "loss": 5.4713, "step": 41700 }, { "epoch": 0.2517456374139241, "grad_norm": 0.7847824692726135, "learning_rate": 0.0002991717826216686, "loss": 5.5998, "step": 41750 }, { "epoch": 0.2520471291952581, "grad_norm": 0.7922324538230896, "learning_rate": 0.0002991696458277907, "loss": 5.5616, "step": 41800 }, { "epoch": 0.25234862097659216, "grad_norm": 0.6981843113899231, "learning_rate": 0.0002991675062886584, "loss": 5.6759, "step": 41850 }, { "epoch": 0.2526501127579262, "grad_norm": 0.8958069682121277, "learning_rate": 0.000299165364004311, "loss": 5.8352, "step": 41900 }, { "epoch": 0.25295160453926024, "grad_norm": 0.6846047639846802, "learning_rate": 0.000299163218974788, "loss": 5.4378, "step": 41950 }, { "epoch": 0.2532530963205943, "grad_norm": 0.31502676010131836, "learning_rate": 0.00029916107120012874, "loss": 5.5195, "step": 42000 }, { "epoch": 0.2532530963205943, "eval_loss": 6.1359333992004395, "eval_runtime": 39.0259, "eval_samples_per_second": 13.119, "eval_steps_per_second": 6.56, "eval_tts_loss": 6.714728359801998, "step": 42000 }, { "epoch": 0.2535545881019283, "grad_norm": 0.986343502998352, "learning_rate": 0.00029915892068037287, "loss": 5.6008, "step": 42050 }, { "epoch": 0.25385607988326236, "grad_norm": 1.168291687965393, "learning_rate": 0.0002991567674155599, "loss": 5.4662, "step": 42100 }, { "epoch": 0.2541575716645964, "grad_norm": 1.509806752204895, "learning_rate": 0.0002991546114057295, "loss": 6.0371, "step": 42150 }, { "epoch": 0.25445906344593044, "grad_norm": 1.3297762870788574, "learning_rate": 0.0002991524526509213, "loss": 5.4547, "step": 42200 }, { "epoch": 0.2547605552272645, "grad_norm": 0.9540511965751648, "learning_rate": 0.0002991502911511751, "loss": 5.1218, "step": 42250 }, { "epoch": 0.2550620470085985, "grad_norm": 0.48395660519599915, "learning_rate": 0.0002991481269065307, "loss": 6.0519, "step": 42300 }, { "epoch": 0.25536353878993256, "grad_norm": 0.7590111494064331, "learning_rate": 0.00029914595991702777, "loss": 5.6558, "step": 42350 }, { "epoch": 0.2556650305712666, "grad_norm": 0.8144111633300781, "learning_rate": 0.0002991437901827063, "loss": 5.5417, "step": 42400 }, { "epoch": 0.25596652235260065, "grad_norm": 0.8104939460754395, "learning_rate": 0.00029914161770360626, "loss": 5.9543, "step": 42450 }, { "epoch": 0.2562680141339347, "grad_norm": 1.0394072532653809, "learning_rate": 0.00029913944247976753, "loss": 5.4444, "step": 42500 }, { "epoch": 0.2565695059152687, "grad_norm": 0.7149035930633545, "learning_rate": 0.00029913726451123024, "loss": 5.776, "step": 42550 }, { "epoch": 0.25687099769660277, "grad_norm": 1.1012612581253052, "learning_rate": 0.0002991350837980344, "loss": 5.5794, "step": 42600 }, { "epoch": 0.2571724894779368, "grad_norm": 0.7112302184104919, "learning_rate": 0.0002991329003402202, "loss": 5.2619, "step": 42650 }, { "epoch": 0.25747398125927085, "grad_norm": 0.6514622569084167, "learning_rate": 0.0002991307141378278, "loss": 6.0864, "step": 42700 }, { "epoch": 0.2577754730406049, "grad_norm": 0.769156277179718, "learning_rate": 0.00029912852519089735, "loss": 5.5187, "step": 42750 }, { "epoch": 0.258076964821939, "grad_norm": 0.8244367241859436, "learning_rate": 0.0002991263334994693, "loss": 5.4582, "step": 42800 }, { "epoch": 0.258378456603273, "grad_norm": 0.7192371487617493, "learning_rate": 0.00029912413906358384, "loss": 5.6309, "step": 42850 }, { "epoch": 0.25867994838460706, "grad_norm": 0.8060984015464783, "learning_rate": 0.00029912194188328145, "loss": 5.9035, "step": 42900 }, { "epoch": 0.2589814401659411, "grad_norm": 1.1362627744674683, "learning_rate": 0.0002991197419586025, "loss": 5.6825, "step": 42950 }, { "epoch": 0.25928293194727514, "grad_norm": 0.7644184231758118, "learning_rate": 0.0002991175392895875, "loss": 6.1461, "step": 43000 }, { "epoch": 0.2595844237286092, "grad_norm": 0.42324090003967285, "learning_rate": 0.000299115333876277, "loss": 5.7534, "step": 43050 }, { "epoch": 0.2598859155099432, "grad_norm": 1.013892650604248, "learning_rate": 0.0002991131257187116, "loss": 6.0472, "step": 43100 }, { "epoch": 0.26018740729127726, "grad_norm": 0.7068168520927429, "learning_rate": 0.0002991109148169319, "loss": 5.3374, "step": 43150 }, { "epoch": 0.2604888990726113, "grad_norm": 1.0577305555343628, "learning_rate": 0.0002991087011709786, "loss": 5.3949, "step": 43200 }, { "epoch": 0.26079039085394534, "grad_norm": 0.8929175138473511, "learning_rate": 0.00029910648478089245, "loss": 5.5261, "step": 43250 }, { "epoch": 0.2610918826352794, "grad_norm": 0.8764124512672424, "learning_rate": 0.0002991042656467143, "loss": 5.3816, "step": 43300 }, { "epoch": 0.2613933744166134, "grad_norm": 1.294346570968628, "learning_rate": 0.0002991020437684848, "loss": 6.0614, "step": 43350 }, { "epoch": 0.26169486619794746, "grad_norm": 0.2839429974555969, "learning_rate": 0.000299099819146245, "loss": 5.4754, "step": 43400 }, { "epoch": 0.2619963579792815, "grad_norm": 1.0284963846206665, "learning_rate": 0.00029909759178003585, "loss": 5.9854, "step": 43450 }, { "epoch": 0.26229784976061554, "grad_norm": 1.1160862445831299, "learning_rate": 0.0002990953616698983, "loss": 5.5209, "step": 43500 }, { "epoch": 0.2625993415419496, "grad_norm": 0.6992689967155457, "learning_rate": 0.00029909312881587334, "loss": 5.7228, "step": 43550 }, { "epoch": 0.2629008333232836, "grad_norm": 0.8646759390830994, "learning_rate": 0.0002990908932180022, "loss": 5.5784, "step": 43600 }, { "epoch": 0.26320232510461766, "grad_norm": 1.1452974081039429, "learning_rate": 0.00029908865487632584, "loss": 5.1649, "step": 43650 }, { "epoch": 0.2635038168859517, "grad_norm": 0.9141587018966675, "learning_rate": 0.0002990864137908855, "loss": 5.48, "step": 43700 }, { "epoch": 0.26380530866728574, "grad_norm": 0.7809168100357056, "learning_rate": 0.0002990841699617226, "loss": 5.6589, "step": 43750 }, { "epoch": 0.2641068004486198, "grad_norm": 0.4229006767272949, "learning_rate": 0.00029908192338887825, "loss": 6.0394, "step": 43800 }, { "epoch": 0.2644082922299538, "grad_norm": 0.8755468726158142, "learning_rate": 0.0002990796740723939, "loss": 5.9225, "step": 43850 }, { "epoch": 0.26470978401128786, "grad_norm": 0.7260099649429321, "learning_rate": 0.0002990774220123108, "loss": 5.486, "step": 43900 }, { "epoch": 0.2650112757926219, "grad_norm": 0.533941924571991, "learning_rate": 0.00029907516720867057, "loss": 5.2152, "step": 43950 }, { "epoch": 0.26531276757395594, "grad_norm": 1.0117948055267334, "learning_rate": 0.0002990729096615146, "loss": 5.7104, "step": 44000 }, { "epoch": 0.26561425935529, "grad_norm": 0.957792341709137, "learning_rate": 0.00029907064937088445, "loss": 6.0643, "step": 44050 }, { "epoch": 0.265915751136624, "grad_norm": 0.7912729978561401, "learning_rate": 0.0002990683863368218, "loss": 5.5176, "step": 44100 }, { "epoch": 0.26621724291795806, "grad_norm": 0.6352227330207825, "learning_rate": 0.0002990661205593682, "loss": 5.7922, "step": 44150 }, { "epoch": 0.2665187346992921, "grad_norm": 0.9064685106277466, "learning_rate": 0.0002990638520385653, "loss": 5.6389, "step": 44200 }, { "epoch": 0.26682022648062614, "grad_norm": 0.7445216774940491, "learning_rate": 0.00029906158077445505, "loss": 5.6276, "step": 44250 }, { "epoch": 0.2671217182619602, "grad_norm": 0.9422538876533508, "learning_rate": 0.0002990593067670791, "loss": 5.3563, "step": 44300 }, { "epoch": 0.2674232100432942, "grad_norm": 0.6338539719581604, "learning_rate": 0.0002990570300164793, "loss": 5.2149, "step": 44350 }, { "epoch": 0.26772470182462826, "grad_norm": 0.6467666029930115, "learning_rate": 0.0002990547505226976, "loss": 5.6295, "step": 44400 }, { "epoch": 0.2680261936059623, "grad_norm": 0.678004264831543, "learning_rate": 0.0002990524682857759, "loss": 5.3998, "step": 44450 }, { "epoch": 0.26832768538729634, "grad_norm": 1.395552635192871, "learning_rate": 0.00029905018330575624, "loss": 5.6426, "step": 44500 }, { "epoch": 0.2686291771686304, "grad_norm": 0.6872416138648987, "learning_rate": 0.00029904789558268065, "loss": 5.6147, "step": 44550 }, { "epoch": 0.2689306689499644, "grad_norm": 0.9608258605003357, "learning_rate": 0.00029904560511659123, "loss": 5.5799, "step": 44600 }, { "epoch": 0.26923216073129846, "grad_norm": 1.2683055400848389, "learning_rate": 0.0002990433119075302, "loss": 5.7982, "step": 44650 }, { "epoch": 0.2695336525126325, "grad_norm": 0.8259441256523132, "learning_rate": 0.0002990410159555397, "loss": 6.0028, "step": 44700 }, { "epoch": 0.26983514429396654, "grad_norm": 0.8305184841156006, "learning_rate": 0.00029903871726066195, "loss": 5.391, "step": 44750 }, { "epoch": 0.2701366360753006, "grad_norm": 0.7408981919288635, "learning_rate": 0.0002990364158229393, "loss": 5.1206, "step": 44800 }, { "epoch": 0.2704381278566346, "grad_norm": 1.1408891677856445, "learning_rate": 0.00029903411164241417, "loss": 5.3358, "step": 44850 }, { "epoch": 0.27073961963796866, "grad_norm": 0.965039074420929, "learning_rate": 0.0002990318047191289, "loss": 5.6792, "step": 44900 }, { "epoch": 0.2710411114193027, "grad_norm": 0.6136263012886047, "learning_rate": 0.00029902949505312585, "loss": 5.6312, "step": 44950 }, { "epoch": 0.27134260320063674, "grad_norm": 1.1231293678283691, "learning_rate": 0.0002990271826444477, "loss": 5.5449, "step": 45000 }, { "epoch": 0.27134260320063674, "eval_loss": 6.118330001831055, "eval_runtime": 39.0347, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 6.734658819592678, "step": 45000 }, { "epoch": 0.2716440949819708, "grad_norm": 0.3674224019050598, "learning_rate": 0.0002990248674931369, "loss": 5.4635, "step": 45050 }, { "epoch": 0.2719455867633048, "grad_norm": 0.5948559045791626, "learning_rate": 0.0002990225495992361, "loss": 5.502, "step": 45100 }, { "epoch": 0.27224707854463887, "grad_norm": 1.0841397047042847, "learning_rate": 0.0002990202289627879, "loss": 5.4196, "step": 45150 }, { "epoch": 0.2725485703259729, "grad_norm": 0.7377861142158508, "learning_rate": 0.0002990179055838351, "loss": 5.3202, "step": 45200 }, { "epoch": 0.27285006210730695, "grad_norm": 0.6696345210075378, "learning_rate": 0.00029901557946242034, "loss": 5.6562, "step": 45250 }, { "epoch": 0.273151553888641, "grad_norm": 0.6774969696998596, "learning_rate": 0.0002990132505985866, "loss": 4.9966, "step": 45300 }, { "epoch": 0.273453045669975, "grad_norm": 1.2874786853790283, "learning_rate": 0.00029901091899237655, "loss": 5.5777, "step": 45350 }, { "epoch": 0.27375453745130907, "grad_norm": 0.795248806476593, "learning_rate": 0.0002990085846438332, "loss": 5.5331, "step": 45400 }, { "epoch": 0.2740560292326431, "grad_norm": 0.676826000213623, "learning_rate": 0.00029900624755299946, "loss": 5.7035, "step": 45450 }, { "epoch": 0.27435752101397715, "grad_norm": 1.0592461824417114, "learning_rate": 0.0002990039077199185, "loss": 5.3584, "step": 45500 }, { "epoch": 0.2746590127953112, "grad_norm": 1.6144452095031738, "learning_rate": 0.0002990015651446332, "loss": 6.0251, "step": 45550 }, { "epoch": 0.2749605045766452, "grad_norm": 1.6776220798492432, "learning_rate": 0.00029899921982718666, "loss": 5.7116, "step": 45600 }, { "epoch": 0.27526199635797927, "grad_norm": 0.7779431939125061, "learning_rate": 0.00029899687176762217, "loss": 5.7314, "step": 45650 }, { "epoch": 0.2755634881393133, "grad_norm": 1.123846411705017, "learning_rate": 0.0002989945209659829, "loss": 6.0722, "step": 45700 }, { "epoch": 0.27586497992064735, "grad_norm": 0.5639986395835876, "learning_rate": 0.00029899216742231206, "loss": 5.3602, "step": 45750 }, { "epoch": 0.2761664717019814, "grad_norm": 0.39274823665618896, "learning_rate": 0.00029898981113665306, "loss": 5.4744, "step": 45800 }, { "epoch": 0.2764679634833154, "grad_norm": 0.817503035068512, "learning_rate": 0.0002989874521090491, "loss": 5.5063, "step": 45850 }, { "epoch": 0.27676945526464947, "grad_norm": 1.344138741493225, "learning_rate": 0.0002989850903395438, "loss": 5.3439, "step": 45900 }, { "epoch": 0.2770709470459835, "grad_norm": 0.7187539339065552, "learning_rate": 0.00029898272582818054, "loss": 5.3735, "step": 45950 }, { "epoch": 0.27737243882731755, "grad_norm": 1.00385582447052, "learning_rate": 0.00029898035857500277, "loss": 5.7608, "step": 46000 }, { "epoch": 0.2776739306086516, "grad_norm": 1.0441581010818481, "learning_rate": 0.0002989779885800541, "loss": 5.7555, "step": 46050 }, { "epoch": 0.2779754223899856, "grad_norm": 0.8586793541908264, "learning_rate": 0.00029897561584337813, "loss": 5.9081, "step": 46100 }, { "epoch": 0.27827691417131967, "grad_norm": 0.6429148316383362, "learning_rate": 0.0002989732403650186, "loss": 5.3811, "step": 46150 }, { "epoch": 0.2785784059526537, "grad_norm": 0.7160729765892029, "learning_rate": 0.0002989708621450192, "loss": 5.4436, "step": 46200 }, { "epoch": 0.27887989773398775, "grad_norm": 1.2902330160140991, "learning_rate": 0.0002989684811834236, "loss": 5.5048, "step": 46250 }, { "epoch": 0.2791813895153218, "grad_norm": 0.7551021575927734, "learning_rate": 0.0002989660974802757, "loss": 5.5317, "step": 46300 }, { "epoch": 0.2794828812966558, "grad_norm": 0.39231008291244507, "learning_rate": 0.0002989637110356194, "loss": 5.5633, "step": 46350 }, { "epoch": 0.27978437307798987, "grad_norm": 0.71333247423172, "learning_rate": 0.0002989613218494986, "loss": 5.5641, "step": 46400 }, { "epoch": 0.2800858648593239, "grad_norm": 1.4839171171188354, "learning_rate": 0.0002989589299219572, "loss": 5.4718, "step": 46450 }, { "epoch": 0.280387356640658, "grad_norm": 1.1275745630264282, "learning_rate": 0.0002989565352530393, "loss": 5.7774, "step": 46500 }, { "epoch": 0.28068884842199204, "grad_norm": 0.7048165798187256, "learning_rate": 0.0002989541378427889, "loss": 5.6035, "step": 46550 }, { "epoch": 0.2809903402033261, "grad_norm": 0.8127660751342773, "learning_rate": 0.00029895173769125025, "loss": 5.6801, "step": 46600 }, { "epoch": 0.2812918319846601, "grad_norm": 0.8876771330833435, "learning_rate": 0.0002989493347984674, "loss": 5.6575, "step": 46650 }, { "epoch": 0.28159332376599416, "grad_norm": 0.9950936436653137, "learning_rate": 0.0002989469291644845, "loss": 5.1158, "step": 46700 }, { "epoch": 0.2818948155473282, "grad_norm": 0.7188896536827087, "learning_rate": 0.00029894452078934604, "loss": 5.6815, "step": 46750 }, { "epoch": 0.28219630732866224, "grad_norm": 0.6909405589103699, "learning_rate": 0.00029894210967309625, "loss": 6.0133, "step": 46800 }, { "epoch": 0.2824977991099963, "grad_norm": 0.6912239193916321, "learning_rate": 0.0002989396958157794, "loss": 5.6007, "step": 46850 }, { "epoch": 0.2827992908913303, "grad_norm": 0.7441765666007996, "learning_rate": 0.00029893727921744006, "loss": 5.7761, "step": 46900 }, { "epoch": 0.28310078267266436, "grad_norm": 0.913466215133667, "learning_rate": 0.00029893485987812264, "loss": 5.5001, "step": 46950 }, { "epoch": 0.2834022744539984, "grad_norm": 0.45174989104270935, "learning_rate": 0.0002989324377978716, "loss": 5.2456, "step": 47000 }, { "epoch": 0.28370376623533244, "grad_norm": 0.6407889127731323, "learning_rate": 0.0002989300129767316, "loss": 6.13, "step": 47050 }, { "epoch": 0.2840052580166665, "grad_norm": 0.42774856090545654, "learning_rate": 0.00029892758541474727, "loss": 5.8152, "step": 47100 }, { "epoch": 0.2843067497980005, "grad_norm": 0.8379168510437012, "learning_rate": 0.0002989251551119633, "loss": 5.8085, "step": 47150 }, { "epoch": 0.28460824157933456, "grad_norm": 1.0452708005905151, "learning_rate": 0.0002989227220684243, "loss": 5.4874, "step": 47200 }, { "epoch": 0.2849097333606686, "grad_norm": 0.7005821466445923, "learning_rate": 0.0002989202862841752, "loss": 5.5007, "step": 47250 }, { "epoch": 0.28521122514200264, "grad_norm": 1.9644780158996582, "learning_rate": 0.0002989178477592607, "loss": 5.7546, "step": 47300 }, { "epoch": 0.2855127169233367, "grad_norm": 1.0736310482025146, "learning_rate": 0.0002989154064937257, "loss": 5.7018, "step": 47350 }, { "epoch": 0.2858142087046707, "grad_norm": 0.8613811135292053, "learning_rate": 0.0002989129624876152, "loss": 5.8439, "step": 47400 }, { "epoch": 0.28611570048600476, "grad_norm": 0.8550467491149902, "learning_rate": 0.0002989105157409742, "loss": 5.438, "step": 47450 }, { "epoch": 0.2864171922673388, "grad_norm": 0.9675105214118958, "learning_rate": 0.00029890806625384753, "loss": 5.4027, "step": 47500 }, { "epoch": 0.28671868404867284, "grad_norm": 1.437467336654663, "learning_rate": 0.0002989056140262805, "loss": 6.1047, "step": 47550 }, { "epoch": 0.2870201758300069, "grad_norm": 1.6224071979522705, "learning_rate": 0.0002989031590583181, "loss": 5.5341, "step": 47600 }, { "epoch": 0.2873216676113409, "grad_norm": 0.9506914019584656, "learning_rate": 0.0002989007013500056, "loss": 5.8976, "step": 47650 }, { "epoch": 0.28762315939267497, "grad_norm": 0.8174239993095398, "learning_rate": 0.0002988982409013881, "loss": 5.6462, "step": 47700 }, { "epoch": 0.287924651174009, "grad_norm": 0.9702172875404358, "learning_rate": 0.00029889577771251107, "loss": 5.3937, "step": 47750 }, { "epoch": 0.28822614295534305, "grad_norm": 0.6575331091880798, "learning_rate": 0.00029889331178341967, "loss": 5.4703, "step": 47800 }, { "epoch": 0.2885276347366771, "grad_norm": 0.6607297658920288, "learning_rate": 0.00029889084311415933, "loss": 5.8056, "step": 47850 }, { "epoch": 0.2888291265180111, "grad_norm": 0.9974874258041382, "learning_rate": 0.00029888837170477555, "loss": 5.5452, "step": 47900 }, { "epoch": 0.28913061829934517, "grad_norm": 0.8976995348930359, "learning_rate": 0.00029888589755531375, "loss": 5.1286, "step": 47950 }, { "epoch": 0.2894321100806792, "grad_norm": 0.4544975757598877, "learning_rate": 0.0002988834206658195, "loss": 5.3969, "step": 48000 }, { "epoch": 0.2894321100806792, "eval_loss": 6.108334541320801, "eval_runtime": 39.1904, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 6.829522304078356, "step": 48000 }, { "epoch": 0.28973360186201325, "grad_norm": 0.9578820466995239, "learning_rate": 0.0002988809410363383, "loss": 6.115, "step": 48050 }, { "epoch": 0.2900350936433473, "grad_norm": 1.036237120628357, "learning_rate": 0.00029887845866691584, "loss": 5.1384, "step": 48100 }, { "epoch": 0.2903365854246813, "grad_norm": 0.968936026096344, "learning_rate": 0.0002988759735575979, "loss": 5.5353, "step": 48150 }, { "epoch": 0.29063807720601537, "grad_norm": 0.7865502238273621, "learning_rate": 0.0002988734857084301, "loss": 5.5454, "step": 48200 }, { "epoch": 0.2909395689873494, "grad_norm": 1.1848315000534058, "learning_rate": 0.0002988709951194582, "loss": 5.5723, "step": 48250 }, { "epoch": 0.29124106076868345, "grad_norm": 0.9845591187477112, "learning_rate": 0.0002988685017907281, "loss": 5.1614, "step": 48300 }, { "epoch": 0.2915425525500175, "grad_norm": 1.219523549079895, "learning_rate": 0.00029886600572228566, "loss": 5.7257, "step": 48350 }, { "epoch": 0.2918440443313515, "grad_norm": 0.5474830269813538, "learning_rate": 0.0002988635069141769, "loss": 5.5996, "step": 48400 }, { "epoch": 0.29214553611268557, "grad_norm": 0.9715328216552734, "learning_rate": 0.00029886100536644767, "loss": 5.387, "step": 48450 }, { "epoch": 0.2924470278940196, "grad_norm": 0.9452165961265564, "learning_rate": 0.00029885850107914404, "loss": 5.4988, "step": 48500 }, { "epoch": 0.29274851967535365, "grad_norm": 0.708035409450531, "learning_rate": 0.0002988559940523122, "loss": 5.7823, "step": 48550 }, { "epoch": 0.2930500114566877, "grad_norm": 0.7995492219924927, "learning_rate": 0.0002988534842859982, "loss": 5.4347, "step": 48600 }, { "epoch": 0.2933515032380217, "grad_norm": 0.8792651891708374, "learning_rate": 0.00029885097178024826, "loss": 5.1886, "step": 48650 }, { "epoch": 0.29365299501935577, "grad_norm": 0.8746998906135559, "learning_rate": 0.0002988484565351086, "loss": 5.702, "step": 48700 }, { "epoch": 0.2939544868006898, "grad_norm": 1.1560460329055786, "learning_rate": 0.0002988459385506255, "loss": 5.403, "step": 48750 }, { "epoch": 0.29425597858202385, "grad_norm": 0.6681215763092041, "learning_rate": 0.0002988434178268453, "loss": 5.3224, "step": 48800 }, { "epoch": 0.2945574703633579, "grad_norm": 0.9536110758781433, "learning_rate": 0.00029884089436381443, "loss": 5.4878, "step": 48850 }, { "epoch": 0.2948589621446919, "grad_norm": 0.21892736852169037, "learning_rate": 0.0002988383681615793, "loss": 5.4317, "step": 48900 }, { "epoch": 0.29516045392602597, "grad_norm": 0.6683943867683411, "learning_rate": 0.0002988358392201864, "loss": 5.9332, "step": 48950 }, { "epoch": 0.29546194570736, "grad_norm": 0.5291156768798828, "learning_rate": 0.0002988333075396823, "loss": 5.5303, "step": 49000 }, { "epoch": 0.29576343748869405, "grad_norm": 0.8439637422561646, "learning_rate": 0.00029883077312011354, "loss": 5.7815, "step": 49050 }, { "epoch": 0.2960649292700281, "grad_norm": 1.2303354740142822, "learning_rate": 0.0002988282359615268, "loss": 5.4775, "step": 49100 }, { "epoch": 0.29636642105136213, "grad_norm": 0.6827579140663147, "learning_rate": 0.0002988256960639688, "loss": 5.7728, "step": 49150 }, { "epoch": 0.29666791283269617, "grad_norm": 0.6998234987258911, "learning_rate": 0.00029882315342748625, "loss": 5.6405, "step": 49200 }, { "epoch": 0.2969694046140302, "grad_norm": 1.6253019571304321, "learning_rate": 0.0002988206080521259, "loss": 5.5546, "step": 49250 }, { "epoch": 0.29727089639536425, "grad_norm": 1.230993390083313, "learning_rate": 0.0002988180599379346, "loss": 5.3361, "step": 49300 }, { "epoch": 0.2975723881766983, "grad_norm": 0.7373805642127991, "learning_rate": 0.0002988155090849594, "loss": 5.2917, "step": 49350 }, { "epoch": 0.29787387995803233, "grad_norm": 0.7858177423477173, "learning_rate": 0.00029881295549324705, "loss": 5.609, "step": 49400 }, { "epoch": 0.29817537173936637, "grad_norm": 1.0689878463745117, "learning_rate": 0.00029881039916284464, "loss": 5.0601, "step": 49450 }, { "epoch": 0.2984768635207004, "grad_norm": 1.1877319812774658, "learning_rate": 0.0002988078400937992, "loss": 5.7874, "step": 49500 }, { "epoch": 0.29877835530203445, "grad_norm": 0.4232880473136902, "learning_rate": 0.0002988052782861578, "loss": 5.7465, "step": 49550 }, { "epoch": 0.2990798470833685, "grad_norm": 1.0186774730682373, "learning_rate": 0.00029880271373996766, "loss": 5.7774, "step": 49600 }, { "epoch": 0.29938133886470253, "grad_norm": 1.8221079111099243, "learning_rate": 0.00029880014645527586, "loss": 5.7762, "step": 49650 }, { "epoch": 0.29968283064603657, "grad_norm": 1.038773775100708, "learning_rate": 0.0002987975764321298, "loss": 5.6942, "step": 49700 }, { "epoch": 0.2999843224273706, "grad_norm": 1.4456876516342163, "learning_rate": 0.0002987950036705766, "loss": 5.8975, "step": 49750 }, { "epoch": 0.30028581420870465, "grad_norm": 1.288627028465271, "learning_rate": 0.0002987924281706638, "loss": 5.6008, "step": 49800 }, { "epoch": 0.3005873059900387, "grad_norm": 0.5458110570907593, "learning_rate": 0.0002987898499324386, "loss": 5.831, "step": 49850 }, { "epoch": 0.30088879777137273, "grad_norm": 0.3377249836921692, "learning_rate": 0.0002987872689559486, "loss": 5.4353, "step": 49900 }, { "epoch": 0.30119028955270677, "grad_norm": 1.3093922138214111, "learning_rate": 0.0002987846852412412, "loss": 5.4966, "step": 49950 }, { "epoch": 0.3014917813340408, "grad_norm": 0.23869210481643677, "learning_rate": 0.00029878209878836403, "loss": 5.4726, "step": 50000 }, { "epoch": 0.30179327311537485, "grad_norm": 0.7585697770118713, "learning_rate": 0.0002987795095973647, "loss": 5.9291, "step": 50050 }, { "epoch": 0.3020947648967089, "grad_norm": 1.047821044921875, "learning_rate": 0.0002987769176682907, "loss": 5.1427, "step": 50100 }, { "epoch": 0.30239625667804293, "grad_norm": 0.49871549010276794, "learning_rate": 0.00029877432300118995, "loss": 5.4951, "step": 50150 }, { "epoch": 0.302697748459377, "grad_norm": 1.0167160034179688, "learning_rate": 0.00029877172559611003, "loss": 5.6879, "step": 50200 }, { "epoch": 0.30299924024071107, "grad_norm": 0.9227567315101624, "learning_rate": 0.00029876912545309886, "loss": 5.5354, "step": 50250 }, { "epoch": 0.3033007320220451, "grad_norm": 0.36893022060394287, "learning_rate": 0.0002987665225722042, "loss": 5.7762, "step": 50300 }, { "epoch": 0.30360222380337915, "grad_norm": 1.2336435317993164, "learning_rate": 0.000298763916953474, "loss": 5.577, "step": 50350 }, { "epoch": 0.3039037155847132, "grad_norm": 0.6709215641021729, "learning_rate": 0.00029876130859695624, "loss": 5.6331, "step": 50400 }, { "epoch": 0.3042052073660472, "grad_norm": 0.4164135456085205, "learning_rate": 0.00029875869750269887, "loss": 5.6418, "step": 50450 }, { "epoch": 0.30450669914738127, "grad_norm": 1.1430959701538086, "learning_rate": 0.00029875608367074997, "loss": 5.6552, "step": 50500 }, { "epoch": 0.3048081909287153, "grad_norm": 0.8271132111549377, "learning_rate": 0.00029875346710115763, "loss": 5.251, "step": 50550 }, { "epoch": 0.30510968271004935, "grad_norm": 0.8485331535339355, "learning_rate": 0.00029875084779397, "loss": 5.5458, "step": 50600 }, { "epoch": 0.3054111744913834, "grad_norm": 1.1268301010131836, "learning_rate": 0.0002987482257492353, "loss": 5.7512, "step": 50650 }, { "epoch": 0.3057126662727174, "grad_norm": 0.6918211579322815, "learning_rate": 0.00029874560096700175, "loss": 5.92, "step": 50700 }, { "epoch": 0.30601415805405147, "grad_norm": 1.1269550323486328, "learning_rate": 0.00029874297344731777, "loss": 4.9433, "step": 50750 }, { "epoch": 0.3063156498353855, "grad_norm": 1.0592695474624634, "learning_rate": 0.0002987403431902315, "loss": 5.6852, "step": 50800 }, { "epoch": 0.30661714161671955, "grad_norm": 0.6500821113586426, "learning_rate": 0.0002987377101957916, "loss": 5.943, "step": 50850 }, { "epoch": 0.3069186333980536, "grad_norm": 0.9471777677536011, "learning_rate": 0.00029873507446404637, "loss": 5.6909, "step": 50900 }, { "epoch": 0.3072201251793876, "grad_norm": 0.732241690158844, "learning_rate": 0.00029873243599504433, "loss": 5.6742, "step": 50950 }, { "epoch": 0.30752161696072167, "grad_norm": 0.884598970413208, "learning_rate": 0.0002987297947888341, "loss": 5.7821, "step": 51000 }, { "epoch": 0.30752161696072167, "eval_loss": 6.109760761260986, "eval_runtime": 39.0767, "eval_samples_per_second": 13.102, "eval_steps_per_second": 6.551, "eval_tts_loss": 6.753831111954548, "step": 51000 }, { "epoch": 0.3078231087420557, "grad_norm": 0.8772879242897034, "learning_rate": 0.0002987271508454642, "loss": 5.7434, "step": 51050 }, { "epoch": 0.30812460052338975, "grad_norm": 0.8536843061447144, "learning_rate": 0.00029872450416498336, "loss": 5.2202, "step": 51100 }, { "epoch": 0.3084260923047238, "grad_norm": 1.1156245470046997, "learning_rate": 0.00029872185474744025, "loss": 5.4875, "step": 51150 }, { "epoch": 0.3087275840860578, "grad_norm": 1.1014959812164307, "learning_rate": 0.00029871920259288363, "loss": 5.5004, "step": 51200 }, { "epoch": 0.30902907586739187, "grad_norm": 1.286461591720581, "learning_rate": 0.0002987165477013624, "loss": 5.2174, "step": 51250 }, { "epoch": 0.3093305676487259, "grad_norm": 1.1943386793136597, "learning_rate": 0.00029871389007292523, "loss": 5.6375, "step": 51300 }, { "epoch": 0.30963205943005995, "grad_norm": 0.7633079290390015, "learning_rate": 0.0002987112297076212, "loss": 5.5426, "step": 51350 }, { "epoch": 0.309933551211394, "grad_norm": 0.8073362112045288, "learning_rate": 0.00029870856660549913, "loss": 5.1799, "step": 51400 }, { "epoch": 0.310235042992728, "grad_norm": 0.5166399478912354, "learning_rate": 0.0002987059007666082, "loss": 4.976, "step": 51450 }, { "epoch": 0.31053653477406207, "grad_norm": 0.3612758219242096, "learning_rate": 0.0002987032321909974, "loss": 5.3148, "step": 51500 }, { "epoch": 0.3108380265553961, "grad_norm": 0.7279371023178101, "learning_rate": 0.0002987005608787157, "loss": 5.5339, "step": 51550 }, { "epoch": 0.31113951833673015, "grad_norm": 1.004072904586792, "learning_rate": 0.0002986978868298125, "loss": 5.2573, "step": 51600 }, { "epoch": 0.3114410101180642, "grad_norm": 0.702451765537262, "learning_rate": 0.0002986952100443368, "loss": 5.7327, "step": 51650 }, { "epoch": 0.3117425018993982, "grad_norm": 0.8771138787269592, "learning_rate": 0.00029869253052233806, "loss": 5.7856, "step": 51700 }, { "epoch": 0.31204399368073227, "grad_norm": 1.1317213773727417, "learning_rate": 0.00029868984826386544, "loss": 5.7195, "step": 51750 }, { "epoch": 0.3123454854620663, "grad_norm": 1.1676057577133179, "learning_rate": 0.0002986871632689684, "loss": 5.7653, "step": 51800 }, { "epoch": 0.31264697724340035, "grad_norm": 1.1117799282073975, "learning_rate": 0.00029868447553769626, "loss": 5.565, "step": 51850 }, { "epoch": 0.3129484690247344, "grad_norm": 0.5492343306541443, "learning_rate": 0.0002986817850700985, "loss": 5.6039, "step": 51900 }, { "epoch": 0.31324996080606843, "grad_norm": 1.2327568531036377, "learning_rate": 0.0002986790918662247, "loss": 5.0663, "step": 51950 }, { "epoch": 0.31355145258740247, "grad_norm": 0.7635775208473206, "learning_rate": 0.0002986763959261244, "loss": 5.6641, "step": 52000 }, { "epoch": 0.3138529443687365, "grad_norm": 0.7389959096908569, "learning_rate": 0.0002986736972498472, "loss": 5.5149, "step": 52050 }, { "epoch": 0.31415443615007055, "grad_norm": 1.2413599491119385, "learning_rate": 0.00029867099583744274, "loss": 5.6746, "step": 52100 }, { "epoch": 0.3144559279314046, "grad_norm": 0.41629692912101746, "learning_rate": 0.00029866829168896083, "loss": 5.7561, "step": 52150 }, { "epoch": 0.31475741971273863, "grad_norm": 0.6740813255310059, "learning_rate": 0.00029866558480445116, "loss": 5.7207, "step": 52200 }, { "epoch": 0.31505891149407267, "grad_norm": 1.0823897123336792, "learning_rate": 0.0002986628751839635, "loss": 5.7889, "step": 52250 }, { "epoch": 0.3153604032754067, "grad_norm": 0.40529385209083557, "learning_rate": 0.00029866016282754785, "loss": 4.8168, "step": 52300 }, { "epoch": 0.31566189505674075, "grad_norm": 0.556481122970581, "learning_rate": 0.000298657447735254, "loss": 5.5907, "step": 52350 }, { "epoch": 0.3159633868380748, "grad_norm": 0.7257757782936096, "learning_rate": 0.000298654729907132, "loss": 5.7053, "step": 52400 }, { "epoch": 0.31626487861940883, "grad_norm": 1.0373954772949219, "learning_rate": 0.0002986520093432318, "loss": 6.0085, "step": 52450 }, { "epoch": 0.31656637040074287, "grad_norm": 0.9873820543289185, "learning_rate": 0.0002986492860436036, "loss": 5.8451, "step": 52500 }, { "epoch": 0.3168678621820769, "grad_norm": 0.8159734606742859, "learning_rate": 0.0002986465600082973, "loss": 5.8182, "step": 52550 }, { "epoch": 0.31716935396341095, "grad_norm": 1.067101240158081, "learning_rate": 0.0002986438312373633, "loss": 5.6335, "step": 52600 }, { "epoch": 0.317470845744745, "grad_norm": 0.8555975556373596, "learning_rate": 0.00029864109973085166, "loss": 5.5152, "step": 52650 }, { "epoch": 0.31777233752607903, "grad_norm": 1.0405560731887817, "learning_rate": 0.00029863836548881266, "loss": 4.966, "step": 52700 }, { "epoch": 0.31807382930741307, "grad_norm": 0.8879379034042358, "learning_rate": 0.0002986356285112967, "loss": 5.4574, "step": 52750 }, { "epoch": 0.3183753210887471, "grad_norm": 0.8935927748680115, "learning_rate": 0.0002986328887983541, "loss": 5.8949, "step": 52800 }, { "epoch": 0.31867681287008115, "grad_norm": 0.8568660020828247, "learning_rate": 0.00029863014635003537, "loss": 5.7407, "step": 52850 }, { "epoch": 0.3189783046514152, "grad_norm": 1.120964527130127, "learning_rate": 0.00029862740116639084, "loss": 5.9574, "step": 52900 }, { "epoch": 0.31927979643274923, "grad_norm": 0.8942052125930786, "learning_rate": 0.0002986246532474711, "loss": 5.4436, "step": 52950 }, { "epoch": 0.31958128821408327, "grad_norm": 0.797913134098053, "learning_rate": 0.0002986219025933267, "loss": 5.6169, "step": 53000 }, { "epoch": 0.3198827799954173, "grad_norm": 0.6821560263633728, "learning_rate": 0.0002986191492040083, "loss": 5.7917, "step": 53050 }, { "epoch": 0.32018427177675135, "grad_norm": 0.5008537173271179, "learning_rate": 0.0002986163930795665, "loss": 5.3271, "step": 53100 }, { "epoch": 0.3204857635580854, "grad_norm": 1.1280933618545532, "learning_rate": 0.0002986136342200521, "loss": 5.3237, "step": 53150 }, { "epoch": 0.32078725533941943, "grad_norm": 1.1649082899093628, "learning_rate": 0.0002986108726255159, "loss": 5.8633, "step": 53200 }, { "epoch": 0.32108874712075347, "grad_norm": 1.5156731605529785, "learning_rate": 0.0002986081082960086, "loss": 5.417, "step": 53250 }, { "epoch": 0.3213902389020875, "grad_norm": 0.7964416742324829, "learning_rate": 0.00029860534123158115, "loss": 5.5732, "step": 53300 }, { "epoch": 0.32169173068342155, "grad_norm": 0.4814507067203522, "learning_rate": 0.0002986025714322845, "loss": 5.6127, "step": 53350 }, { "epoch": 0.3219932224647556, "grad_norm": 0.6560469269752502, "learning_rate": 0.00029859979889816955, "loss": 5.5876, "step": 53400 }, { "epoch": 0.32229471424608963, "grad_norm": 0.8748001456260681, "learning_rate": 0.0002985970236292874, "loss": 5.6601, "step": 53450 }, { "epoch": 0.32259620602742367, "grad_norm": 0.759776771068573, "learning_rate": 0.00029859424562568907, "loss": 5.5176, "step": 53500 }, { "epoch": 0.3228976978087577, "grad_norm": 1.545841097831726, "learning_rate": 0.0002985914648874257, "loss": 5.3059, "step": 53550 }, { "epoch": 0.32319918959009175, "grad_norm": 1.0237040519714355, "learning_rate": 0.0002985886814145485, "loss": 5.8799, "step": 53600 }, { "epoch": 0.3235006813714258, "grad_norm": 1.0234451293945312, "learning_rate": 0.00029858589520710867, "loss": 5.6862, "step": 53650 }, { "epoch": 0.32380217315275983, "grad_norm": 0.9254264235496521, "learning_rate": 0.0002985831062651574, "loss": 5.5684, "step": 53700 }, { "epoch": 0.32410366493409387, "grad_norm": 0.6939722299575806, "learning_rate": 0.00029858031458874617, "loss": 5.7621, "step": 53750 }, { "epoch": 0.3244051567154279, "grad_norm": 0.7266665697097778, "learning_rate": 0.0002985775201779263, "loss": 5.411, "step": 53800 }, { "epoch": 0.32470664849676195, "grad_norm": 0.5725377798080444, "learning_rate": 0.00029857472303274914, "loss": 5.2994, "step": 53850 }, { "epoch": 0.325008140278096, "grad_norm": 0.7244455814361572, "learning_rate": 0.0002985719231532663, "loss": 5.5366, "step": 53900 }, { "epoch": 0.3253096320594301, "grad_norm": 1.4813069105148315, "learning_rate": 0.0002985691205395292, "loss": 5.9715, "step": 53950 }, { "epoch": 0.3256111238407641, "grad_norm": 0.7868010401725769, "learning_rate": 0.00029856631519158946, "loss": 5.4965, "step": 54000 }, { "epoch": 0.3256111238407641, "eval_loss": 6.0783538818359375, "eval_runtime": 38.8919, "eval_samples_per_second": 13.165, "eval_steps_per_second": 6.582, "eval_tts_loss": 6.854259607327995, "step": 54000 }, { "epoch": 0.32591261562209817, "grad_norm": 1.2126339673995972, "learning_rate": 0.0002985635071094987, "loss": 5.6263, "step": 54050 }, { "epoch": 0.3262141074034322, "grad_norm": 1.360251545906067, "learning_rate": 0.0002985606962933086, "loss": 5.5297, "step": 54100 }, { "epoch": 0.32651559918476625, "grad_norm": 1.0696525573730469, "learning_rate": 0.00029855788274307094, "loss": 5.5717, "step": 54150 }, { "epoch": 0.3268170909661003, "grad_norm": 1.2020008563995361, "learning_rate": 0.0002985550664588374, "loss": 5.7923, "step": 54200 }, { "epoch": 0.3271185827474343, "grad_norm": 0.8208432197570801, "learning_rate": 0.00029855224744065986, "loss": 5.5399, "step": 54250 }, { "epoch": 0.32742007452876837, "grad_norm": 0.24377647042274475, "learning_rate": 0.00029854942568859023, "loss": 5.8755, "step": 54300 }, { "epoch": 0.3277215663101024, "grad_norm": 1.193087100982666, "learning_rate": 0.0002985466012026804, "loss": 5.7769, "step": 54350 }, { "epoch": 0.32802305809143645, "grad_norm": 0.5893236994743347, "learning_rate": 0.0002985437739829824, "loss": 5.4401, "step": 54400 }, { "epoch": 0.3283245498727705, "grad_norm": 0.6907051801681519, "learning_rate": 0.00029854094402954813, "loss": 5.5958, "step": 54450 }, { "epoch": 0.32862604165410453, "grad_norm": 0.7426232695579529, "learning_rate": 0.0002985381113424298, "loss": 6.0263, "step": 54500 }, { "epoch": 0.32892753343543857, "grad_norm": 0.7051056623458862, "learning_rate": 0.00029853527592167955, "loss": 5.6358, "step": 54550 }, { "epoch": 0.3292290252167726, "grad_norm": 1.3861123323440552, "learning_rate": 0.0002985324377673495, "loss": 5.7064, "step": 54600 }, { "epoch": 0.32953051699810665, "grad_norm": 1.0730327367782593, "learning_rate": 0.0002985295968794919, "loss": 5.9123, "step": 54650 }, { "epoch": 0.3298320087794407, "grad_norm": 1.0072053670883179, "learning_rate": 0.000298526753258159, "loss": 5.3346, "step": 54700 }, { "epoch": 0.33013350056077473, "grad_norm": 1.1062756776809692, "learning_rate": 0.0002985239069034032, "loss": 5.6629, "step": 54750 }, { "epoch": 0.33043499234210877, "grad_norm": 1.000396728515625, "learning_rate": 0.0002985210578152768, "loss": 5.4411, "step": 54800 }, { "epoch": 0.3307364841234428, "grad_norm": 0.7359741926193237, "learning_rate": 0.0002985182059938323, "loss": 5.4826, "step": 54850 }, { "epoch": 0.33103797590477685, "grad_norm": 1.0396335124969482, "learning_rate": 0.0002985153514391222, "loss": 6.0545, "step": 54900 }, { "epoch": 0.3313394676861109, "grad_norm": 1.2045825719833374, "learning_rate": 0.00029851249415119904, "loss": 5.3245, "step": 54950 }, { "epoch": 0.33164095946744493, "grad_norm": 0.8202127814292908, "learning_rate": 0.0002985096341301153, "loss": 5.6469, "step": 55000 }, { "epoch": 0.33194245124877897, "grad_norm": 0.9464853405952454, "learning_rate": 0.00029850677137592367, "loss": 5.1718, "step": 55050 }, { "epoch": 0.332243943030113, "grad_norm": 0.7601866722106934, "learning_rate": 0.00029850390588867683, "loss": 5.9335, "step": 55100 }, { "epoch": 0.33254543481144705, "grad_norm": 1.2793779373168945, "learning_rate": 0.0002985010376684276, "loss": 5.3052, "step": 55150 }, { "epoch": 0.3328469265927811, "grad_norm": 0.7971204519271851, "learning_rate": 0.0002984981667152286, "loss": 5.5174, "step": 55200 }, { "epoch": 0.33314841837411513, "grad_norm": 0.7168843746185303, "learning_rate": 0.0002984952930291328, "loss": 5.1904, "step": 55250 }, { "epoch": 0.33344991015544917, "grad_norm": 0.9350612163543701, "learning_rate": 0.0002984924166101931, "loss": 5.4465, "step": 55300 }, { "epoch": 0.3337514019367832, "grad_norm": 0.8704623579978943, "learning_rate": 0.0002984895374584623, "loss": 5.5419, "step": 55350 }, { "epoch": 0.33405289371811725, "grad_norm": 0.7984592914581299, "learning_rate": 0.0002984866555739935, "loss": 5.9597, "step": 55400 }, { "epoch": 0.3343543854994513, "grad_norm": 1.1099300384521484, "learning_rate": 0.00029848377095683976, "loss": 5.8384, "step": 55450 }, { "epoch": 0.33465587728078533, "grad_norm": 1.0529706478118896, "learning_rate": 0.00029848088360705406, "loss": 5.9636, "step": 55500 }, { "epoch": 0.33495736906211937, "grad_norm": 1.3574689626693726, "learning_rate": 0.0002984779935246896, "loss": 6.1523, "step": 55550 }, { "epoch": 0.3352588608434534, "grad_norm": 0.8839330673217773, "learning_rate": 0.0002984751007097995, "loss": 5.6063, "step": 55600 }, { "epoch": 0.33556035262478745, "grad_norm": 0.6571201086044312, "learning_rate": 0.0002984722051624372, "loss": 5.9669, "step": 55650 }, { "epoch": 0.3358618444061215, "grad_norm": 0.5828574299812317, "learning_rate": 0.0002984693068826558, "loss": 5.0438, "step": 55700 }, { "epoch": 0.33616333618745553, "grad_norm": 0.8036131858825684, "learning_rate": 0.0002984664058705086, "loss": 5.6706, "step": 55750 }, { "epoch": 0.33646482796878957, "grad_norm": 0.4496825337409973, "learning_rate": 0.0002984635021260491, "loss": 5.6781, "step": 55800 }, { "epoch": 0.3367663197501236, "grad_norm": 0.9468846917152405, "learning_rate": 0.00029846059564933076, "loss": 5.6981, "step": 55850 }, { "epoch": 0.33706781153145765, "grad_norm": 1.1230944395065308, "learning_rate": 0.000298457686440407, "loss": 5.7545, "step": 55900 }, { "epoch": 0.3373693033127917, "grad_norm": 1.1313971281051636, "learning_rate": 0.0002984547744993313, "loss": 4.9628, "step": 55950 }, { "epoch": 0.33767079509412573, "grad_norm": 0.8400096893310547, "learning_rate": 0.0002984518598261574, "loss": 5.6989, "step": 56000 }, { "epoch": 0.33797228687545977, "grad_norm": 0.9196832180023193, "learning_rate": 0.0002984489424209389, "loss": 5.4616, "step": 56050 }, { "epoch": 0.3382737786567938, "grad_norm": 0.9966282248497009, "learning_rate": 0.00029844602228372945, "loss": 5.7667, "step": 56100 }, { "epoch": 0.33857527043812785, "grad_norm": 1.00424063205719, "learning_rate": 0.0002984430994145828, "loss": 5.5124, "step": 56150 }, { "epoch": 0.3388767622194619, "grad_norm": 0.5065834522247314, "learning_rate": 0.0002984401738135527, "loss": 4.8203, "step": 56200 }, { "epoch": 0.33917825400079593, "grad_norm": 0.974287748336792, "learning_rate": 0.0002984372454806931, "loss": 5.6119, "step": 56250 }, { "epoch": 0.33947974578212997, "grad_norm": 1.0984435081481934, "learning_rate": 0.0002984343144160578, "loss": 5.2683, "step": 56300 }, { "epoch": 0.339781237563464, "grad_norm": 0.8775320053100586, "learning_rate": 0.0002984313806197008, "loss": 5.2728, "step": 56350 }, { "epoch": 0.34008272934479805, "grad_norm": 0.8922436833381653, "learning_rate": 0.000298428444091676, "loss": 5.9115, "step": 56400 }, { "epoch": 0.3403842211261321, "grad_norm": 0.8025553226470947, "learning_rate": 0.00029842550483203756, "loss": 5.5572, "step": 56450 }, { "epoch": 0.34068571290746613, "grad_norm": 0.6843119859695435, "learning_rate": 0.0002984225628408395, "loss": 5.6307, "step": 56500 }, { "epoch": 0.34098720468880017, "grad_norm": 1.0536187887191772, "learning_rate": 0.00029841961811813596, "loss": 5.7175, "step": 56550 }, { "epoch": 0.3412886964701342, "grad_norm": 1.0104771852493286, "learning_rate": 0.0002984166706639812, "loss": 5.7635, "step": 56600 }, { "epoch": 0.34159018825146825, "grad_norm": 0.9906367659568787, "learning_rate": 0.00029841372047842934, "loss": 5.8656, "step": 56650 }, { "epoch": 0.3418916800328023, "grad_norm": 0.8121002912521362, "learning_rate": 0.0002984107675615348, "loss": 5.2387, "step": 56700 }, { "epoch": 0.34219317181413633, "grad_norm": 0.7296305298805237, "learning_rate": 0.0002984078119133519, "loss": 5.1308, "step": 56750 }, { "epoch": 0.34249466359547037, "grad_norm": 0.8026958107948303, "learning_rate": 0.000298404853533935, "loss": 5.7223, "step": 56800 }, { "epoch": 0.3427961553768044, "grad_norm": 1.2175474166870117, "learning_rate": 0.0002984018924233386, "loss": 5.7578, "step": 56850 }, { "epoch": 0.34309764715813845, "grad_norm": 1.0211974382400513, "learning_rate": 0.00029839892858161705, "loss": 5.6485, "step": 56900 }, { "epoch": 0.3433991389394725, "grad_norm": 0.6056095361709595, "learning_rate": 0.00029839596200882505, "loss": 5.7277, "step": 56950 }, { "epoch": 0.34370063072080653, "grad_norm": 1.6548784971237183, "learning_rate": 0.0002983929927050171, "loss": 5.897, "step": 57000 }, { "epoch": 0.34370063072080653, "eval_loss": 6.046796798706055, "eval_runtime": 39.1671, "eval_samples_per_second": 13.072, "eval_steps_per_second": 6.536, "eval_tts_loss": 6.8863961853145685, "step": 57000 }, { "epoch": 0.3440021225021406, "grad_norm": 1.2385656833648682, "learning_rate": 0.00029839002067024795, "loss": 5.9664, "step": 57050 }, { "epoch": 0.3443036142834746, "grad_norm": 0.298421174287796, "learning_rate": 0.0002983870459045722, "loss": 5.1947, "step": 57100 }, { "epoch": 0.34460510606480865, "grad_norm": 0.786775529384613, "learning_rate": 0.00029838406840804465, "loss": 5.3891, "step": 57150 }, { "epoch": 0.3449065978461427, "grad_norm": 1.088732361793518, "learning_rate": 0.00029838108818072007, "loss": 5.4047, "step": 57200 }, { "epoch": 0.34520808962747673, "grad_norm": 0.45781251788139343, "learning_rate": 0.0002983781052226533, "loss": 6.074, "step": 57250 }, { "epoch": 0.3455095814088108, "grad_norm": 1.328945517539978, "learning_rate": 0.00029837511953389925, "loss": 5.1828, "step": 57300 }, { "epoch": 0.3458110731901448, "grad_norm": 1.0101701021194458, "learning_rate": 0.0002983721311145129, "loss": 5.7174, "step": 57350 }, { "epoch": 0.34611256497147885, "grad_norm": 0.7661852836608887, "learning_rate": 0.00029836913996454916, "loss": 5.2393, "step": 57400 }, { "epoch": 0.3464140567528129, "grad_norm": 1.4983270168304443, "learning_rate": 0.0002983661460840632, "loss": 5.8131, "step": 57450 }, { "epoch": 0.34671554853414693, "grad_norm": 0.5325705409049988, "learning_rate": 0.00029836314947311, "loss": 5.0131, "step": 57500 }, { "epoch": 0.347017040315481, "grad_norm": 0.6093450784683228, "learning_rate": 0.0002983601501317448, "loss": 5.4989, "step": 57550 }, { "epoch": 0.347318532096815, "grad_norm": 1.0794061422348022, "learning_rate": 0.0002983571480600227, "loss": 5.1989, "step": 57600 }, { "epoch": 0.3476200238781491, "grad_norm": 0.8284804821014404, "learning_rate": 0.000298354143257999, "loss": 6.0741, "step": 57650 }, { "epoch": 0.34792151565948315, "grad_norm": 0.8067139387130737, "learning_rate": 0.0002983511357257291, "loss": 5.7702, "step": 57700 }, { "epoch": 0.3482230074408172, "grad_norm": 0.2389974147081375, "learning_rate": 0.0002983481254632682, "loss": 5.5784, "step": 57750 }, { "epoch": 0.34852449922215123, "grad_norm": 0.46118423342704773, "learning_rate": 0.0002983451124706717, "loss": 5.4129, "step": 57800 }, { "epoch": 0.34882599100348527, "grad_norm": 0.6862304210662842, "learning_rate": 0.0002983420967479952, "loss": 5.5937, "step": 57850 }, { "epoch": 0.3491274827848193, "grad_norm": 1.5366333723068237, "learning_rate": 0.0002983390782952941, "loss": 5.7171, "step": 57900 }, { "epoch": 0.34942897456615335, "grad_norm": 1.2994920015335083, "learning_rate": 0.0002983360571126239, "loss": 5.7399, "step": 57950 }, { "epoch": 0.3497304663474874, "grad_norm": 0.6896786689758301, "learning_rate": 0.0002983330332000403, "loss": 5.9445, "step": 58000 }, { "epoch": 0.35003195812882143, "grad_norm": 0.7184028029441833, "learning_rate": 0.00029833000655759884, "loss": 5.3281, "step": 58050 }, { "epoch": 0.35033344991015547, "grad_norm": 1.0272027254104614, "learning_rate": 0.00029832697718535533, "loss": 5.439, "step": 58100 }, { "epoch": 0.3506349416914895, "grad_norm": 0.8139055371284485, "learning_rate": 0.0002983239450833655, "loss": 5.7487, "step": 58150 }, { "epoch": 0.35093643347282355, "grad_norm": 0.7248658537864685, "learning_rate": 0.0002983209102516851, "loss": 5.4718, "step": 58200 }, { "epoch": 0.3512379252541576, "grad_norm": 0.8201451301574707, "learning_rate": 0.00029831787269037, "loss": 5.575, "step": 58250 }, { "epoch": 0.35153941703549163, "grad_norm": 2.0246455669403076, "learning_rate": 0.00029831483239947616, "loss": 5.5759, "step": 58300 }, { "epoch": 0.35184090881682567, "grad_norm": 0.9382103085517883, "learning_rate": 0.00029831178937905946, "loss": 5.2648, "step": 58350 }, { "epoch": 0.3521424005981597, "grad_norm": 0.8070986866950989, "learning_rate": 0.00029830874362917593, "loss": 5.6116, "step": 58400 }, { "epoch": 0.35244389237949375, "grad_norm": 1.1126052141189575, "learning_rate": 0.00029830569514988157, "loss": 5.6368, "step": 58450 }, { "epoch": 0.3527453841608278, "grad_norm": 0.7996619343757629, "learning_rate": 0.00029830264394123266, "loss": 5.2385, "step": 58500 }, { "epoch": 0.35304687594216183, "grad_norm": 0.7799546122550964, "learning_rate": 0.00029829959000328513, "loss": 5.1142, "step": 58550 }, { "epoch": 0.35334836772349587, "grad_norm": 0.7126521468162537, "learning_rate": 0.00029829653333609526, "loss": 5.2199, "step": 58600 }, { "epoch": 0.3536498595048299, "grad_norm": 1.254685878753662, "learning_rate": 0.0002982934739397194, "loss": 6.0042, "step": 58650 }, { "epoch": 0.35395135128616395, "grad_norm": 1.1558090448379517, "learning_rate": 0.00029829041181421374, "loss": 5.277, "step": 58700 }, { "epoch": 0.354252843067498, "grad_norm": 0.599417507648468, "learning_rate": 0.0002982873469596347, "loss": 5.225, "step": 58750 }, { "epoch": 0.35455433484883203, "grad_norm": 1.186550259590149, "learning_rate": 0.00029828427937603864, "loss": 5.7604, "step": 58800 }, { "epoch": 0.35485582663016607, "grad_norm": 0.5926681160926819, "learning_rate": 0.00029828120906348203, "loss": 5.6695, "step": 58850 }, { "epoch": 0.3551573184115001, "grad_norm": 0.7129924297332764, "learning_rate": 0.00029827813602202136, "loss": 5.4065, "step": 58900 }, { "epoch": 0.35545881019283415, "grad_norm": 0.32303985953330994, "learning_rate": 0.0002982750602517132, "loss": 5.4574, "step": 58950 }, { "epoch": 0.3557603019741682, "grad_norm": 1.0186265707015991, "learning_rate": 0.00029827198175261417, "loss": 5.5433, "step": 59000 }, { "epoch": 0.35606179375550223, "grad_norm": 0.6500131487846375, "learning_rate": 0.0002982689005247809, "loss": 5.8062, "step": 59050 }, { "epoch": 0.35636328553683627, "grad_norm": 0.8428109884262085, "learning_rate": 0.0002982658165682701, "loss": 5.1416, "step": 59100 }, { "epoch": 0.3566647773181703, "grad_norm": 0.8713069558143616, "learning_rate": 0.00029826272988313856, "loss": 5.65, "step": 59150 }, { "epoch": 0.35696626909950435, "grad_norm": 0.7756221294403076, "learning_rate": 0.0002982596404694431, "loss": 5.2796, "step": 59200 }, { "epoch": 0.3572677608808384, "grad_norm": 1.308018684387207, "learning_rate": 0.0002982565483272405, "loss": 5.7974, "step": 59250 }, { "epoch": 0.35756925266217243, "grad_norm": 0.3622822165489197, "learning_rate": 0.00029825345345658766, "loss": 5.3303, "step": 59300 }, { "epoch": 0.35787074444350647, "grad_norm": 0.672706663608551, "learning_rate": 0.0002982503558575416, "loss": 5.6045, "step": 59350 }, { "epoch": 0.3581722362248405, "grad_norm": 0.7606387138366699, "learning_rate": 0.00029824725553015933, "loss": 5.5265, "step": 59400 }, { "epoch": 0.35847372800617455, "grad_norm": 1.0037168264389038, "learning_rate": 0.00029824415247449784, "loss": 6.0302, "step": 59450 }, { "epoch": 0.3587752197875086, "grad_norm": 0.6942145824432373, "learning_rate": 0.0002982410466906143, "loss": 5.7084, "step": 59500 }, { "epoch": 0.35907671156884263, "grad_norm": 1.060225248336792, "learning_rate": 0.00029823793817856584, "loss": 6.0968, "step": 59550 }, { "epoch": 0.3593782033501767, "grad_norm": 0.5661870837211609, "learning_rate": 0.00029823482693840965, "loss": 6.2852, "step": 59600 }, { "epoch": 0.3596796951315107, "grad_norm": 0.6671584248542786, "learning_rate": 0.0002982317129702031, "loss": 5.4965, "step": 59650 }, { "epoch": 0.35998118691284475, "grad_norm": 0.7291431427001953, "learning_rate": 0.0002982285962740033, "loss": 5.7284, "step": 59700 }, { "epoch": 0.3602826786941788, "grad_norm": 0.7977154850959778, "learning_rate": 0.0002982254768498678, "loss": 5.5139, "step": 59750 }, { "epoch": 0.36058417047551283, "grad_norm": 0.9739433526992798, "learning_rate": 0.00029822235469785394, "loss": 5.5443, "step": 59800 }, { "epoch": 0.3608856622568469, "grad_norm": 0.6293472051620483, "learning_rate": 0.0002982192298180191, "loss": 5.3644, "step": 59850 }, { "epoch": 0.3611871540381809, "grad_norm": 1.1726881265640259, "learning_rate": 0.0002982161022104209, "loss": 5.7996, "step": 59900 }, { "epoch": 0.36148864581951495, "grad_norm": 0.8323904871940613, "learning_rate": 0.00029821297187511685, "loss": 5.4463, "step": 59950 }, { "epoch": 0.361790137600849, "grad_norm": 0.8337196707725525, "learning_rate": 0.00029820983881216455, "loss": 5.7757, "step": 60000 }, { "epoch": 0.361790137600849, "eval_loss": 6.017859935760498, "eval_runtime": 39.2235, "eval_samples_per_second": 13.053, "eval_steps_per_second": 6.527, "eval_tts_loss": 6.793063328565676, "step": 60000 }, { "epoch": 0.36209162938218303, "grad_norm": 0.9707840085029602, "learning_rate": 0.0002982067030216217, "loss": 5.4221, "step": 60050 }, { "epoch": 0.3623931211635171, "grad_norm": 0.33754396438598633, "learning_rate": 0.00029820356450354594, "loss": 5.3791, "step": 60100 }, { "epoch": 0.3626946129448511, "grad_norm": 1.2815696001052856, "learning_rate": 0.0002982004232579951, "loss": 5.5075, "step": 60150 }, { "epoch": 0.36299610472618515, "grad_norm": 0.9637876152992249, "learning_rate": 0.0002981972792850269, "loss": 5.3709, "step": 60200 }, { "epoch": 0.3632975965075192, "grad_norm": 0.6664658188819885, "learning_rate": 0.00029819413258469935, "loss": 5.479, "step": 60250 }, { "epoch": 0.36359908828885323, "grad_norm": 0.8945205807685852, "learning_rate": 0.0002981909831570703, "loss": 5.6378, "step": 60300 }, { "epoch": 0.3639005800701873, "grad_norm": 0.8341527581214905, "learning_rate": 0.0002981878310021976, "loss": 5.6638, "step": 60350 }, { "epoch": 0.3642020718515213, "grad_norm": 1.1351202726364136, "learning_rate": 0.0002981846761201393, "loss": 5.498, "step": 60400 }, { "epoch": 0.36450356363285535, "grad_norm": 0.7468136548995972, "learning_rate": 0.00029818151851095364, "loss": 5.9795, "step": 60450 }, { "epoch": 0.3648050554141894, "grad_norm": 0.5965693593025208, "learning_rate": 0.0002981783581746985, "loss": 5.3093, "step": 60500 }, { "epoch": 0.36510654719552343, "grad_norm": 0.8381109833717346, "learning_rate": 0.00029817519511143224, "loss": 5.9, "step": 60550 }, { "epoch": 0.3654080389768575, "grad_norm": 1.033879280090332, "learning_rate": 0.0002981720293212129, "loss": 5.7973, "step": 60600 }, { "epoch": 0.3657095307581915, "grad_norm": 0.30049997568130493, "learning_rate": 0.0002981688608040988, "loss": 5.352, "step": 60650 }, { "epoch": 0.36601102253952555, "grad_norm": 1.0027724504470825, "learning_rate": 0.0002981656895601483, "loss": 5.8627, "step": 60700 }, { "epoch": 0.3663125143208596, "grad_norm": 0.8298447728157043, "learning_rate": 0.0002981625155894198, "loss": 5.9376, "step": 60750 }, { "epoch": 0.36661400610219363, "grad_norm": 0.17957951128482819, "learning_rate": 0.0002981593388919715, "loss": 5.6122, "step": 60800 }, { "epoch": 0.3669154978835277, "grad_norm": 1.3753695487976074, "learning_rate": 0.0002981561594678621, "loss": 5.5034, "step": 60850 }, { "epoch": 0.3672169896648617, "grad_norm": 1.0566976070404053, "learning_rate": 0.00029815297731714994, "loss": 5.6097, "step": 60900 }, { "epoch": 0.36751848144619575, "grad_norm": 0.6932763457298279, "learning_rate": 0.0002981497924398937, "loss": 5.7569, "step": 60950 }, { "epoch": 0.3678199732275298, "grad_norm": 0.9134290814399719, "learning_rate": 0.0002981466048361519, "loss": 5.4381, "step": 61000 }, { "epoch": 0.36812146500886384, "grad_norm": 1.3405344486236572, "learning_rate": 0.0002981434145059833, "loss": 5.2635, "step": 61050 }, { "epoch": 0.3684229567901979, "grad_norm": 1.746137022972107, "learning_rate": 0.0002981402214494466, "loss": 5.8374, "step": 61100 }, { "epoch": 0.3687244485715319, "grad_norm": 1.4100160598754883, "learning_rate": 0.0002981370256666005, "loss": 5.548, "step": 61150 }, { "epoch": 0.36902594035286596, "grad_norm": 1.0520262718200684, "learning_rate": 0.0002981338271575038, "loss": 5.8266, "step": 61200 }, { "epoch": 0.3693274321342, "grad_norm": 0.6944697499275208, "learning_rate": 0.00029813062592221545, "loss": 6.0741, "step": 61250 }, { "epoch": 0.36962892391553404, "grad_norm": 0.9421022534370422, "learning_rate": 0.00029812742196079435, "loss": 5.4757, "step": 61300 }, { "epoch": 0.3699304156968681, "grad_norm": 0.7989382743835449, "learning_rate": 0.0002981242152732994, "loss": 5.4283, "step": 61350 }, { "epoch": 0.37023190747820217, "grad_norm": 0.6719452142715454, "learning_rate": 0.0002981210058597897, "loss": 5.8938, "step": 61400 }, { "epoch": 0.3705333992595362, "grad_norm": 0.9602781534194946, "learning_rate": 0.00029811779372032425, "loss": 5.7307, "step": 61450 }, { "epoch": 0.37083489104087025, "grad_norm": 1.016156554222107, "learning_rate": 0.00029811457885496216, "loss": 5.9107, "step": 61500 }, { "epoch": 0.3711363828222043, "grad_norm": 0.9170010685920715, "learning_rate": 0.0002981113612637626, "loss": 5.6456, "step": 61550 }, { "epoch": 0.37143787460353833, "grad_norm": 0.7237151861190796, "learning_rate": 0.0002981081409467848, "loss": 5.7562, "step": 61600 }, { "epoch": 0.37173936638487237, "grad_norm": 0.41403019428253174, "learning_rate": 0.00029810491790408807, "loss": 5.4837, "step": 61650 }, { "epoch": 0.3720408581662064, "grad_norm": 0.2508154511451721, "learning_rate": 0.00029810169213573165, "loss": 5.7797, "step": 61700 }, { "epoch": 0.37234234994754045, "grad_norm": 0.7954480648040771, "learning_rate": 0.000298098463641775, "loss": 4.9349, "step": 61750 }, { "epoch": 0.3726438417288745, "grad_norm": 1.0932607650756836, "learning_rate": 0.0002980952324222774, "loss": 5.2465, "step": 61800 }, { "epoch": 0.37294533351020853, "grad_norm": 0.7345856428146362, "learning_rate": 0.00029809199847729846, "loss": 5.465, "step": 61850 }, { "epoch": 0.37324682529154257, "grad_norm": 0.8163089156150818, "learning_rate": 0.0002980887618068975, "loss": 5.556, "step": 61900 }, { "epoch": 0.3735483170728766, "grad_norm": 0.805067241191864, "learning_rate": 0.0002980855224111344, "loss": 5.4669, "step": 61950 }, { "epoch": 0.37384980885421065, "grad_norm": 0.5516752600669861, "learning_rate": 0.00029808228029006847, "loss": 5.4402, "step": 62000 }, { "epoch": 0.3741513006355447, "grad_norm": 1.1279020309448242, "learning_rate": 0.0002980790354437595, "loss": 5.9357, "step": 62050 }, { "epoch": 0.37445279241687873, "grad_norm": 0.8906154632568359, "learning_rate": 0.00029807578787226723, "loss": 5.572, "step": 62100 }, { "epoch": 0.3747542841982128, "grad_norm": 1.6733239889144897, "learning_rate": 0.00029807253757565134, "loss": 5.283, "step": 62150 }, { "epoch": 0.3750557759795468, "grad_norm": 0.9398669004440308, "learning_rate": 0.0002980692845539718, "loss": 5.0041, "step": 62200 }, { "epoch": 0.37535726776088085, "grad_norm": 0.6108828783035278, "learning_rate": 0.0002980660288072883, "loss": 5.4186, "step": 62250 }, { "epoch": 0.3756587595422149, "grad_norm": 0.7625749707221985, "learning_rate": 0.0002980627703356609, "loss": 5.7783, "step": 62300 }, { "epoch": 0.37596025132354893, "grad_norm": 0.7533839344978333, "learning_rate": 0.0002980595091391495, "loss": 5.4267, "step": 62350 }, { "epoch": 0.376261743104883, "grad_norm": 1.2376712560653687, "learning_rate": 0.0002980562452178141, "loss": 5.5586, "step": 62400 }, { "epoch": 0.376563234886217, "grad_norm": 0.9196678996086121, "learning_rate": 0.0002980529785717148, "loss": 5.632, "step": 62450 }, { "epoch": 0.37686472666755105, "grad_norm": 1.628251314163208, "learning_rate": 0.0002980497092009117, "loss": 5.3634, "step": 62500 }, { "epoch": 0.3771662184488851, "grad_norm": 1.1161786317825317, "learning_rate": 0.00029804643710546505, "loss": 5.3999, "step": 62550 }, { "epoch": 0.37746771023021913, "grad_norm": 0.522364616394043, "learning_rate": 0.00029804316228543493, "loss": 5.3379, "step": 62600 }, { "epoch": 0.3777692020115532, "grad_norm": 0.7702578902244568, "learning_rate": 0.0002980398847408817, "loss": 5.5999, "step": 62650 }, { "epoch": 0.3780706937928872, "grad_norm": 0.7814573645591736, "learning_rate": 0.00029803660447186563, "loss": 5.349, "step": 62700 }, { "epoch": 0.37837218557422125, "grad_norm": 0.64893639087677, "learning_rate": 0.00029803332147844717, "loss": 5.1111, "step": 62750 }, { "epoch": 0.3786736773555553, "grad_norm": 0.4736332893371582, "learning_rate": 0.0002980300357606867, "loss": 5.7343, "step": 62800 }, { "epoch": 0.37897516913688933, "grad_norm": 0.8790179491043091, "learning_rate": 0.0002980267473186446, "loss": 5.8265, "step": 62850 }, { "epoch": 0.3792766609182234, "grad_norm": 0.6373598575592041, "learning_rate": 0.00029802345615238145, "loss": 5.6006, "step": 62900 }, { "epoch": 0.3795781526995574, "grad_norm": 1.0485203266143799, "learning_rate": 0.00029802016226195785, "loss": 5.3447, "step": 62950 }, { "epoch": 0.37987964448089145, "grad_norm": 0.7814056277275085, "learning_rate": 0.00029801686564743444, "loss": 5.4907, "step": 63000 }, { "epoch": 0.37987964448089145, "eval_loss": 6.02476692199707, "eval_runtime": 39.1354, "eval_samples_per_second": 13.083, "eval_steps_per_second": 6.541, "eval_tts_loss": 6.893042390048836, "step": 63000 }, { "epoch": 0.3801811362622255, "grad_norm": 0.911869466304779, "learning_rate": 0.00029801356630887183, "loss": 5.6008, "step": 63050 }, { "epoch": 0.38048262804355953, "grad_norm": 1.0811924934387207, "learning_rate": 0.00029801026424633074, "loss": 5.7972, "step": 63100 }, { "epoch": 0.3807841198248936, "grad_norm": 0.250000536441803, "learning_rate": 0.000298006959459872, "loss": 5.3434, "step": 63150 }, { "epoch": 0.3810856116062276, "grad_norm": 1.667839765548706, "learning_rate": 0.0002980036519495563, "loss": 5.5526, "step": 63200 }, { "epoch": 0.38138710338756165, "grad_norm": 1.0083378553390503, "learning_rate": 0.00029800034171544466, "loss": 5.9368, "step": 63250 }, { "epoch": 0.3816885951688957, "grad_norm": 1.1038482189178467, "learning_rate": 0.00029799702875759796, "loss": 5.17, "step": 63300 }, { "epoch": 0.38199008695022973, "grad_norm": 1.3188165426254272, "learning_rate": 0.0002979937130760771, "loss": 5.6594, "step": 63350 }, { "epoch": 0.3822915787315638, "grad_norm": 0.9305115938186646, "learning_rate": 0.0002979903946709432, "loss": 5.5019, "step": 63400 }, { "epoch": 0.3825930705128978, "grad_norm": 0.8791376948356628, "learning_rate": 0.00029798707354225724, "loss": 5.288, "step": 63450 }, { "epoch": 0.38289456229423185, "grad_norm": 1.0127328634262085, "learning_rate": 0.00029798374969008045, "loss": 5.6935, "step": 63500 }, { "epoch": 0.3831960540755659, "grad_norm": 0.5214490294456482, "learning_rate": 0.00029798042311447384, "loss": 5.8037, "step": 63550 }, { "epoch": 0.38349754585689994, "grad_norm": 0.9296566843986511, "learning_rate": 0.00029797709381549877, "loss": 5.817, "step": 63600 }, { "epoch": 0.383799037638234, "grad_norm": 0.9897987246513367, "learning_rate": 0.00029797376179321644, "loss": 5.7735, "step": 63650 }, { "epoch": 0.384100529419568, "grad_norm": 0.9371635317802429, "learning_rate": 0.0002979704270476882, "loss": 5.7919, "step": 63700 }, { "epoch": 0.38440202120090206, "grad_norm": 0.38590207695961, "learning_rate": 0.0002979670895789754, "loss": 5.5436, "step": 63750 }, { "epoch": 0.3847035129822361, "grad_norm": 1.1541776657104492, "learning_rate": 0.0002979637493871395, "loss": 5.4871, "step": 63800 }, { "epoch": 0.38500500476357014, "grad_norm": 1.2932274341583252, "learning_rate": 0.000297960406472242, "loss": 5.7783, "step": 63850 }, { "epoch": 0.3853064965449042, "grad_norm": 0.9359835386276245, "learning_rate": 0.0002979570608343443, "loss": 5.4105, "step": 63900 }, { "epoch": 0.3856079883262382, "grad_norm": 0.7734107375144958, "learning_rate": 0.00029795371247350803, "loss": 5.5675, "step": 63950 }, { "epoch": 0.38590948010757226, "grad_norm": 0.30938148498535156, "learning_rate": 0.00029795036138979483, "loss": 5.3299, "step": 64000 }, { "epoch": 0.3862109718889063, "grad_norm": 1.2357035875320435, "learning_rate": 0.0002979470075832664, "loss": 5.6367, "step": 64050 }, { "epoch": 0.38651246367024034, "grad_norm": 0.4041103422641754, "learning_rate": 0.00029794365105398444, "loss": 5.6076, "step": 64100 }, { "epoch": 0.3868139554515744, "grad_norm": 0.8205692172050476, "learning_rate": 0.0002979402918020107, "loss": 5.3438, "step": 64150 }, { "epoch": 0.3871154472329084, "grad_norm": 0.8718674182891846, "learning_rate": 0.00029793692982740697, "loss": 5.324, "step": 64200 }, { "epoch": 0.38741693901424246, "grad_norm": 0.672122061252594, "learning_rate": 0.00029793356513023515, "loss": 5.822, "step": 64250 }, { "epoch": 0.3877184307955765, "grad_norm": 0.928280234336853, "learning_rate": 0.0002979301977105572, "loss": 5.3473, "step": 64300 }, { "epoch": 0.38801992257691054, "grad_norm": 0.22104208171367645, "learning_rate": 0.00029792682756843507, "loss": 5.1329, "step": 64350 }, { "epoch": 0.3883214143582446, "grad_norm": 0.8759163618087769, "learning_rate": 0.00029792345470393077, "loss": 5.2858, "step": 64400 }, { "epoch": 0.3886229061395786, "grad_norm": 1.2829617261886597, "learning_rate": 0.0002979200791171065, "loss": 5.6309, "step": 64450 }, { "epoch": 0.38892439792091266, "grad_norm": 1.3073735237121582, "learning_rate": 0.00029791670080802415, "loss": 5.266, "step": 64500 }, { "epoch": 0.3892258897022467, "grad_norm": 1.19869065284729, "learning_rate": 0.00029791331977674606, "loss": 5.485, "step": 64550 }, { "epoch": 0.38952738148358074, "grad_norm": 1.3668086528778076, "learning_rate": 0.0002979099360233343, "loss": 5.4061, "step": 64600 }, { "epoch": 0.3898288732649148, "grad_norm": 0.6387830972671509, "learning_rate": 0.00029790654954785136, "loss": 5.5824, "step": 64650 }, { "epoch": 0.3901303650462488, "grad_norm": 1.0295064449310303, "learning_rate": 0.0002979031603503594, "loss": 5.815, "step": 64700 }, { "epoch": 0.39043185682758286, "grad_norm": 0.9128040671348572, "learning_rate": 0.00029789976843092085, "loss": 5.5946, "step": 64750 }, { "epoch": 0.3907333486089169, "grad_norm": 1.093897819519043, "learning_rate": 0.00029789637378959815, "loss": 5.2976, "step": 64800 }, { "epoch": 0.39103484039025094, "grad_norm": 0.5554790496826172, "learning_rate": 0.0002978929764264537, "loss": 5.7675, "step": 64850 }, { "epoch": 0.391336332171585, "grad_norm": 0.787680983543396, "learning_rate": 0.00029788957634155013, "loss": 5.2183, "step": 64900 }, { "epoch": 0.391637823952919, "grad_norm": 1.223617434501648, "learning_rate": 0.00029788617353494995, "loss": 5.1746, "step": 64950 }, { "epoch": 0.39193931573425306, "grad_norm": 0.8006744384765625, "learning_rate": 0.00029788276800671575, "loss": 5.6298, "step": 65000 }, { "epoch": 0.3922408075155871, "grad_norm": 0.7483025789260864, "learning_rate": 0.00029787935975691027, "loss": 5.3388, "step": 65050 }, { "epoch": 0.3925422992969212, "grad_norm": 1.0675327777862549, "learning_rate": 0.0002978759487855962, "loss": 5.1935, "step": 65100 }, { "epoch": 0.39284379107825523, "grad_norm": 1.2119039297103882, "learning_rate": 0.00029787253509283633, "loss": 5.3893, "step": 65150 }, { "epoch": 0.3931452828595893, "grad_norm": 0.9626854658126831, "learning_rate": 0.0002978691186786935, "loss": 5.4706, "step": 65200 }, { "epoch": 0.3934467746409233, "grad_norm": 0.7634212374687195, "learning_rate": 0.0002978656995432305, "loss": 5.7606, "step": 65250 }, { "epoch": 0.39374826642225735, "grad_norm": 1.4302875995635986, "learning_rate": 0.0002978622776865104, "loss": 5.3176, "step": 65300 }, { "epoch": 0.3940497582035914, "grad_norm": 0.769821047782898, "learning_rate": 0.00029785885310859605, "loss": 5.4166, "step": 65350 }, { "epoch": 0.39435124998492543, "grad_norm": 0.6554358005523682, "learning_rate": 0.0002978554258095505, "loss": 5.2699, "step": 65400 }, { "epoch": 0.3946527417662595, "grad_norm": 1.0819811820983887, "learning_rate": 0.0002978519957894369, "loss": 5.5763, "step": 65450 }, { "epoch": 0.3949542335475935, "grad_norm": 1.6238775253295898, "learning_rate": 0.00029784856304831823, "loss": 5.6121, "step": 65500 }, { "epoch": 0.39525572532892755, "grad_norm": 0.8199433088302612, "learning_rate": 0.0002978451275862577, "loss": 5.5999, "step": 65550 }, { "epoch": 0.3955572171102616, "grad_norm": 1.09218430519104, "learning_rate": 0.00029784168940331874, "loss": 5.1524, "step": 65600 }, { "epoch": 0.39585870889159563, "grad_norm": 0.6436928510665894, "learning_rate": 0.0002978382484995644, "loss": 5.6229, "step": 65650 }, { "epoch": 0.3961602006729297, "grad_norm": 0.9180819988250732, "learning_rate": 0.00029783480487505806, "loss": 5.3647, "step": 65700 }, { "epoch": 0.3964616924542637, "grad_norm": 0.9269749522209167, "learning_rate": 0.0002978313585298631, "loss": 5.7966, "step": 65750 }, { "epoch": 0.39676318423559775, "grad_norm": 0.7323442101478577, "learning_rate": 0.00029782790946404294, "loss": 5.6595, "step": 65800 }, { "epoch": 0.3970646760169318, "grad_norm": 0.8316158056259155, "learning_rate": 0.0002978244576776611, "loss": 5.386, "step": 65850 }, { "epoch": 0.39736616779826583, "grad_norm": 0.7434049844741821, "learning_rate": 0.00029782100317078105, "loss": 5.6, "step": 65900 }, { "epoch": 0.3976676595795999, "grad_norm": 0.480934202671051, "learning_rate": 0.0002978175459434664, "loss": 5.0931, "step": 65950 }, { "epoch": 0.3979691513609339, "grad_norm": 0.9257269501686096, "learning_rate": 0.0002978140859957808, "loss": 5.3244, "step": 66000 }, { "epoch": 0.3979691513609339, "eval_loss": 6.010476589202881, "eval_runtime": 39.1155, "eval_samples_per_second": 13.089, "eval_steps_per_second": 6.545, "eval_tts_loss": 6.865248970393404, "step": 66000 }, { "epoch": 0.39827064314226795, "grad_norm": 0.23291046917438507, "learning_rate": 0.00029781062332778785, "loss": 5.4172, "step": 66050 }, { "epoch": 0.398572134923602, "grad_norm": 0.7996150255203247, "learning_rate": 0.0002978071579395513, "loss": 5.4848, "step": 66100 }, { "epoch": 0.39887362670493604, "grad_norm": 1.4791289567947388, "learning_rate": 0.000297803689831135, "loss": 5.6525, "step": 66150 }, { "epoch": 0.3991751184862701, "grad_norm": 0.823848307132721, "learning_rate": 0.00029780021900260265, "loss": 5.7577, "step": 66200 }, { "epoch": 0.3994766102676041, "grad_norm": 0.8444792628288269, "learning_rate": 0.00029779674545401825, "loss": 5.4455, "step": 66250 }, { "epoch": 0.39977810204893816, "grad_norm": 1.4171346426010132, "learning_rate": 0.0002977932691854456, "loss": 5.8735, "step": 66300 }, { "epoch": 0.4000795938302722, "grad_norm": 0.7297936081886292, "learning_rate": 0.0002977897901969488, "loss": 5.8094, "step": 66350 }, { "epoch": 0.40038108561160624, "grad_norm": 1.0235059261322021, "learning_rate": 0.0002977863084885918, "loss": 5.7926, "step": 66400 }, { "epoch": 0.4006825773929403, "grad_norm": 1.3886281251907349, "learning_rate": 0.00029778282406043866, "loss": 5.6374, "step": 66450 }, { "epoch": 0.4009840691742743, "grad_norm": 1.271442174911499, "learning_rate": 0.0002977793369125536, "loss": 5.3015, "step": 66500 }, { "epoch": 0.40128556095560836, "grad_norm": 0.9120314717292786, "learning_rate": 0.00029777584704500067, "loss": 5.4818, "step": 66550 }, { "epoch": 0.4015870527369424, "grad_norm": 0.9560785293579102, "learning_rate": 0.00029777235445784423, "loss": 5.5742, "step": 66600 }, { "epoch": 0.40188854451827644, "grad_norm": 0.7142675518989563, "learning_rate": 0.00029776885915114846, "loss": 5.4446, "step": 66650 }, { "epoch": 0.4021900362996105, "grad_norm": 1.0035794973373413, "learning_rate": 0.0002977653611249777, "loss": 5.0701, "step": 66700 }, { "epoch": 0.4024915280809445, "grad_norm": 0.39110270142555237, "learning_rate": 0.0002977618603793964, "loss": 5.3156, "step": 66750 }, { "epoch": 0.40279301986227856, "grad_norm": 0.7268269062042236, "learning_rate": 0.00029775835691446886, "loss": 5.7929, "step": 66800 }, { "epoch": 0.4030945116436126, "grad_norm": 0.8916475176811218, "learning_rate": 0.00029775485073025963, "loss": 5.6623, "step": 66850 }, { "epoch": 0.40339600342494664, "grad_norm": 0.897605299949646, "learning_rate": 0.0002977513418268333, "loss": 5.8317, "step": 66900 }, { "epoch": 0.4036974952062807, "grad_norm": 0.9518166780471802, "learning_rate": 0.00029774783020425425, "loss": 5.254, "step": 66950 }, { "epoch": 0.4039989869876147, "grad_norm": 0.7701130509376526, "learning_rate": 0.00029774431586258735, "loss": 5.4684, "step": 67000 }, { "epoch": 0.40430047876894876, "grad_norm": 1.4107767343521118, "learning_rate": 0.00029774079880189714, "loss": 5.5162, "step": 67050 }, { "epoch": 0.4046019705502828, "grad_norm": 0.9920675754547119, "learning_rate": 0.0002977372790222483, "loss": 5.3005, "step": 67100 }, { "epoch": 0.40490346233161684, "grad_norm": 1.0957945585250854, "learning_rate": 0.0002977337565237057, "loss": 5.2377, "step": 67150 }, { "epoch": 0.4052049541129509, "grad_norm": 0.6903366446495056, "learning_rate": 0.0002977302313063342, "loss": 5.6077, "step": 67200 }, { "epoch": 0.4055064458942849, "grad_norm": 1.0206553936004639, "learning_rate": 0.00029772670337019857, "loss": 5.7085, "step": 67250 }, { "epoch": 0.40580793767561896, "grad_norm": 0.8907656669616699, "learning_rate": 0.0002977231727153638, "loss": 5.7381, "step": 67300 }, { "epoch": 0.406109429456953, "grad_norm": 1.3018453121185303, "learning_rate": 0.0002977196393418948, "loss": 5.6865, "step": 67350 }, { "epoch": 0.40641092123828704, "grad_norm": 0.9241821765899658, "learning_rate": 0.00029771610324985664, "loss": 5.4226, "step": 67400 }, { "epoch": 0.4067124130196211, "grad_norm": 0.857705295085907, "learning_rate": 0.00029771256443931446, "loss": 5.7349, "step": 67450 }, { "epoch": 0.4070139048009551, "grad_norm": 0.7284466624259949, "learning_rate": 0.0002977090229103333, "loss": 5.3926, "step": 67500 }, { "epoch": 0.40731539658228916, "grad_norm": 0.9253411889076233, "learning_rate": 0.0002977054786629784, "loss": 5.8913, "step": 67550 }, { "epoch": 0.4076168883636232, "grad_norm": 1.1662101745605469, "learning_rate": 0.0002977019316973149, "loss": 5.5222, "step": 67600 }, { "epoch": 0.40791838014495724, "grad_norm": 1.0271962881088257, "learning_rate": 0.0002976983820134081, "loss": 5.6128, "step": 67650 }, { "epoch": 0.4082198719262913, "grad_norm": 0.2857980728149414, "learning_rate": 0.0002976948296113234, "loss": 5.1875, "step": 67700 }, { "epoch": 0.4085213637076253, "grad_norm": 0.9315298199653625, "learning_rate": 0.0002976912744911262, "loss": 4.994, "step": 67750 }, { "epoch": 0.40882285548895936, "grad_norm": 1.315632700920105, "learning_rate": 0.00029768771665288177, "loss": 5.9333, "step": 67800 }, { "epoch": 0.4091243472702934, "grad_norm": 1.0289833545684814, "learning_rate": 0.0002976841560966557, "loss": 5.0881, "step": 67850 }, { "epoch": 0.40942583905162744, "grad_norm": 1.613336205482483, "learning_rate": 0.0002976805928225135, "loss": 5.3447, "step": 67900 }, { "epoch": 0.4097273308329615, "grad_norm": 0.642764151096344, "learning_rate": 0.00029767702683052076, "loss": 5.6044, "step": 67950 }, { "epoch": 0.4100288226142955, "grad_norm": 0.43819692730903625, "learning_rate": 0.00029767345812074304, "loss": 5.0533, "step": 68000 }, { "epoch": 0.41033031439562956, "grad_norm": 0.5635634064674377, "learning_rate": 0.00029766988669324614, "loss": 5.5131, "step": 68050 }, { "epoch": 0.4106318061769636, "grad_norm": 1.5713058710098267, "learning_rate": 0.0002976663125480956, "loss": 5.5012, "step": 68100 }, { "epoch": 0.41093329795829764, "grad_norm": 0.729307234287262, "learning_rate": 0.00029766273568535744, "loss": 5.2243, "step": 68150 }, { "epoch": 0.4112347897396317, "grad_norm": 1.3683489561080933, "learning_rate": 0.00029765915610509725, "loss": 5.1736, "step": 68200 }, { "epoch": 0.4115362815209657, "grad_norm": 0.7494198679924011, "learning_rate": 0.00029765557380738103, "loss": 5.254, "step": 68250 }, { "epoch": 0.41183777330229976, "grad_norm": 0.7104106545448303, "learning_rate": 0.00029765198879227474, "loss": 5.1302, "step": 68300 }, { "epoch": 0.4121392650836338, "grad_norm": 0.8988791108131409, "learning_rate": 0.00029764840105984426, "loss": 5.7474, "step": 68350 }, { "epoch": 0.41244075686496784, "grad_norm": 1.1162325143814087, "learning_rate": 0.00029764481061015567, "loss": 5.7374, "step": 68400 }, { "epoch": 0.4127422486463019, "grad_norm": 0.9623829126358032, "learning_rate": 0.00029764121744327505, "loss": 5.6633, "step": 68450 }, { "epoch": 0.4130437404276359, "grad_norm": 1.799211025238037, "learning_rate": 0.00029763762155926853, "loss": 5.9531, "step": 68500 }, { "epoch": 0.41334523220896996, "grad_norm": 1.1162340641021729, "learning_rate": 0.00029763402295820225, "loss": 5.8485, "step": 68550 }, { "epoch": 0.413646723990304, "grad_norm": 1.1240646839141846, "learning_rate": 0.0002976304216401425, "loss": 5.4343, "step": 68600 }, { "epoch": 0.41394821577163804, "grad_norm": 1.1420578956604004, "learning_rate": 0.0002976268176051555, "loss": 5.4248, "step": 68650 }, { "epoch": 0.4142497075529721, "grad_norm": 0.8938031196594238, "learning_rate": 0.0002976232108533076, "loss": 5.659, "step": 68700 }, { "epoch": 0.4145511993343061, "grad_norm": 0.7405080795288086, "learning_rate": 0.00029761960138466515, "loss": 5.1819, "step": 68750 }, { "epoch": 0.4148526911156402, "grad_norm": 1.1816387176513672, "learning_rate": 0.0002976159891992946, "loss": 5.3496, "step": 68800 }, { "epoch": 0.41515418289697426, "grad_norm": 0.7501606941223145, "learning_rate": 0.0002976123742972625, "loss": 5.7916, "step": 68850 }, { "epoch": 0.4154556746783083, "grad_norm": 1.2785476446151733, "learning_rate": 0.00029760875667863526, "loss": 5.6278, "step": 68900 }, { "epoch": 0.41575716645964234, "grad_norm": 0.5768195986747742, "learning_rate": 0.00029760513634347947, "loss": 4.9024, "step": 68950 }, { "epoch": 0.4160586582409764, "grad_norm": 0.806246817111969, "learning_rate": 0.0002976015132918618, "loss": 5.4994, "step": 69000 }, { "epoch": 0.4160586582409764, "eval_loss": 6.00975227355957, "eval_runtime": 39.3157, "eval_samples_per_second": 13.023, "eval_steps_per_second": 6.511, "eval_tts_loss": 6.858713644475784, "step": 69000 }, { "epoch": 0.4163601500223104, "grad_norm": 0.749523937702179, "learning_rate": 0.000297597887523849, "loss": 5.35, "step": 69050 }, { "epoch": 0.41666164180364446, "grad_norm": 0.4866684675216675, "learning_rate": 0.0002975942590395076, "loss": 5.5269, "step": 69100 }, { "epoch": 0.4169631335849785, "grad_norm": 1.2285975217819214, "learning_rate": 0.0002975906278389045, "loss": 5.1189, "step": 69150 }, { "epoch": 0.41726462536631254, "grad_norm": 1.234284520149231, "learning_rate": 0.0002975869939221066, "loss": 5.8034, "step": 69200 }, { "epoch": 0.4175661171476466, "grad_norm": 1.578583002090454, "learning_rate": 0.00029758335728918064, "loss": 5.7351, "step": 69250 }, { "epoch": 0.4178676089289806, "grad_norm": 0.981745183467865, "learning_rate": 0.0002975797179401936, "loss": 5.7316, "step": 69300 }, { "epoch": 0.41816910071031466, "grad_norm": 0.9798596501350403, "learning_rate": 0.00029757607587521244, "loss": 5.2496, "step": 69350 }, { "epoch": 0.4184705924916487, "grad_norm": 0.7315351963043213, "learning_rate": 0.00029757243109430426, "loss": 5.9435, "step": 69400 }, { "epoch": 0.41877208427298274, "grad_norm": 0.42303940653800964, "learning_rate": 0.00029756878359753604, "loss": 5.0776, "step": 69450 }, { "epoch": 0.4190735760543168, "grad_norm": 1.4397201538085938, "learning_rate": 0.000297565133384975, "loss": 5.6613, "step": 69500 }, { "epoch": 0.4193750678356508, "grad_norm": 1.3348782062530518, "learning_rate": 0.0002975614804566882, "loss": 5.6771, "step": 69550 }, { "epoch": 0.41967655961698486, "grad_norm": 1.425757646560669, "learning_rate": 0.000297557824812743, "loss": 5.3147, "step": 69600 }, { "epoch": 0.4199780513983189, "grad_norm": 0.9898025393486023, "learning_rate": 0.00029755416645320653, "loss": 5.8196, "step": 69650 }, { "epoch": 0.42027954317965294, "grad_norm": 0.2799665927886963, "learning_rate": 0.0002975505053781463, "loss": 5.2293, "step": 69700 }, { "epoch": 0.420581034960987, "grad_norm": 1.15394926071167, "learning_rate": 0.00029754684158762954, "loss": 5.8445, "step": 69750 }, { "epoch": 0.420882526742321, "grad_norm": 1.1765013933181763, "learning_rate": 0.0002975431750817237, "loss": 5.33, "step": 69800 }, { "epoch": 0.42118401852365506, "grad_norm": 0.6815205812454224, "learning_rate": 0.00029753950586049627, "loss": 5.5117, "step": 69850 }, { "epoch": 0.4214855103049891, "grad_norm": 0.71918785572052, "learning_rate": 0.0002975358339240148, "loss": 5.288, "step": 69900 }, { "epoch": 0.42178700208632314, "grad_norm": 1.0456948280334473, "learning_rate": 0.00029753215927234687, "loss": 5.6179, "step": 69950 }, { "epoch": 0.4220884938676572, "grad_norm": 0.7725948691368103, "learning_rate": 0.0002975284819055601, "loss": 5.2773, "step": 70000 }, { "epoch": 0.4223899856489912, "grad_norm": 1.294581413269043, "learning_rate": 0.0002975248018237221, "loss": 5.1507, "step": 70050 }, { "epoch": 0.42269147743032526, "grad_norm": 1.5923742055892944, "learning_rate": 0.00029752111902690067, "loss": 5.541, "step": 70100 }, { "epoch": 0.4229929692116593, "grad_norm": 1.1530650854110718, "learning_rate": 0.0002975174335151636, "loss": 5.3326, "step": 70150 }, { "epoch": 0.42329446099299334, "grad_norm": 0.7758811712265015, "learning_rate": 0.00029751374528857864, "loss": 5.322, "step": 70200 }, { "epoch": 0.4235959527743274, "grad_norm": 0.42408764362335205, "learning_rate": 0.00029751005434721376, "loss": 5.7015, "step": 70250 }, { "epoch": 0.4238974445556614, "grad_norm": 0.571323573589325, "learning_rate": 0.0002975063606911368, "loss": 5.4004, "step": 70300 }, { "epoch": 0.42419893633699546, "grad_norm": 2.329010248184204, "learning_rate": 0.0002975026643204158, "loss": 5.2128, "step": 70350 }, { "epoch": 0.4245004281183295, "grad_norm": 0.9724481701850891, "learning_rate": 0.0002974989652351187, "loss": 5.5423, "step": 70400 }, { "epoch": 0.42480191989966354, "grad_norm": 1.012682557106018, "learning_rate": 0.0002974952634353137, "loss": 5.6206, "step": 70450 }, { "epoch": 0.4251034116809976, "grad_norm": 1.0905206203460693, "learning_rate": 0.00029749155892106887, "loss": 5.4382, "step": 70500 }, { "epoch": 0.4254049034623316, "grad_norm": 1.3622685670852661, "learning_rate": 0.0002974878516924524, "loss": 5.2779, "step": 70550 }, { "epoch": 0.42570639524366566, "grad_norm": 0.9513403177261353, "learning_rate": 0.0002974841417495324, "loss": 5.4852, "step": 70600 }, { "epoch": 0.4260078870249997, "grad_norm": 0.7947118878364563, "learning_rate": 0.00029748042909237733, "loss": 5.6708, "step": 70650 }, { "epoch": 0.42630937880633374, "grad_norm": 0.36032283306121826, "learning_rate": 0.00029747671372105537, "loss": 5.6564, "step": 70700 }, { "epoch": 0.4266108705876678, "grad_norm": 0.6721035242080688, "learning_rate": 0.000297472995635635, "loss": 5.8378, "step": 70750 }, { "epoch": 0.4269123623690018, "grad_norm": 0.8551837205886841, "learning_rate": 0.0002974692748361846, "loss": 5.5061, "step": 70800 }, { "epoch": 0.42721385415033586, "grad_norm": 0.8279239535331726, "learning_rate": 0.0002974655513227726, "loss": 5.4771, "step": 70850 }, { "epoch": 0.4275153459316699, "grad_norm": 0.8888903260231018, "learning_rate": 0.0002974618250954676, "loss": 4.8803, "step": 70900 }, { "epoch": 0.42781683771300394, "grad_norm": 0.8158321976661682, "learning_rate": 0.00029745809615433814, "loss": 5.4258, "step": 70950 }, { "epoch": 0.428118329494338, "grad_norm": 1.3889409303665161, "learning_rate": 0.00029745436449945287, "loss": 5.4864, "step": 71000 }, { "epoch": 0.428419821275672, "grad_norm": 1.0143166780471802, "learning_rate": 0.00029745063013088046, "loss": 5.6985, "step": 71050 }, { "epoch": 0.42872131305700606, "grad_norm": 0.7676172852516174, "learning_rate": 0.0002974468930486896, "loss": 5.3749, "step": 71100 }, { "epoch": 0.4290228048383401, "grad_norm": 1.1923211812973022, "learning_rate": 0.00029744315325294906, "loss": 5.2937, "step": 71150 }, { "epoch": 0.42932429661967414, "grad_norm": 1.139027714729309, "learning_rate": 0.0002974394107437278, "loss": 5.5542, "step": 71200 }, { "epoch": 0.4296257884010082, "grad_norm": 1.287854552268982, "learning_rate": 0.00029743566552109453, "loss": 5.3818, "step": 71250 }, { "epoch": 0.4299272801823422, "grad_norm": 0.808631181716919, "learning_rate": 0.0002974319175851182, "loss": 5.8895, "step": 71300 }, { "epoch": 0.43022877196367626, "grad_norm": 0.6962984204292297, "learning_rate": 0.0002974281669358679, "loss": 5.379, "step": 71350 }, { "epoch": 0.4305302637450103, "grad_norm": 1.2034885883331299, "learning_rate": 0.0002974244135734125, "loss": 5.8101, "step": 71400 }, { "epoch": 0.43083175552634434, "grad_norm": 1.8343721628189087, "learning_rate": 0.0002974206574978212, "loss": 5.851, "step": 71450 }, { "epoch": 0.4311332473076784, "grad_norm": 1.006738305091858, "learning_rate": 0.0002974168987091631, "loss": 5.2371, "step": 71500 }, { "epoch": 0.4314347390890124, "grad_norm": 0.49184292554855347, "learning_rate": 0.0002974131372075073, "loss": 5.6934, "step": 71550 }, { "epoch": 0.43173623087034646, "grad_norm": 1.5441361665725708, "learning_rate": 0.0002974093729929231, "loss": 4.9822, "step": 71600 }, { "epoch": 0.4320377226516805, "grad_norm": 0.8337753415107727, "learning_rate": 0.00029740560606547975, "loss": 5.2545, "step": 71650 }, { "epoch": 0.43233921443301454, "grad_norm": 0.749809741973877, "learning_rate": 0.0002974018364252466, "loss": 5.2622, "step": 71700 }, { "epoch": 0.4326407062143486, "grad_norm": 0.30531632900238037, "learning_rate": 0.00029739806407229303, "loss": 5.7322, "step": 71750 }, { "epoch": 0.4329421979956826, "grad_norm": 0.8639238476753235, "learning_rate": 0.0002973942890066884, "loss": 5.8946, "step": 71800 }, { "epoch": 0.43324368977701666, "grad_norm": 1.2139533758163452, "learning_rate": 0.0002973905112285022, "loss": 5.9051, "step": 71850 }, { "epoch": 0.4335451815583507, "grad_norm": 1.0293803215026855, "learning_rate": 0.00029738673073780405, "loss": 5.255, "step": 71900 }, { "epoch": 0.43384667333968474, "grad_norm": 0.40490052103996277, "learning_rate": 0.0002973829475346634, "loss": 5.7204, "step": 71950 }, { "epoch": 0.4341481651210188, "grad_norm": 0.9136404395103455, "learning_rate": 0.00029737916161914993, "loss": 5.2993, "step": 72000 }, { "epoch": 0.4341481651210188, "eval_loss": 5.992713928222656, "eval_runtime": 39.3197, "eval_samples_per_second": 13.021, "eval_steps_per_second": 6.511, "eval_tts_loss": 6.8454064810082, "step": 72000 }, { "epoch": 0.4344496569023528, "grad_norm": 1.787865161895752, "learning_rate": 0.0002973753729913333, "loss": 5.5706, "step": 72050 }, { "epoch": 0.43475114868368686, "grad_norm": 0.43244561553001404, "learning_rate": 0.0002973715816512833, "loss": 5.8754, "step": 72100 }, { "epoch": 0.4350526404650209, "grad_norm": 0.9285842180252075, "learning_rate": 0.0002973677875990696, "loss": 5.5619, "step": 72150 }, { "epoch": 0.43535413224635494, "grad_norm": 0.6384845972061157, "learning_rate": 0.0002973639908347621, "loss": 5.615, "step": 72200 }, { "epoch": 0.435655624027689, "grad_norm": 0.755869448184967, "learning_rate": 0.0002973601913584306, "loss": 5.5988, "step": 72250 }, { "epoch": 0.435957115809023, "grad_norm": 1.2240440845489502, "learning_rate": 0.00029735638917014514, "loss": 5.6989, "step": 72300 }, { "epoch": 0.43625860759035706, "grad_norm": 0.9993360638618469, "learning_rate": 0.00029735258426997563, "loss": 5.4075, "step": 72350 }, { "epoch": 0.4365600993716911, "grad_norm": 0.8361043334007263, "learning_rate": 0.00029734877665799206, "loss": 5.8824, "step": 72400 }, { "epoch": 0.43686159115302514, "grad_norm": 0.7827766537666321, "learning_rate": 0.0002973449663342645, "loss": 5.5851, "step": 72450 }, { "epoch": 0.4371630829343592, "grad_norm": 0.9817952513694763, "learning_rate": 0.00029734115329886314, "loss": 4.9979, "step": 72500 }, { "epoch": 0.4374645747156933, "grad_norm": 0.9510376453399658, "learning_rate": 0.0002973373375518581, "loss": 6.067, "step": 72550 }, { "epoch": 0.4377660664970273, "grad_norm": 0.5686951279640198, "learning_rate": 0.00029733351909331965, "loss": 5.1169, "step": 72600 }, { "epoch": 0.43806755827836136, "grad_norm": 0.9538442492485046, "learning_rate": 0.00029732969792331803, "loss": 5.7011, "step": 72650 }, { "epoch": 0.4383690500596954, "grad_norm": 1.073512315750122, "learning_rate": 0.00029732587404192356, "loss": 5.5429, "step": 72700 }, { "epoch": 0.43867054184102944, "grad_norm": 1.0051424503326416, "learning_rate": 0.00029732204744920666, "loss": 5.4875, "step": 72750 }, { "epoch": 0.4389720336223635, "grad_norm": 0.6300674676895142, "learning_rate": 0.0002973182181452377, "loss": 5.3616, "step": 72800 }, { "epoch": 0.4392735254036975, "grad_norm": 1.1184922456741333, "learning_rate": 0.0002973143861300871, "loss": 5.2937, "step": 72850 }, { "epoch": 0.43957501718503156, "grad_norm": 1.1815348863601685, "learning_rate": 0.00029731055140382555, "loss": 5.5404, "step": 72900 }, { "epoch": 0.4398765089663656, "grad_norm": 1.2967262268066406, "learning_rate": 0.0002973067139665235, "loss": 5.4369, "step": 72950 }, { "epoch": 0.44017800074769964, "grad_norm": 0.6458814144134521, "learning_rate": 0.0002973028738182516, "loss": 5.7342, "step": 73000 }, { "epoch": 0.4404794925290337, "grad_norm": 0.7760760188102722, "learning_rate": 0.00029729903095908046, "loss": 5.7052, "step": 73050 }, { "epoch": 0.4407809843103677, "grad_norm": 0.7565313577651978, "learning_rate": 0.0002972951853890809, "loss": 5.583, "step": 73100 }, { "epoch": 0.44108247609170176, "grad_norm": 1.1758830547332764, "learning_rate": 0.00029729133710832366, "loss": 5.6909, "step": 73150 }, { "epoch": 0.4413839678730358, "grad_norm": 0.3073970675468445, "learning_rate": 0.0002972874861168796, "loss": 4.945, "step": 73200 }, { "epoch": 0.44168545965436984, "grad_norm": 1.2094026803970337, "learning_rate": 0.0002972836324148195, "loss": 5.0704, "step": 73250 }, { "epoch": 0.4419869514357039, "grad_norm": 1.0833210945129395, "learning_rate": 0.00029727977600221427, "loss": 5.488, "step": 73300 }, { "epoch": 0.4422884432170379, "grad_norm": 0.9758421778678894, "learning_rate": 0.00029727591687913506, "loss": 5.2963, "step": 73350 }, { "epoch": 0.44258993499837196, "grad_norm": 0.981188178062439, "learning_rate": 0.0002972720550456527, "loss": 5.6541, "step": 73400 }, { "epoch": 0.442891426779706, "grad_norm": 1.456020712852478, "learning_rate": 0.00029726819050183837, "loss": 5.6694, "step": 73450 }, { "epoch": 0.44319291856104004, "grad_norm": 0.7926260828971863, "learning_rate": 0.00029726432324776313, "loss": 5.2105, "step": 73500 }, { "epoch": 0.4434944103423741, "grad_norm": 1.173313021659851, "learning_rate": 0.00029726045328349823, "loss": 5.6488, "step": 73550 }, { "epoch": 0.4437959021237081, "grad_norm": 1.407318115234375, "learning_rate": 0.00029725658060911474, "loss": 5.6195, "step": 73600 }, { "epoch": 0.44409739390504216, "grad_norm": 0.7395715117454529, "learning_rate": 0.0002972527052246841, "loss": 5.838, "step": 73650 }, { "epoch": 0.4443988856863762, "grad_norm": 1.0592786073684692, "learning_rate": 0.0002972488271302776, "loss": 5.4067, "step": 73700 }, { "epoch": 0.44470037746771024, "grad_norm": 1.5510344505310059, "learning_rate": 0.0002972449463259665, "loss": 5.5492, "step": 73750 }, { "epoch": 0.4450018692490443, "grad_norm": 1.4226680994033813, "learning_rate": 0.0002972410628118223, "loss": 5.7074, "step": 73800 }, { "epoch": 0.4453033610303783, "grad_norm": 1.5343610048294067, "learning_rate": 0.0002972371765879165, "loss": 5.6287, "step": 73850 }, { "epoch": 0.44560485281171236, "grad_norm": 1.6840111017227173, "learning_rate": 0.0002972332876543206, "loss": 5.1418, "step": 73900 }, { "epoch": 0.4459063445930464, "grad_norm": 1.152022361755371, "learning_rate": 0.0002972293960111061, "loss": 5.8877, "step": 73950 }, { "epoch": 0.44620783637438044, "grad_norm": 1.1323624849319458, "learning_rate": 0.00029722550165834467, "loss": 5.5542, "step": 74000 }, { "epoch": 0.4465093281557145, "grad_norm": 1.2778420448303223, "learning_rate": 0.000297221604596108, "loss": 5.1368, "step": 74050 }, { "epoch": 0.4468108199370485, "grad_norm": 1.0781426429748535, "learning_rate": 0.00029721770482446784, "loss": 5.6995, "step": 74100 }, { "epoch": 0.44711231171838256, "grad_norm": 0.7402617335319519, "learning_rate": 0.0002972138023434959, "loss": 5.2228, "step": 74150 }, { "epoch": 0.4474138034997166, "grad_norm": 0.7835879325866699, "learning_rate": 0.00029720989715326394, "loss": 5.2223, "step": 74200 }, { "epoch": 0.44771529528105064, "grad_norm": 0.2585538923740387, "learning_rate": 0.000297205989253844, "loss": 5.7597, "step": 74250 }, { "epoch": 0.4480167870623847, "grad_norm": 1.2460399866104126, "learning_rate": 0.0002972020786453078, "loss": 5.048, "step": 74300 }, { "epoch": 0.4483182788437187, "grad_norm": 1.028441071510315, "learning_rate": 0.00029719816532772746, "loss": 5.7667, "step": 74350 }, { "epoch": 0.44861977062505276, "grad_norm": 1.0867533683776855, "learning_rate": 0.00029719424930117495, "loss": 5.4132, "step": 74400 }, { "epoch": 0.4489212624063868, "grad_norm": 1.9760404825210571, "learning_rate": 0.00029719033056572233, "loss": 5.8173, "step": 74450 }, { "epoch": 0.44922275418772084, "grad_norm": 0.9913554787635803, "learning_rate": 0.00029718640912144176, "loss": 5.4435, "step": 74500 }, { "epoch": 0.4495242459690549, "grad_norm": 0.4308446943759918, "learning_rate": 0.0002971824849684054, "loss": 5.3899, "step": 74550 }, { "epoch": 0.4498257377503889, "grad_norm": 1.099312663078308, "learning_rate": 0.00029717855810668537, "loss": 5.6217, "step": 74600 }, { "epoch": 0.45012722953172296, "grad_norm": 0.7306317687034607, "learning_rate": 0.000297174628536354, "loss": 5.6137, "step": 74650 }, { "epoch": 0.450428721313057, "grad_norm": 0.9609135985374451, "learning_rate": 0.00029717069625748366, "loss": 5.9911, "step": 74700 }, { "epoch": 0.45073021309439104, "grad_norm": 1.0635172128677368, "learning_rate": 0.0002971667612701467, "loss": 5.25, "step": 74750 }, { "epoch": 0.4510317048757251, "grad_norm": 0.8892016410827637, "learning_rate": 0.00029716282357441555, "loss": 5.5146, "step": 74800 }, { "epoch": 0.4513331966570591, "grad_norm": 1.0106509923934937, "learning_rate": 0.0002971588831703626, "loss": 5.3632, "step": 74850 }, { "epoch": 0.45163468843839316, "grad_norm": 0.9105972647666931, "learning_rate": 0.00029715494005806036, "loss": 5.1619, "step": 74900 }, { "epoch": 0.4519361802197272, "grad_norm": 1.2586970329284668, "learning_rate": 0.00029715099423758157, "loss": 5.503, "step": 74950 }, { "epoch": 0.45223767200106124, "grad_norm": 1.3249962329864502, "learning_rate": 0.00029714704570899864, "loss": 6.012, "step": 75000 }, { "epoch": 0.45223767200106124, "eval_loss": 5.968864440917969, "eval_runtime": 39.0811, "eval_samples_per_second": 13.101, "eval_steps_per_second": 6.55, "eval_tts_loss": 6.941050806793469, "step": 75000 }, { "epoch": 0.4525391637823953, "grad_norm": 0.44307783246040344, "learning_rate": 0.00029714309447238427, "loss": 5.2195, "step": 75050 }, { "epoch": 0.4528406555637293, "grad_norm": 0.7589360475540161, "learning_rate": 0.0002971391405278113, "loss": 5.4055, "step": 75100 }, { "epoch": 0.45314214734506336, "grad_norm": 1.1733956336975098, "learning_rate": 0.00029713518387535246, "loss": 5.7853, "step": 75150 }, { "epoch": 0.4534436391263974, "grad_norm": 0.9368916749954224, "learning_rate": 0.0002971312245150805, "loss": 5.4649, "step": 75200 }, { "epoch": 0.45374513090773144, "grad_norm": 1.5307005643844604, "learning_rate": 0.00029712726244706837, "loss": 5.656, "step": 75250 }, { "epoch": 0.4540466226890655, "grad_norm": 0.31537938117980957, "learning_rate": 0.00029712329767138883, "loss": 5.5216, "step": 75300 }, { "epoch": 0.4543481144703995, "grad_norm": 1.4584078788757324, "learning_rate": 0.000297119330188115, "loss": 5.6116, "step": 75350 }, { "epoch": 0.45464960625173356, "grad_norm": 0.4466046988964081, "learning_rate": 0.00029711535999731985, "loss": 5.63, "step": 75400 }, { "epoch": 0.4549510980330676, "grad_norm": 0.636944055557251, "learning_rate": 0.0002971113870990765, "loss": 5.4303, "step": 75450 }, { "epoch": 0.45525258981440164, "grad_norm": 0.590165913105011, "learning_rate": 0.000297107411493458, "loss": 5.1837, "step": 75500 }, { "epoch": 0.4555540815957357, "grad_norm": 0.8413363099098206, "learning_rate": 0.00029710343318053745, "loss": 5.1617, "step": 75550 }, { "epoch": 0.4558555733770697, "grad_norm": 0.8955807685852051, "learning_rate": 0.0002970994521603882, "loss": 5.5238, "step": 75600 }, { "epoch": 0.45615706515840376, "grad_norm": 1.3330796957015991, "learning_rate": 0.00029709546843308343, "loss": 5.4364, "step": 75650 }, { "epoch": 0.4564585569397378, "grad_norm": 0.49501532316207886, "learning_rate": 0.00029709148199869647, "loss": 5.6891, "step": 75700 }, { "epoch": 0.45676004872107184, "grad_norm": 1.500314474105835, "learning_rate": 0.00029708749285730073, "loss": 5.5593, "step": 75750 }, { "epoch": 0.4570615405024059, "grad_norm": 0.3600963354110718, "learning_rate": 0.00029708350100896957, "loss": 5.3136, "step": 75800 }, { "epoch": 0.4573630322837399, "grad_norm": 0.7682241797447205, "learning_rate": 0.00029707950645377644, "loss": 5.6452, "step": 75850 }, { "epoch": 0.45766452406507396, "grad_norm": 0.838597297668457, "learning_rate": 0.0002970755091917949, "loss": 5.4289, "step": 75900 }, { "epoch": 0.457966015846408, "grad_norm": 1.4342913627624512, "learning_rate": 0.0002970715092230985, "loss": 5.5103, "step": 75950 }, { "epoch": 0.45826750762774204, "grad_norm": 0.6998138427734375, "learning_rate": 0.0002970675065477609, "loss": 5.5825, "step": 76000 }, { "epoch": 0.4585689994090761, "grad_norm": 0.8695118427276611, "learning_rate": 0.00029706350116585575, "loss": 4.9472, "step": 76050 }, { "epoch": 0.4588704911904101, "grad_norm": 1.4185765981674194, "learning_rate": 0.00029705949307745665, "loss": 5.8181, "step": 76100 }, { "epoch": 0.45917198297174416, "grad_norm": 0.8473349213600159, "learning_rate": 0.00029705548228263745, "loss": 5.4667, "step": 76150 }, { "epoch": 0.4594734747530782, "grad_norm": 0.6940177083015442, "learning_rate": 0.000297051468781472, "loss": 5.3366, "step": 76200 }, { "epoch": 0.4597749665344123, "grad_norm": 1.4103384017944336, "learning_rate": 0.00029704745257403407, "loss": 5.4649, "step": 76250 }, { "epoch": 0.46007645831574634, "grad_norm": 0.4873877763748169, "learning_rate": 0.00029704343366039767, "loss": 5.3244, "step": 76300 }, { "epoch": 0.4603779500970804, "grad_norm": 1.2304246425628662, "learning_rate": 0.0002970394120406367, "loss": 5.5378, "step": 76350 }, { "epoch": 0.4606794418784144, "grad_norm": 0.5246436595916748, "learning_rate": 0.00029703538771482516, "loss": 5.4069, "step": 76400 }, { "epoch": 0.46098093365974846, "grad_norm": 0.30000126361846924, "learning_rate": 0.00029703136068303716, "loss": 5.438, "step": 76450 }, { "epoch": 0.4612824254410825, "grad_norm": 0.6377673745155334, "learning_rate": 0.0002970273309453468, "loss": 5.4703, "step": 76500 }, { "epoch": 0.46158391722241654, "grad_norm": 1.3181360960006714, "learning_rate": 0.00029702329850182823, "loss": 5.4122, "step": 76550 }, { "epoch": 0.4618854090037506, "grad_norm": 0.418816477060318, "learning_rate": 0.0002970192633525557, "loss": 5.8162, "step": 76600 }, { "epoch": 0.4621869007850846, "grad_norm": 1.062147617340088, "learning_rate": 0.00029701522549760337, "loss": 5.1683, "step": 76650 }, { "epoch": 0.46248839256641866, "grad_norm": 1.3457822799682617, "learning_rate": 0.0002970111849370457, "loss": 5.5988, "step": 76700 }, { "epoch": 0.4627898843477527, "grad_norm": 1.154507040977478, "learning_rate": 0.0002970071416709569, "loss": 5.4996, "step": 76750 }, { "epoch": 0.46309137612908674, "grad_norm": 0.7483363151550293, "learning_rate": 0.00029700309569941145, "loss": 5.448, "step": 76800 }, { "epoch": 0.4633928679104208, "grad_norm": 1.0814242362976074, "learning_rate": 0.00029699904702248383, "loss": 5.4248, "step": 76850 }, { "epoch": 0.4636943596917548, "grad_norm": 1.1077516078948975, "learning_rate": 0.0002969949956402485, "loss": 5.6891, "step": 76900 }, { "epoch": 0.46399585147308886, "grad_norm": 0.24430623650550842, "learning_rate": 0.0002969909415527801, "loss": 5.4433, "step": 76950 }, { "epoch": 0.4642973432544229, "grad_norm": 0.9930973052978516, "learning_rate": 0.00029698688476015317, "loss": 5.9078, "step": 77000 }, { "epoch": 0.46459883503575694, "grad_norm": 0.8889061212539673, "learning_rate": 0.0002969828252624423, "loss": 5.8593, "step": 77050 }, { "epoch": 0.464900326817091, "grad_norm": 1.4263575077056885, "learning_rate": 0.0002969787630597224, "loss": 5.3247, "step": 77100 }, { "epoch": 0.465201818598425, "grad_norm": 0.9487208127975464, "learning_rate": 0.0002969746981520681, "loss": 5.6282, "step": 77150 }, { "epoch": 0.46550331037975906, "grad_norm": 1.2874988317489624, "learning_rate": 0.0002969706305395542, "loss": 5.299, "step": 77200 }, { "epoch": 0.4658048021610931, "grad_norm": 1.4308826923370361, "learning_rate": 0.0002969665602222556, "loss": 5.6806, "step": 77250 }, { "epoch": 0.46610629394242714, "grad_norm": 1.2150838375091553, "learning_rate": 0.0002969624872002471, "loss": 4.9292, "step": 77300 }, { "epoch": 0.4664077857237612, "grad_norm": 1.1619287729263306, "learning_rate": 0.0002969584114736039, "loss": 5.7538, "step": 77350 }, { "epoch": 0.4667092775050952, "grad_norm": 1.606849193572998, "learning_rate": 0.0002969543330424007, "loss": 5.1169, "step": 77400 }, { "epoch": 0.46701076928642926, "grad_norm": 1.6578997373580933, "learning_rate": 0.00029695025190671286, "loss": 5.6213, "step": 77450 }, { "epoch": 0.4673122610677633, "grad_norm": 0.772125780582428, "learning_rate": 0.0002969461680666153, "loss": 5.3508, "step": 77500 }, { "epoch": 0.46761375284909734, "grad_norm": 1.0534485578536987, "learning_rate": 0.0002969420815221832, "loss": 5.5776, "step": 77550 }, { "epoch": 0.4679152446304314, "grad_norm": 1.5296164751052856, "learning_rate": 0.0002969379922734918, "loss": 6.087, "step": 77600 }, { "epoch": 0.4682167364117654, "grad_norm": 1.6921236515045166, "learning_rate": 0.0002969339003206163, "loss": 5.4066, "step": 77650 }, { "epoch": 0.46851822819309946, "grad_norm": 0.8419989943504333, "learning_rate": 0.0002969298056636321, "loss": 5.4254, "step": 77700 }, { "epoch": 0.4688197199744335, "grad_norm": 0.8434841632843018, "learning_rate": 0.0002969257083026145, "loss": 5.4875, "step": 77750 }, { "epoch": 0.46912121175576754, "grad_norm": 0.8605629205703735, "learning_rate": 0.00029692160823763887, "loss": 4.8735, "step": 77800 }, { "epoch": 0.4694227035371016, "grad_norm": 1.131161093711853, "learning_rate": 0.0002969175054687807, "loss": 5.4493, "step": 77850 }, { "epoch": 0.4697241953184356, "grad_norm": 0.7222887873649597, "learning_rate": 0.0002969133999961156, "loss": 5.1449, "step": 77900 }, { "epoch": 0.47002568709976966, "grad_norm": 1.2346384525299072, "learning_rate": 0.000296909291819719, "loss": 5.3922, "step": 77950 }, { "epoch": 0.4703271788811037, "grad_norm": 1.260045051574707, "learning_rate": 0.00029690518093966644, "loss": 5.6698, "step": 78000 }, { "epoch": 0.4703271788811037, "eval_loss": 5.9620184898376465, "eval_runtime": 39.1288, "eval_samples_per_second": 13.085, "eval_steps_per_second": 6.542, "eval_tts_loss": 6.910727731851027, "step": 78000 }, { "epoch": 0.47062867066243774, "grad_norm": 0.8269534111022949, "learning_rate": 0.0002969010673560338, "loss": 5.5578, "step": 78050 }, { "epoch": 0.4709301624437718, "grad_norm": 1.181211233139038, "learning_rate": 0.0002968969510688966, "loss": 5.1738, "step": 78100 }, { "epoch": 0.4712316542251058, "grad_norm": 0.5247254967689514, "learning_rate": 0.00029689283207833065, "loss": 5.5386, "step": 78150 }, { "epoch": 0.47153314600643986, "grad_norm": 0.33246809244155884, "learning_rate": 0.00029688871038441176, "loss": 5.1788, "step": 78200 }, { "epoch": 0.4718346377877739, "grad_norm": 1.293890357017517, "learning_rate": 0.00029688458598721575, "loss": 5.7622, "step": 78250 }, { "epoch": 0.47213612956910794, "grad_norm": 1.8805826902389526, "learning_rate": 0.00029688045888681854, "loss": 5.2554, "step": 78300 }, { "epoch": 0.472437621350442, "grad_norm": 1.2806355953216553, "learning_rate": 0.0002968763290832961, "loss": 5.6791, "step": 78350 }, { "epoch": 0.472739113131776, "grad_norm": 1.2846570014953613, "learning_rate": 0.00029687219657672446, "loss": 5.749, "step": 78400 }, { "epoch": 0.47304060491311006, "grad_norm": 0.8646582961082458, "learning_rate": 0.00029686806136717967, "loss": 5.2439, "step": 78450 }, { "epoch": 0.4733420966944441, "grad_norm": 1.0554583072662354, "learning_rate": 0.0002968639234547377, "loss": 5.542, "step": 78500 }, { "epoch": 0.47364358847577814, "grad_norm": 1.1340880393981934, "learning_rate": 0.0002968597828394749, "loss": 5.262, "step": 78550 }, { "epoch": 0.4739450802571122, "grad_norm": 0.532855749130249, "learning_rate": 0.00029685563952146736, "loss": 5.5503, "step": 78600 }, { "epoch": 0.4742465720384462, "grad_norm": 0.5302792191505432, "learning_rate": 0.0002968514935007913, "loss": 5.5127, "step": 78650 }, { "epoch": 0.47454806381978026, "grad_norm": 1.8584284782409668, "learning_rate": 0.0002968473447775231, "loss": 5.6061, "step": 78700 }, { "epoch": 0.4748495556011143, "grad_norm": 1.277764916419983, "learning_rate": 0.0002968431933517391, "loss": 5.6491, "step": 78750 }, { "epoch": 0.47515104738244834, "grad_norm": 1.9586448669433594, "learning_rate": 0.0002968390392235156, "loss": 5.0089, "step": 78800 }, { "epoch": 0.4754525391637824, "grad_norm": 1.049437165260315, "learning_rate": 0.00029683488239292923, "loss": 5.1839, "step": 78850 }, { "epoch": 0.4757540309451164, "grad_norm": 1.0131796598434448, "learning_rate": 0.00029683072286005636, "loss": 5.7506, "step": 78900 }, { "epoch": 0.47605552272645046, "grad_norm": 1.1836272478103638, "learning_rate": 0.00029682656062497356, "loss": 5.3738, "step": 78950 }, { "epoch": 0.4763570145077845, "grad_norm": 0.22917966544628143, "learning_rate": 0.0002968223956877575, "loss": 5.5409, "step": 79000 }, { "epoch": 0.47665850628911854, "grad_norm": 1.0831756591796875, "learning_rate": 0.0002968182280484847, "loss": 5.4883, "step": 79050 }, { "epoch": 0.4769599980704526, "grad_norm": 0.9845784306526184, "learning_rate": 0.00029681405770723196, "loss": 5.2493, "step": 79100 }, { "epoch": 0.4772614898517866, "grad_norm": 0.9840935468673706, "learning_rate": 0.00029680988466407603, "loss": 5.3734, "step": 79150 }, { "epoch": 0.47756298163312066, "grad_norm": 2.747628927230835, "learning_rate": 0.00029680570891909365, "loss": 5.4345, "step": 79200 }, { "epoch": 0.4778644734144547, "grad_norm": 1.6615885496139526, "learning_rate": 0.00029680153047236175, "loss": 5.2714, "step": 79250 }, { "epoch": 0.47816596519578874, "grad_norm": 3.3958709239959717, "learning_rate": 0.00029679734932395714, "loss": 5.0925, "step": 79300 }, { "epoch": 0.4784674569771228, "grad_norm": 0.7486457228660583, "learning_rate": 0.0002967931654739568, "loss": 5.5079, "step": 79350 }, { "epoch": 0.4787689487584568, "grad_norm": 1.2583531141281128, "learning_rate": 0.00029678897892243777, "loss": 5.6396, "step": 79400 }, { "epoch": 0.47907044053979086, "grad_norm": 1.1655434370040894, "learning_rate": 0.000296784789669477, "loss": 5.4852, "step": 79450 }, { "epoch": 0.4793719323211249, "grad_norm": 0.968452513217926, "learning_rate": 0.0002967805977151517, "loss": 5.9675, "step": 79500 }, { "epoch": 0.47967342410245895, "grad_norm": 1.2086753845214844, "learning_rate": 0.00029677640305953893, "loss": 5.2671, "step": 79550 }, { "epoch": 0.479974915883793, "grad_norm": 1.303040623664856, "learning_rate": 0.0002967722057027159, "loss": 5.5308, "step": 79600 }, { "epoch": 0.480276407665127, "grad_norm": 1.3369760513305664, "learning_rate": 0.00029676800564476, "loss": 5.7204, "step": 79650 }, { "epoch": 0.48057789944646107, "grad_norm": 0.19921384751796722, "learning_rate": 0.0002967638028857483, "loss": 5.2227, "step": 79700 }, { "epoch": 0.4808793912277951, "grad_norm": 1.106431484222412, "learning_rate": 0.00029675959742575823, "loss": 5.7613, "step": 79750 }, { "epoch": 0.48118088300912915, "grad_norm": 1.0413012504577637, "learning_rate": 0.0002967553892648673, "loss": 5.7266, "step": 79800 }, { "epoch": 0.4814823747904632, "grad_norm": 1.0932990312576294, "learning_rate": 0.0002967511784031528, "loss": 5.5258, "step": 79850 }, { "epoch": 0.4817838665717972, "grad_norm": 1.0389549732208252, "learning_rate": 0.00029674696484069233, "loss": 5.907, "step": 79900 }, { "epoch": 0.4820853583531313, "grad_norm": 0.8505406975746155, "learning_rate": 0.00029674274857756334, "loss": 5.4518, "step": 79950 }, { "epoch": 0.48238685013446536, "grad_norm": 0.9348762631416321, "learning_rate": 0.0002967385296138435, "loss": 5.8084, "step": 80000 }, { "epoch": 0.4826883419157994, "grad_norm": 0.30288946628570557, "learning_rate": 0.0002967343079496104, "loss": 5.467, "step": 80050 }, { "epoch": 0.48298983369713344, "grad_norm": 0.9844419360160828, "learning_rate": 0.00029673008358494185, "loss": 5.5362, "step": 80100 }, { "epoch": 0.4832913254784675, "grad_norm": 1.50944185256958, "learning_rate": 0.00029672585651991545, "loss": 5.4556, "step": 80150 }, { "epoch": 0.4835928172598015, "grad_norm": 0.9055566191673279, "learning_rate": 0.00029672162675460905, "loss": 5.3833, "step": 80200 }, { "epoch": 0.48389430904113556, "grad_norm": 1.1137325763702393, "learning_rate": 0.0002967173942891005, "loss": 5.6985, "step": 80250 }, { "epoch": 0.4841958008224696, "grad_norm": 1.1370432376861572, "learning_rate": 0.0002967131591234677, "loss": 5.5853, "step": 80300 }, { "epoch": 0.48449729260380364, "grad_norm": 1.2774473428726196, "learning_rate": 0.0002967089212577885, "loss": 5.1867, "step": 80350 }, { "epoch": 0.4847987843851377, "grad_norm": 1.208559513092041, "learning_rate": 0.00029670468069214106, "loss": 5.1143, "step": 80400 }, { "epoch": 0.4851002761664717, "grad_norm": 1.304936170578003, "learning_rate": 0.0002967004374266033, "loss": 5.6533, "step": 80450 }, { "epoch": 0.48540176794780576, "grad_norm": 1.22727632522583, "learning_rate": 0.0002966961914612533, "loss": 5.427, "step": 80500 }, { "epoch": 0.4857032597291398, "grad_norm": 0.8548924326896667, "learning_rate": 0.0002966919427961693, "loss": 5.1692, "step": 80550 }, { "epoch": 0.48600475151047384, "grad_norm": 0.6398463845252991, "learning_rate": 0.00029668769143142943, "loss": 5.2912, "step": 80600 }, { "epoch": 0.4863062432918079, "grad_norm": 0.6830285787582397, "learning_rate": 0.0002966834373671119, "loss": 5.0841, "step": 80650 }, { "epoch": 0.4866077350731419, "grad_norm": 1.339910864830017, "learning_rate": 0.00029667918060329507, "loss": 5.6013, "step": 80700 }, { "epoch": 0.48690922685447596, "grad_norm": 0.9510960578918457, "learning_rate": 0.0002966749211400572, "loss": 5.8169, "step": 80750 }, { "epoch": 0.48721071863581, "grad_norm": 0.7609521150588989, "learning_rate": 0.00029667065897747677, "loss": 5.9547, "step": 80800 }, { "epoch": 0.48751221041714404, "grad_norm": 0.5653070211410522, "learning_rate": 0.0002966663941156321, "loss": 5.2605, "step": 80850 }, { "epoch": 0.4878137021984781, "grad_norm": 1.0964018106460571, "learning_rate": 0.0002966621265546018, "loss": 5.66, "step": 80900 }, { "epoch": 0.4881151939798121, "grad_norm": 1.0000110864639282, "learning_rate": 0.0002966578562944644, "loss": 5.2816, "step": 80950 }, { "epoch": 0.48841668576114616, "grad_norm": 1.4042826890945435, "learning_rate": 0.0002966535833352984, "loss": 5.2441, "step": 81000 }, { "epoch": 0.48841668576114616, "eval_loss": 5.941708564758301, "eval_runtime": 38.8741, "eval_samples_per_second": 13.171, "eval_steps_per_second": 6.585, "eval_tts_loss": 6.916577768714163, "step": 81000 }, { "epoch": 0.4887181775424802, "grad_norm": 0.7111455798149109, "learning_rate": 0.00029664930767718246, "loss": 5.5238, "step": 81050 }, { "epoch": 0.48901966932381424, "grad_norm": 1.508941888809204, "learning_rate": 0.00029664502932019536, "loss": 5.6527, "step": 81100 }, { "epoch": 0.4893211611051483, "grad_norm": 1.1744884252548218, "learning_rate": 0.0002966407482644157, "loss": 5.0906, "step": 81150 }, { "epoch": 0.4896226528864823, "grad_norm": 1.4764869213104248, "learning_rate": 0.00029663646450992236, "loss": 5.3979, "step": 81200 }, { "epoch": 0.48992414466781636, "grad_norm": 0.7821881175041199, "learning_rate": 0.0002966321780567942, "loss": 5.4643, "step": 81250 }, { "epoch": 0.4902256364491504, "grad_norm": 1.6371777057647705, "learning_rate": 0.00029662788890511, "loss": 5.4499, "step": 81300 }, { "epoch": 0.49052712823048444, "grad_norm": 0.737937867641449, "learning_rate": 0.00029662359705494876, "loss": 5.1628, "step": 81350 }, { "epoch": 0.4908286200118185, "grad_norm": 1.213423252105713, "learning_rate": 0.0002966193025063895, "loss": 5.5077, "step": 81400 }, { "epoch": 0.4911301117931525, "grad_norm": 0.9045541286468506, "learning_rate": 0.00029661500525951116, "loss": 5.4829, "step": 81450 }, { "epoch": 0.49143160357448656, "grad_norm": 1.2195051908493042, "learning_rate": 0.00029661070531439293, "loss": 5.5261, "step": 81500 }, { "epoch": 0.4917330953558206, "grad_norm": 1.3019957542419434, "learning_rate": 0.00029660640267111383, "loss": 4.8683, "step": 81550 }, { "epoch": 0.49203458713715464, "grad_norm": 0.5789055824279785, "learning_rate": 0.0002966020973297532, "loss": 5.3057, "step": 81600 }, { "epoch": 0.4923360789184887, "grad_norm": 0.7689682245254517, "learning_rate": 0.0002965977892903901, "loss": 5.6641, "step": 81650 }, { "epoch": 0.4926375706998227, "grad_norm": 1.2568269968032837, "learning_rate": 0.00029659347855310393, "loss": 5.6835, "step": 81700 }, { "epoch": 0.49293906248115676, "grad_norm": 1.0251436233520508, "learning_rate": 0.00029658916511797397, "loss": 5.4541, "step": 81750 }, { "epoch": 0.4932405542624908, "grad_norm": 0.7457462549209595, "learning_rate": 0.00029658484898507963, "loss": 5.0554, "step": 81800 }, { "epoch": 0.49354204604382484, "grad_norm": 1.0542194843292236, "learning_rate": 0.00029658053015450033, "loss": 5.6991, "step": 81850 }, { "epoch": 0.4938435378251589, "grad_norm": 0.8511309623718262, "learning_rate": 0.00029657620862631555, "loss": 5.3717, "step": 81900 }, { "epoch": 0.4941450296064929, "grad_norm": 1.234178066253662, "learning_rate": 0.0002965718844006048, "loss": 5.2972, "step": 81950 }, { "epoch": 0.49444652138782696, "grad_norm": 1.3021661043167114, "learning_rate": 0.00029656755747744774, "loss": 5.5051, "step": 82000 }, { "epoch": 0.494748013169161, "grad_norm": 1.023585557937622, "learning_rate": 0.0002965632278569239, "loss": 5.3322, "step": 82050 }, { "epoch": 0.49504950495049505, "grad_norm": 1.016404151916504, "learning_rate": 0.000296558895539113, "loss": 5.6404, "step": 82100 }, { "epoch": 0.4953509967318291, "grad_norm": 1.1777606010437012, "learning_rate": 0.00029655456052409484, "loss": 5.3009, "step": 82150 }, { "epoch": 0.4956524885131631, "grad_norm": 0.8641856908798218, "learning_rate": 0.0002965502228119491, "loss": 5.7206, "step": 82200 }, { "epoch": 0.49595398029449717, "grad_norm": 0.5006489157676697, "learning_rate": 0.0002965458824027557, "loss": 5.3934, "step": 82250 }, { "epoch": 0.4962554720758312, "grad_norm": 0.9028316140174866, "learning_rate": 0.0002965415392965944, "loss": 5.3273, "step": 82300 }, { "epoch": 0.49655696385716525, "grad_norm": 0.8884007334709167, "learning_rate": 0.0002965371934935452, "loss": 5.3266, "step": 82350 }, { "epoch": 0.4968584556384993, "grad_norm": 1.0007665157318115, "learning_rate": 0.0002965328449936881, "loss": 5.1978, "step": 82400 }, { "epoch": 0.4971599474198333, "grad_norm": 1.060962200164795, "learning_rate": 0.0002965284937971031, "loss": 5.4143, "step": 82450 }, { "epoch": 0.49746143920116737, "grad_norm": 1.2199385166168213, "learning_rate": 0.0002965241399038703, "loss": 5.3939, "step": 82500 }, { "epoch": 0.4977629309825014, "grad_norm": 0.8604403138160706, "learning_rate": 0.0002965197833140698, "loss": 6.006, "step": 82550 }, { "epoch": 0.49806442276383545, "grad_norm": 0.6586524248123169, "learning_rate": 0.0002965154240277817, "loss": 5.6898, "step": 82600 }, { "epoch": 0.4983659145451695, "grad_norm": 0.4351431429386139, "learning_rate": 0.00029651106204508636, "loss": 5.0407, "step": 82650 }, { "epoch": 0.4986674063265035, "grad_norm": 1.0020102262496948, "learning_rate": 0.000296506697366064, "loss": 4.8001, "step": 82700 }, { "epoch": 0.49896889810783757, "grad_norm": 0.8024177551269531, "learning_rate": 0.00029650232999079497, "loss": 5.4484, "step": 82750 }, { "epoch": 0.4992703898891716, "grad_norm": 0.8950265049934387, "learning_rate": 0.00029649795991935966, "loss": 5.376, "step": 82800 }, { "epoch": 0.49957188167050565, "grad_norm": 2.1043708324432373, "learning_rate": 0.00029649358715183843, "loss": 5.3118, "step": 82850 }, { "epoch": 0.4998733734518397, "grad_norm": 2.049985885620117, "learning_rate": 0.0002964892116883118, "loss": 5.181, "step": 82900 }, { "epoch": 0.5001748652331738, "grad_norm": 1.4109222888946533, "learning_rate": 0.00029648483352886026, "loss": 5.9187, "step": 82950 }, { "epoch": 0.5004763570145078, "grad_norm": 0.7514991164207458, "learning_rate": 0.0002964804526735644, "loss": 5.472, "step": 83000 }, { "epoch": 0.5007778487958419, "grad_norm": 0.7477379441261292, "learning_rate": 0.00029647606912250486, "loss": 4.8014, "step": 83050 }, { "epoch": 0.5010793405771758, "grad_norm": 1.1122407913208008, "learning_rate": 0.0002964716828757623, "loss": 4.7121, "step": 83100 }, { "epoch": 0.5013808323585099, "grad_norm": 1.143982172012329, "learning_rate": 0.0002964672939334174, "loss": 5.1547, "step": 83150 }, { "epoch": 0.5016823241398439, "grad_norm": 0.9422672986984253, "learning_rate": 0.00029646290229555104, "loss": 4.8581, "step": 83200 }, { "epoch": 0.501983815921178, "grad_norm": 0.8290767073631287, "learning_rate": 0.000296458507962244, "loss": 5.3183, "step": 83250 }, { "epoch": 0.502285307702512, "grad_norm": 0.8387415409088135, "learning_rate": 0.000296454110933577, "loss": 5.4116, "step": 83300 }, { "epoch": 0.5025867994838461, "grad_norm": 1.3713529109954834, "learning_rate": 0.0002964497112096312, "loss": 5.5178, "step": 83350 }, { "epoch": 0.5028882912651801, "grad_norm": 1.4528504610061646, "learning_rate": 0.00029644530879048743, "loss": 5.2243, "step": 83400 }, { "epoch": 0.5031897830465142, "grad_norm": 0.8307807445526123, "learning_rate": 0.0002964409036762267, "loss": 5.4304, "step": 83450 }, { "epoch": 0.5034912748278482, "grad_norm": 0.902966320514679, "learning_rate": 0.00029643649586693016, "loss": 5.1011, "step": 83500 }, { "epoch": 0.5037927666091823, "grad_norm": 1.3196786642074585, "learning_rate": 0.0002964320853626789, "loss": 5.5364, "step": 83550 }, { "epoch": 0.5040942583905162, "grad_norm": 0.8948118686676025, "learning_rate": 0.0002964276721635541, "loss": 5.7282, "step": 83600 }, { "epoch": 0.5043957501718503, "grad_norm": 0.9051743149757385, "learning_rate": 0.0002964232562696369, "loss": 5.4241, "step": 83650 }, { "epoch": 0.5046972419531843, "grad_norm": 0.9922083020210266, "learning_rate": 0.00029641883768100866, "loss": 5.7088, "step": 83700 }, { "epoch": 0.5049987337345184, "grad_norm": 0.482597678899765, "learning_rate": 0.00029641441639775064, "loss": 4.8462, "step": 83750 }, { "epoch": 0.5053002255158524, "grad_norm": 1.143714189529419, "learning_rate": 0.00029640999241994427, "loss": 5.2544, "step": 83800 }, { "epoch": 0.5056017172971865, "grad_norm": 0.8007168173789978, "learning_rate": 0.00029640556574767084, "loss": 4.9798, "step": 83850 }, { "epoch": 0.5059032090785205, "grad_norm": 0.7514850497245789, "learning_rate": 0.000296401136381012, "loss": 5.4275, "step": 83900 }, { "epoch": 0.5062047008598546, "grad_norm": 1.6581178903579712, "learning_rate": 0.00029639670432004915, "loss": 4.9706, "step": 83950 }, { "epoch": 0.5065061926411886, "grad_norm": 1.0887938737869263, "learning_rate": 0.00029639226956486384, "loss": 5.666, "step": 84000 }, { "epoch": 0.5065061926411886, "eval_loss": 5.926547050476074, "eval_runtime": 39.2096, "eval_samples_per_second": 13.058, "eval_steps_per_second": 6.529, "eval_tts_loss": 6.930029721987113, "step": 84000 }, { "epoch": 0.5068076844225227, "grad_norm": 1.383786678314209, "learning_rate": 0.0002963878321155378, "loss": 6.0019, "step": 84050 }, { "epoch": 0.5071091762038566, "grad_norm": 1.0560262203216553, "learning_rate": 0.00029638339197215254, "loss": 5.2718, "step": 84100 }, { "epoch": 0.5074106679851907, "grad_norm": 0.9631621241569519, "learning_rate": 0.00029637894913478983, "loss": 4.9997, "step": 84150 }, { "epoch": 0.5077121597665247, "grad_norm": 0.8397107124328613, "learning_rate": 0.0002963745036035315, "loss": 5.2139, "step": 84200 }, { "epoch": 0.5080136515478588, "grad_norm": 0.7584561109542847, "learning_rate": 0.0002963700553784593, "loss": 5.3536, "step": 84250 }, { "epoch": 0.5083151433291928, "grad_norm": 1.544567584991455, "learning_rate": 0.0002963656044596551, "loss": 5.2901, "step": 84300 }, { "epoch": 0.5086166351105269, "grad_norm": 0.6083757877349854, "learning_rate": 0.0002963611508472009, "loss": 5.9236, "step": 84350 }, { "epoch": 0.5089181268918609, "grad_norm": 0.8062569499015808, "learning_rate": 0.00029635669454117854, "loss": 5.7083, "step": 84400 }, { "epoch": 0.509219618673195, "grad_norm": 1.0806671380996704, "learning_rate": 0.0002963522355416701, "loss": 5.299, "step": 84450 }, { "epoch": 0.509521110454529, "grad_norm": 1.004625678062439, "learning_rate": 0.00029634777384875756, "loss": 5.5976, "step": 84500 }, { "epoch": 0.5098226022358631, "grad_norm": 0.9023197293281555, "learning_rate": 0.0002963433094625231, "loss": 5.7333, "step": 84550 }, { "epoch": 0.510124094017197, "grad_norm": 0.8792926073074341, "learning_rate": 0.00029633884238304894, "loss": 5.4804, "step": 84600 }, { "epoch": 0.5104255857985311, "grad_norm": 0.869053840637207, "learning_rate": 0.00029633437261041717, "loss": 5.3615, "step": 84650 }, { "epoch": 0.5107270775798651, "grad_norm": 0.7250701785087585, "learning_rate": 0.00029632990014471014, "loss": 5.391, "step": 84700 }, { "epoch": 0.5110285693611992, "grad_norm": 1.052751064300537, "learning_rate": 0.00029632542498601015, "loss": 5.4083, "step": 84750 }, { "epoch": 0.5113300611425332, "grad_norm": 1.0572576522827148, "learning_rate": 0.00029632094713439943, "loss": 5.0461, "step": 84800 }, { "epoch": 0.5116315529238673, "grad_norm": 0.871200680732727, "learning_rate": 0.00029631646658996054, "loss": 5.7797, "step": 84850 }, { "epoch": 0.5119330447052013, "grad_norm": 0.7569305896759033, "learning_rate": 0.0002963119833527759, "loss": 5.4865, "step": 84900 }, { "epoch": 0.5122345364865354, "grad_norm": 0.8235620856285095, "learning_rate": 0.000296307497422928, "loss": 5.8864, "step": 84950 }, { "epoch": 0.5125360282678694, "grad_norm": 0.8417738676071167, "learning_rate": 0.0002963030088004994, "loss": 5.6301, "step": 85000 }, { "epoch": 0.5128375200492035, "grad_norm": 0.9542450308799744, "learning_rate": 0.0002962985174855727, "loss": 5.256, "step": 85050 }, { "epoch": 0.5131390118305375, "grad_norm": 1.0011111497879028, "learning_rate": 0.00029629402347823054, "loss": 5.2279, "step": 85100 }, { "epoch": 0.5134405036118715, "grad_norm": 0.9443592429161072, "learning_rate": 0.0002962895267785557, "loss": 5.1717, "step": 85150 }, { "epoch": 0.5137419953932055, "grad_norm": 1.470547080039978, "learning_rate": 0.00029628502738663086, "loss": 5.4199, "step": 85200 }, { "epoch": 0.5140434871745396, "grad_norm": 1.6413885354995728, "learning_rate": 0.00029628052530253883, "loss": 5.6561, "step": 85250 }, { "epoch": 0.5143449789558736, "grad_norm": 0.7837339639663696, "learning_rate": 0.00029627602052636256, "loss": 4.967, "step": 85300 }, { "epoch": 0.5146464707372077, "grad_norm": 0.9321562647819519, "learning_rate": 0.0002962715130581848, "loss": 5.6423, "step": 85350 }, { "epoch": 0.5149479625185417, "grad_norm": 1.0105329751968384, "learning_rate": 0.0002962670028980886, "loss": 5.2056, "step": 85400 }, { "epoch": 0.5152494542998758, "grad_norm": 0.8647049069404602, "learning_rate": 0.00029626249004615696, "loss": 5.7835, "step": 85450 }, { "epoch": 0.5155509460812098, "grad_norm": 0.47922226786613464, "learning_rate": 0.00029625797450247293, "loss": 5.3565, "step": 85500 }, { "epoch": 0.5158524378625439, "grad_norm": 0.23402847349643707, "learning_rate": 0.0002962534562671196, "loss": 5.3735, "step": 85550 }, { "epoch": 0.516153929643878, "grad_norm": 1.0629838705062866, "learning_rate": 0.0002962489353401801, "loss": 5.6394, "step": 85600 }, { "epoch": 0.516455421425212, "grad_norm": 1.5537623167037964, "learning_rate": 0.00029624441172173763, "loss": 5.3464, "step": 85650 }, { "epoch": 0.516756913206546, "grad_norm": 1.4734660387039185, "learning_rate": 0.00029623988541187555, "loss": 5.4604, "step": 85700 }, { "epoch": 0.51705840498788, "grad_norm": 1.4615983963012695, "learning_rate": 0.000296235356410677, "loss": 5.3485, "step": 85750 }, { "epoch": 0.5173598967692141, "grad_norm": 0.8337305188179016, "learning_rate": 0.00029623082471822544, "loss": 5.2589, "step": 85800 }, { "epoch": 0.5176613885505481, "grad_norm": 0.37481123208999634, "learning_rate": 0.0002962262903346042, "loss": 5.1446, "step": 85850 }, { "epoch": 0.5179628803318822, "grad_norm": 0.9868521690368652, "learning_rate": 0.0002962217532598968, "loss": 5.0823, "step": 85900 }, { "epoch": 0.5182643721132162, "grad_norm": 0.830217719078064, "learning_rate": 0.0002962172134941867, "loss": 5.4839, "step": 85950 }, { "epoch": 0.5185658638945503, "grad_norm": 0.4916618764400482, "learning_rate": 0.0002962126710375574, "loss": 5.3138, "step": 86000 }, { "epoch": 0.5188673556758843, "grad_norm": 1.5908881425857544, "learning_rate": 0.0002962081258900926, "loss": 5.5498, "step": 86050 }, { "epoch": 0.5191688474572184, "grad_norm": 0.45264601707458496, "learning_rate": 0.0002962035780518759, "loss": 5.2425, "step": 86100 }, { "epoch": 0.5194703392385523, "grad_norm": 1.4410377740859985, "learning_rate": 0.0002961990275229909, "loss": 5.4836, "step": 86150 }, { "epoch": 0.5197718310198864, "grad_norm": 1.5038870573043823, "learning_rate": 0.00029619447430352154, "loss": 5.4532, "step": 86200 }, { "epoch": 0.5200733228012204, "grad_norm": 1.1186045408248901, "learning_rate": 0.00029618991839355145, "loss": 5.6888, "step": 86250 }, { "epoch": 0.5203748145825545, "grad_norm": 0.3127157986164093, "learning_rate": 0.0002961853597931646, "loss": 5.2987, "step": 86300 }, { "epoch": 0.5206763063638885, "grad_norm": 1.3622223138809204, "learning_rate": 0.0002961807985024448, "loss": 5.7346, "step": 86350 }, { "epoch": 0.5209777981452226, "grad_norm": 1.0363080501556396, "learning_rate": 0.00029617623452147596, "loss": 5.6834, "step": 86400 }, { "epoch": 0.5212792899265566, "grad_norm": 0.8455144762992859, "learning_rate": 0.0002961716678503421, "loss": 5.5937, "step": 86450 }, { "epoch": 0.5215807817078907, "grad_norm": 0.9110015630722046, "learning_rate": 0.00029616709848912735, "loss": 5.6491, "step": 86500 }, { "epoch": 0.5218822734892247, "grad_norm": 0.5161350965499878, "learning_rate": 0.00029616252643791576, "loss": 5.4586, "step": 86550 }, { "epoch": 0.5221837652705588, "grad_norm": 1.6822437047958374, "learning_rate": 0.00029615795169679143, "loss": 5.3096, "step": 86600 }, { "epoch": 0.5224852570518927, "grad_norm": 0.4310929775238037, "learning_rate": 0.0002961533742658386, "loss": 5.2718, "step": 86650 }, { "epoch": 0.5227867488332268, "grad_norm": 1.2260253429412842, "learning_rate": 0.0002961487941451415, "loss": 5.302, "step": 86700 }, { "epoch": 0.5230882406145608, "grad_norm": 0.7922249436378479, "learning_rate": 0.0002961442113347844, "loss": 5.4593, "step": 86750 }, { "epoch": 0.5233897323958949, "grad_norm": 1.2659448385238647, "learning_rate": 0.0002961396258348516, "loss": 5.4191, "step": 86800 }, { "epoch": 0.5236912241772289, "grad_norm": 0.8767409920692444, "learning_rate": 0.00029613503764542757, "loss": 5.9018, "step": 86850 }, { "epoch": 0.523992715958563, "grad_norm": 1.4764957427978516, "learning_rate": 0.0002961304467665967, "loss": 5.1348, "step": 86900 }, { "epoch": 0.524294207739897, "grad_norm": 1.2076737880706787, "learning_rate": 0.00029612585319844354, "loss": 5.4157, "step": 86950 }, { "epoch": 0.5245956995212311, "grad_norm": 0.8235222697257996, "learning_rate": 0.0002961212569410526, "loss": 5.1896, "step": 87000 }, { "epoch": 0.5245956995212311, "eval_loss": 5.919267654418945, "eval_runtime": 39.0685, "eval_samples_per_second": 13.105, "eval_steps_per_second": 6.553, "eval_tts_loss": 6.9460464632279075, "step": 87000 }, { "epoch": 0.5248971913025651, "grad_norm": 0.8016168475151062, "learning_rate": 0.0002961166579945084, "loss": 5.0443, "step": 87050 }, { "epoch": 0.5251986830838992, "grad_norm": 1.28067147731781, "learning_rate": 0.00029611205635889565, "loss": 5.5611, "step": 87100 }, { "epoch": 0.5255001748652331, "grad_norm": 1.1068130731582642, "learning_rate": 0.000296107452034299, "loss": 5.8054, "step": 87150 }, { "epoch": 0.5258016666465672, "grad_norm": 0.9939075112342834, "learning_rate": 0.00029610284502080324, "loss": 5.4417, "step": 87200 }, { "epoch": 0.5261031584279012, "grad_norm": 0.9389002919197083, "learning_rate": 0.0002960982353184931, "loss": 5.5714, "step": 87250 }, { "epoch": 0.5264046502092353, "grad_norm": 0.9879956245422363, "learning_rate": 0.00029609362292745337, "loss": 5.7104, "step": 87300 }, { "epoch": 0.5267061419905693, "grad_norm": 0.8371336460113525, "learning_rate": 0.000296089007847769, "loss": 5.5891, "step": 87350 }, { "epoch": 0.5270076337719034, "grad_norm": 0.8993528485298157, "learning_rate": 0.00029608439007952506, "loss": 5.3184, "step": 87400 }, { "epoch": 0.5273091255532374, "grad_norm": 0.989437460899353, "learning_rate": 0.0002960797696228063, "loss": 5.593, "step": 87450 }, { "epoch": 0.5276106173345715, "grad_norm": 1.0070329904556274, "learning_rate": 0.00029607514647769784, "loss": 5.867, "step": 87500 }, { "epoch": 0.5279121091159055, "grad_norm": 1.6210565567016602, "learning_rate": 0.00029607052064428474, "loss": 4.9554, "step": 87550 }, { "epoch": 0.5282136008972396, "grad_norm": 0.8607068657875061, "learning_rate": 0.0002960658921226522, "loss": 5.0435, "step": 87600 }, { "epoch": 0.5285150926785736, "grad_norm": 0.9629681706428528, "learning_rate": 0.0002960612609128854, "loss": 5.3449, "step": 87650 }, { "epoch": 0.5288165844599076, "grad_norm": 1.3018798828125, "learning_rate": 0.0002960566270150695, "loss": 5.0184, "step": 87700 }, { "epoch": 0.5291180762412416, "grad_norm": 0.9879356026649475, "learning_rate": 0.00029605199042928977, "loss": 5.723, "step": 87750 }, { "epoch": 0.5294195680225757, "grad_norm": 0.7346367835998535, "learning_rate": 0.00029604735115563167, "loss": 4.9311, "step": 87800 }, { "epoch": 0.5297210598039097, "grad_norm": 0.826981246471405, "learning_rate": 0.00029604270919418043, "loss": 5.0309, "step": 87850 }, { "epoch": 0.5300225515852438, "grad_norm": 1.476569652557373, "learning_rate": 0.00029603806454502155, "loss": 5.7561, "step": 87900 }, { "epoch": 0.5303240433665778, "grad_norm": 1.6064908504486084, "learning_rate": 0.0002960334172082405, "loss": 5.7654, "step": 87950 }, { "epoch": 0.5306255351479119, "grad_norm": 0.8115198612213135, "learning_rate": 0.0002960287671839229, "loss": 5.2237, "step": 88000 }, { "epoch": 0.5309270269292459, "grad_norm": 1.1930946111679077, "learning_rate": 0.00029602411447215407, "loss": 5.103, "step": 88050 }, { "epoch": 0.53122851871058, "grad_norm": 1.1153074502944946, "learning_rate": 0.0002960194590730199, "loss": 5.5846, "step": 88100 }, { "epoch": 0.531530010491914, "grad_norm": 1.527891755104065, "learning_rate": 0.00029601480098660597, "loss": 4.9088, "step": 88150 }, { "epoch": 0.531831502273248, "grad_norm": 1.3054035902023315, "learning_rate": 0.000296010140212998, "loss": 5.6524, "step": 88200 }, { "epoch": 0.532132994054582, "grad_norm": 1.3570349216461182, "learning_rate": 0.0002960054767522817, "loss": 5.2411, "step": 88250 }, { "epoch": 0.5324344858359161, "grad_norm": 0.9342194199562073, "learning_rate": 0.00029600081060454305, "loss": 5.594, "step": 88300 }, { "epoch": 0.5327359776172501, "grad_norm": 1.0508936643600464, "learning_rate": 0.0002959961417698678, "loss": 5.5954, "step": 88350 }, { "epoch": 0.5330374693985842, "grad_norm": 1.0958936214447021, "learning_rate": 0.00029599147024834185, "loss": 5.3733, "step": 88400 }, { "epoch": 0.5333389611799182, "grad_norm": 1.6457786560058594, "learning_rate": 0.0002959867960400513, "loss": 5.2009, "step": 88450 }, { "epoch": 0.5336404529612523, "grad_norm": 1.507873296737671, "learning_rate": 0.0002959821191450821, "loss": 5.6467, "step": 88500 }, { "epoch": 0.5339419447425863, "grad_norm": 0.8614032864570618, "learning_rate": 0.0002959774395635203, "loss": 5.2579, "step": 88550 }, { "epoch": 0.5342434365239204, "grad_norm": 1.4928761720657349, "learning_rate": 0.000295972757295452, "loss": 5.4668, "step": 88600 }, { "epoch": 0.5345449283052544, "grad_norm": 0.8353002667427063, "learning_rate": 0.00029596807234096354, "loss": 5.1471, "step": 88650 }, { "epoch": 0.5348464200865884, "grad_norm": 0.9062924981117249, "learning_rate": 0.00029596338470014094, "loss": 5.1977, "step": 88700 }, { "epoch": 0.5351479118679224, "grad_norm": 1.1002637147903442, "learning_rate": 0.00029595869437307057, "loss": 5.3327, "step": 88750 }, { "epoch": 0.5354494036492565, "grad_norm": 0.8412830233573914, "learning_rate": 0.0002959540013598387, "loss": 5.4243, "step": 88800 }, { "epoch": 0.5357508954305905, "grad_norm": 0.6928727030754089, "learning_rate": 0.0002959493056605317, "loss": 5.2383, "step": 88850 }, { "epoch": 0.5360523872119246, "grad_norm": 0.870782196521759, "learning_rate": 0.0002959446072752361, "loss": 5.3942, "step": 88900 }, { "epoch": 0.5363538789932586, "grad_norm": 1.2821412086486816, "learning_rate": 0.00029593990620403813, "loss": 5.3919, "step": 88950 }, { "epoch": 0.5366553707745927, "grad_norm": 1.2082806825637817, "learning_rate": 0.0002959352024470246, "loss": 5.3026, "step": 89000 }, { "epoch": 0.5369568625559267, "grad_norm": 2.1101999282836914, "learning_rate": 0.0002959304960042819, "loss": 5.1243, "step": 89050 }, { "epoch": 0.5372583543372608, "grad_norm": 1.2788255214691162, "learning_rate": 0.00029592578687589663, "loss": 5.3174, "step": 89100 }, { "epoch": 0.5375598461185948, "grad_norm": 0.9972044229507446, "learning_rate": 0.0002959210750619555, "loss": 5.3555, "step": 89150 }, { "epoch": 0.5378613378999288, "grad_norm": 0.5922455787658691, "learning_rate": 0.0002959163605625452, "loss": 5.4449, "step": 89200 }, { "epoch": 0.5381628296812629, "grad_norm": 1.0951459407806396, "learning_rate": 0.0002959116433777526, "loss": 5.4858, "step": 89250 }, { "epoch": 0.5384643214625969, "grad_norm": 1.088260293006897, "learning_rate": 0.0002959069235076644, "loss": 5.3911, "step": 89300 }, { "epoch": 0.538765813243931, "grad_norm": 1.4618911743164062, "learning_rate": 0.00029590220095236744, "loss": 5.1905, "step": 89350 }, { "epoch": 0.539067305025265, "grad_norm": 0.8825082778930664, "learning_rate": 0.00029589747571194877, "loss": 5.3721, "step": 89400 }, { "epoch": 0.5393687968065991, "grad_norm": 1.1924505233764648, "learning_rate": 0.00029589274778649523, "loss": 4.9448, "step": 89450 }, { "epoch": 0.5396702885879331, "grad_norm": 0.5963086485862732, "learning_rate": 0.00029588801717609384, "loss": 5.2072, "step": 89500 }, { "epoch": 0.5399717803692672, "grad_norm": 0.9252137541770935, "learning_rate": 0.0002958832838808317, "loss": 5.2569, "step": 89550 }, { "epoch": 0.5402732721506012, "grad_norm": 0.5325434803962708, "learning_rate": 0.00029587854790079593, "loss": 5.6141, "step": 89600 }, { "epoch": 0.5405747639319353, "grad_norm": 1.03267240524292, "learning_rate": 0.00029587380923607365, "loss": 5.3742, "step": 89650 }, { "epoch": 0.5408762557132692, "grad_norm": 0.7044246196746826, "learning_rate": 0.0002958690678867521, "loss": 5.6692, "step": 89700 }, { "epoch": 0.5411777474946033, "grad_norm": 0.930395781993866, "learning_rate": 0.00029586432385291845, "loss": 5.4407, "step": 89750 }, { "epoch": 0.5414792392759373, "grad_norm": 1.019134521484375, "learning_rate": 0.0002958595771346602, "loss": 5.5451, "step": 89800 }, { "epoch": 0.5417807310572714, "grad_norm": 1.5057804584503174, "learning_rate": 0.0002958548277320645, "loss": 5.1586, "step": 89850 }, { "epoch": 0.5420822228386054, "grad_norm": 0.9856516122817993, "learning_rate": 0.00029585007564521883, "loss": 5.6374, "step": 89900 }, { "epoch": 0.5423837146199395, "grad_norm": 0.9164791107177734, "learning_rate": 0.00029584532087421076, "loss": 5.0127, "step": 89950 }, { "epoch": 0.5426852064012735, "grad_norm": 1.319340467453003, "learning_rate": 0.0002958405634191276, "loss": 5.1767, "step": 90000 }, { "epoch": 0.5426852064012735, "eval_loss": 5.916418075561523, "eval_runtime": 38.8573, "eval_samples_per_second": 13.176, "eval_steps_per_second": 6.588, "eval_tts_loss": 6.9209556892380535, "step": 90000 }, { "epoch": 0.5429866981826076, "grad_norm": 1.027008056640625, "learning_rate": 0.000295835803280057, "loss": 5.4936, "step": 90050 }, { "epoch": 0.5432881899639416, "grad_norm": 1.5209957361221313, "learning_rate": 0.0002958310404570866, "loss": 5.6744, "step": 90100 }, { "epoch": 0.5435896817452757, "grad_norm": 0.8627886176109314, "learning_rate": 0.000295826274950304, "loss": 5.4667, "step": 90150 }, { "epoch": 0.5438911735266097, "grad_norm": 0.8732492327690125, "learning_rate": 0.0002958215067597969, "loss": 5.8721, "step": 90200 }, { "epoch": 0.5441926653079437, "grad_norm": 1.1845695972442627, "learning_rate": 0.00029581673588565305, "loss": 5.6332, "step": 90250 }, { "epoch": 0.5444941570892777, "grad_norm": 1.0399823188781738, "learning_rate": 0.0002958119623279603, "loss": 5.3797, "step": 90300 }, { "epoch": 0.5447956488706118, "grad_norm": 1.5306127071380615, "learning_rate": 0.00029580718608680643, "loss": 5.1072, "step": 90350 }, { "epoch": 0.5450971406519458, "grad_norm": 0.809299111366272, "learning_rate": 0.00029580240716227944, "loss": 5.4646, "step": 90400 }, { "epoch": 0.5453986324332799, "grad_norm": 1.0564802885055542, "learning_rate": 0.00029579762555446715, "loss": 5.3486, "step": 90450 }, { "epoch": 0.5457001242146139, "grad_norm": 1.6013284921646118, "learning_rate": 0.0002957928412634577, "loss": 5.4152, "step": 90500 }, { "epoch": 0.546001615995948, "grad_norm": 2.0678458213806152, "learning_rate": 0.000295788054289339, "loss": 5.2083, "step": 90550 }, { "epoch": 0.546303107777282, "grad_norm": 1.0935783386230469, "learning_rate": 0.0002957832646321992, "loss": 5.2794, "step": 90600 }, { "epoch": 0.5466045995586161, "grad_norm": 1.8490805625915527, "learning_rate": 0.00029577847229212654, "loss": 5.0647, "step": 90650 }, { "epoch": 0.54690609133995, "grad_norm": 0.8930879235267639, "learning_rate": 0.0002957736772692091, "loss": 5.8464, "step": 90700 }, { "epoch": 0.5472075831212841, "grad_norm": 1.245474100112915, "learning_rate": 0.00029576887956353514, "loss": 5.9346, "step": 90750 }, { "epoch": 0.5475090749026181, "grad_norm": 0.8117373585700989, "learning_rate": 0.000295764079175193, "loss": 5.7229, "step": 90800 }, { "epoch": 0.5478105666839522, "grad_norm": 1.0757819414138794, "learning_rate": 0.00029575927610427094, "loss": 5.26, "step": 90850 }, { "epoch": 0.5481120584652862, "grad_norm": 1.1484211683273315, "learning_rate": 0.00029575447035085745, "loss": 5.7233, "step": 90900 }, { "epoch": 0.5484135502466203, "grad_norm": 0.4875519871711731, "learning_rate": 0.00029574966191504094, "loss": 5.444, "step": 90950 }, { "epoch": 0.5487150420279543, "grad_norm": 1.2396948337554932, "learning_rate": 0.00029574485079690994, "loss": 5.488, "step": 91000 }, { "epoch": 0.5490165338092884, "grad_norm": 0.8897046446800232, "learning_rate": 0.00029574003699655286, "loss": 5.3195, "step": 91050 }, { "epoch": 0.5493180255906224, "grad_norm": 1.3373503684997559, "learning_rate": 0.00029573522051405844, "loss": 5.5019, "step": 91100 }, { "epoch": 0.5496195173719565, "grad_norm": 1.3454574346542358, "learning_rate": 0.00029573040134951527, "loss": 5.2181, "step": 91150 }, { "epoch": 0.5499210091532905, "grad_norm": 0.8776136040687561, "learning_rate": 0.000295725579503012, "loss": 5.5189, "step": 91200 }, { "epoch": 0.5502225009346245, "grad_norm": 0.6781423091888428, "learning_rate": 0.0002957207549746374, "loss": 4.9925, "step": 91250 }, { "epoch": 0.5505239927159585, "grad_norm": 0.7907350659370422, "learning_rate": 0.0002957159277644803, "loss": 5.7975, "step": 91300 }, { "epoch": 0.5508254844972926, "grad_norm": 0.7251743078231812, "learning_rate": 0.0002957110978726295, "loss": 5.4187, "step": 91350 }, { "epoch": 0.5511269762786266, "grad_norm": 0.26917263865470886, "learning_rate": 0.00029570626529917387, "loss": 5.2196, "step": 91400 }, { "epoch": 0.5514284680599607, "grad_norm": 0.8050959706306458, "learning_rate": 0.0002957014300442023, "loss": 5.5182, "step": 91450 }, { "epoch": 0.5517299598412947, "grad_norm": 1.075561761856079, "learning_rate": 0.0002956965921078039, "loss": 5.5006, "step": 91500 }, { "epoch": 0.5520314516226288, "grad_norm": 1.4412881135940552, "learning_rate": 0.00029569175149006766, "loss": 5.5795, "step": 91550 }, { "epoch": 0.5523329434039628, "grad_norm": 1.1534234285354614, "learning_rate": 0.0002956869081910826, "loss": 5.3229, "step": 91600 }, { "epoch": 0.5526344351852969, "grad_norm": 0.7754513621330261, "learning_rate": 0.00029568206221093786, "loss": 5.367, "step": 91650 }, { "epoch": 0.5529359269666309, "grad_norm": 0.7983711361885071, "learning_rate": 0.0002956772135497227, "loss": 5.2837, "step": 91700 }, { "epoch": 0.553237418747965, "grad_norm": 0.8333536386489868, "learning_rate": 0.00029567236220752637, "loss": 5.5446, "step": 91750 }, { "epoch": 0.5535389105292989, "grad_norm": 0.39866629242897034, "learning_rate": 0.0002956675081844381, "loss": 5.4599, "step": 91800 }, { "epoch": 0.553840402310633, "grad_norm": 0.9978666305541992, "learning_rate": 0.00029566265148054716, "loss": 5.656, "step": 91850 }, { "epoch": 0.554141894091967, "grad_norm": 0.8354892134666443, "learning_rate": 0.000295657792095943, "loss": 5.5238, "step": 91900 }, { "epoch": 0.5544433858733011, "grad_norm": 1.554002046585083, "learning_rate": 0.00029565293003071513, "loss": 5.6428, "step": 91950 }, { "epoch": 0.5547448776546351, "grad_norm": 0.7707767486572266, "learning_rate": 0.0002956480652849528, "loss": 5.3278, "step": 92000 }, { "epoch": 0.5550463694359692, "grad_norm": 0.8927907943725586, "learning_rate": 0.0002956431978587458, "loss": 5.7211, "step": 92050 }, { "epoch": 0.5553478612173032, "grad_norm": 0.7539495229721069, "learning_rate": 0.0002956383277521835, "loss": 5.2473, "step": 92100 }, { "epoch": 0.5556493529986373, "grad_norm": 1.0779447555541992, "learning_rate": 0.00029563345496535565, "loss": 5.2826, "step": 92150 }, { "epoch": 0.5559508447799713, "grad_norm": 0.823029637336731, "learning_rate": 0.0002956285794983519, "loss": 5.6821, "step": 92200 }, { "epoch": 0.5562523365613053, "grad_norm": 0.9238352179527283, "learning_rate": 0.00029562370135126196, "loss": 5.6488, "step": 92250 }, { "epoch": 0.5565538283426393, "grad_norm": 0.8013512492179871, "learning_rate": 0.0002956188205241756, "loss": 5.141, "step": 92300 }, { "epoch": 0.5568553201239734, "grad_norm": 1.0475106239318848, "learning_rate": 0.00029561393701718267, "loss": 5.8431, "step": 92350 }, { "epoch": 0.5571568119053074, "grad_norm": 1.6498981714248657, "learning_rate": 0.000295609050830373, "loss": 5.3259, "step": 92400 }, { "epoch": 0.5574583036866415, "grad_norm": 1.4449446201324463, "learning_rate": 0.00029560416196383664, "loss": 5.5394, "step": 92450 }, { "epoch": 0.5577597954679755, "grad_norm": 0.9109590649604797, "learning_rate": 0.0002955992704176634, "loss": 5.3314, "step": 92500 }, { "epoch": 0.5580612872493096, "grad_norm": 1.1420055627822876, "learning_rate": 0.00029559437619194336, "loss": 5.183, "step": 92550 }, { "epoch": 0.5583627790306436, "grad_norm": 0.8042157292366028, "learning_rate": 0.0002955894792867666, "loss": 5.531, "step": 92600 }, { "epoch": 0.5586642708119777, "grad_norm": 1.2839215993881226, "learning_rate": 0.0002955845797022233, "loss": 4.937, "step": 92650 }, { "epoch": 0.5589657625933117, "grad_norm": 0.8924357891082764, "learning_rate": 0.0002955796774384035, "loss": 5.0597, "step": 92700 }, { "epoch": 0.5592672543746458, "grad_norm": 1.4216747283935547, "learning_rate": 0.00029557477249539756, "loss": 5.5157, "step": 92750 }, { "epoch": 0.5595687461559797, "grad_norm": 1.1316231489181519, "learning_rate": 0.00029556986487329557, "loss": 4.9426, "step": 92800 }, { "epoch": 0.5598702379373138, "grad_norm": 1.2961785793304443, "learning_rate": 0.00029556495457218806, "loss": 5.1961, "step": 92850 }, { "epoch": 0.5601717297186478, "grad_norm": 1.775773286819458, "learning_rate": 0.00029556004159216525, "loss": 5.7491, "step": 92900 }, { "epoch": 0.5604732214999819, "grad_norm": 1.044104814529419, "learning_rate": 0.0002955551259333176, "loss": 5.612, "step": 92950 }, { "epoch": 0.560774713281316, "grad_norm": 1.1744459867477417, "learning_rate": 0.0002955502075957356, "loss": 5.2724, "step": 93000 }, { "epoch": 0.560774713281316, "eval_loss": 5.889717102050781, "eval_runtime": 39.1081, "eval_samples_per_second": 13.092, "eval_steps_per_second": 6.546, "eval_tts_loss": 6.9917208123877534, "step": 93000 }, { "epoch": 0.56107620506265, "grad_norm": 1.6756510734558105, "learning_rate": 0.00029554528657950967, "loss": 5.5865, "step": 93050 }, { "epoch": 0.5613776968439841, "grad_norm": 0.8031917214393616, "learning_rate": 0.0002955403628847305, "loss": 5.2548, "step": 93100 }, { "epoch": 0.5616791886253181, "grad_norm": 0.5974986553192139, "learning_rate": 0.00029553543651148867, "loss": 5.7795, "step": 93150 }, { "epoch": 0.5619806804066522, "grad_norm": 1.2687832117080688, "learning_rate": 0.0002955305074598748, "loss": 5.5976, "step": 93200 }, { "epoch": 0.5622821721879862, "grad_norm": 0.7445743680000305, "learning_rate": 0.0002955255757299796, "loss": 5.4172, "step": 93250 }, { "epoch": 0.5625836639693202, "grad_norm": 0.8189687728881836, "learning_rate": 0.0002955206413218939, "loss": 5.7694, "step": 93300 }, { "epoch": 0.5628851557506542, "grad_norm": 1.6461641788482666, "learning_rate": 0.0002955157042357084, "loss": 5.1375, "step": 93350 }, { "epoch": 0.5631866475319883, "grad_norm": 1.0595647096633911, "learning_rate": 0.0002955107644715141, "loss": 5.5492, "step": 93400 }, { "epoch": 0.5634881393133223, "grad_norm": 0.5051787495613098, "learning_rate": 0.0002955058220294018, "loss": 4.8428, "step": 93450 }, { "epoch": 0.5637896310946564, "grad_norm": 1.3675822019577026, "learning_rate": 0.00029550087690946244, "loss": 5.479, "step": 93500 }, { "epoch": 0.5640911228759904, "grad_norm": 0.8800107836723328, "learning_rate": 0.00029549592911178713, "loss": 5.5756, "step": 93550 }, { "epoch": 0.5643926146573245, "grad_norm": 1.051252007484436, "learning_rate": 0.0002954909786364669, "loss": 5.1978, "step": 93600 }, { "epoch": 0.5646941064386585, "grad_norm": 0.853618860244751, "learning_rate": 0.0002954860254835928, "loss": 5.116, "step": 93650 }, { "epoch": 0.5649955982199926, "grad_norm": 0.5990920662879944, "learning_rate": 0.00029548106965325604, "loss": 5.2201, "step": 93700 }, { "epoch": 0.5652970900013266, "grad_norm": 1.4609922170639038, "learning_rate": 0.0002954761111455478, "loss": 5.8572, "step": 93750 }, { "epoch": 0.5655985817826606, "grad_norm": 1.415482759475708, "learning_rate": 0.0002954711499605594, "loss": 5.1871, "step": 93800 }, { "epoch": 0.5659000735639946, "grad_norm": 0.6907297372817993, "learning_rate": 0.000295466186098382, "loss": 5.3093, "step": 93850 }, { "epoch": 0.5662015653453287, "grad_norm": 1.3553764820098877, "learning_rate": 0.00029546121955910707, "loss": 5.4845, "step": 93900 }, { "epoch": 0.5665030571266627, "grad_norm": 0.42759907245635986, "learning_rate": 0.000295456250342826, "loss": 5.4478, "step": 93950 }, { "epoch": 0.5668045489079968, "grad_norm": 1.2371617555618286, "learning_rate": 0.0002954512784496302, "loss": 5.5596, "step": 94000 }, { "epoch": 0.5671060406893308, "grad_norm": 1.059735655784607, "learning_rate": 0.00029544630387961113, "loss": 5.1305, "step": 94050 }, { "epoch": 0.5674075324706649, "grad_norm": 1.2887678146362305, "learning_rate": 0.0002954413266328604, "loss": 5.7492, "step": 94100 }, { "epoch": 0.5677090242519989, "grad_norm": 0.6831839084625244, "learning_rate": 0.0002954363467094697, "loss": 4.9885, "step": 94150 }, { "epoch": 0.568010516033333, "grad_norm": 0.8960145115852356, "learning_rate": 0.00029543136410953053, "loss": 5.6091, "step": 94200 }, { "epoch": 0.568312007814667, "grad_norm": 0.8707526922225952, "learning_rate": 0.00029542637883313473, "loss": 5.2918, "step": 94250 }, { "epoch": 0.568613499596001, "grad_norm": 1.7414612770080566, "learning_rate": 0.0002954213908803739, "loss": 5.2493, "step": 94300 }, { "epoch": 0.568914991377335, "grad_norm": 0.6358331441879272, "learning_rate": 0.00029541640025133984, "loss": 5.4745, "step": 94350 }, { "epoch": 0.5692164831586691, "grad_norm": 0.7650109529495239, "learning_rate": 0.00029541140694612457, "loss": 4.999, "step": 94400 }, { "epoch": 0.5695179749400031, "grad_norm": 0.9262018203735352, "learning_rate": 0.00029540641096481975, "loss": 5.5209, "step": 94450 }, { "epoch": 0.5698194667213372, "grad_norm": 0.9247233271598816, "learning_rate": 0.00029540141230751754, "loss": 5.6021, "step": 94500 }, { "epoch": 0.5701209585026712, "grad_norm": 0.9050551056861877, "learning_rate": 0.0002953964109743098, "loss": 5.4158, "step": 94550 }, { "epoch": 0.5704224502840053, "grad_norm": 0.981111466884613, "learning_rate": 0.00029539140696528855, "loss": 5.3791, "step": 94600 }, { "epoch": 0.5707239420653393, "grad_norm": 0.996667742729187, "learning_rate": 0.00029538640028054603, "loss": 5.2808, "step": 94650 }, { "epoch": 0.5710254338466734, "grad_norm": 1.0466632843017578, "learning_rate": 0.00029538139092017427, "loss": 5.3422, "step": 94700 }, { "epoch": 0.5713269256280074, "grad_norm": 0.8787126541137695, "learning_rate": 0.0002953763788842654, "loss": 5.4083, "step": 94750 }, { "epoch": 0.5716284174093414, "grad_norm": 0.9230401515960693, "learning_rate": 0.0002953713641729118, "loss": 4.8325, "step": 94800 }, { "epoch": 0.5719299091906754, "grad_norm": 1.8314507007598877, "learning_rate": 0.00029536634678620574, "loss": 5.0028, "step": 94850 }, { "epoch": 0.5722314009720095, "grad_norm": 0.541100025177002, "learning_rate": 0.00029536132672423947, "loss": 5.3996, "step": 94900 }, { "epoch": 0.5725328927533435, "grad_norm": 1.6427440643310547, "learning_rate": 0.00029535630398710544, "loss": 5.155, "step": 94950 }, { "epoch": 0.5728343845346776, "grad_norm": 0.7480568289756775, "learning_rate": 0.000295351278574896, "loss": 5.6241, "step": 95000 }, { "epoch": 0.5731358763160116, "grad_norm": 1.3986549377441406, "learning_rate": 0.0002953462504877038, "loss": 4.9568, "step": 95050 }, { "epoch": 0.5734373680973457, "grad_norm": 0.9017693400382996, "learning_rate": 0.00029534121972562123, "loss": 5.455, "step": 95100 }, { "epoch": 0.5737388598786797, "grad_norm": 0.6557192802429199, "learning_rate": 0.00029533618628874094, "loss": 5.4549, "step": 95150 }, { "epoch": 0.5740403516600138, "grad_norm": 0.8418014049530029, "learning_rate": 0.00029533115017715556, "loss": 5.3938, "step": 95200 }, { "epoch": 0.5743418434413478, "grad_norm": 1.575558066368103, "learning_rate": 0.00029532611139095776, "loss": 5.2057, "step": 95250 }, { "epoch": 0.5746433352226819, "grad_norm": 1.246448278427124, "learning_rate": 0.00029532106993024023, "loss": 4.9871, "step": 95300 }, { "epoch": 0.5749448270040158, "grad_norm": 0.6200658082962036, "learning_rate": 0.0002953160257950959, "loss": 5.7753, "step": 95350 }, { "epoch": 0.5752463187853499, "grad_norm": 0.39429575204849243, "learning_rate": 0.0002953109789856173, "loss": 5.6409, "step": 95400 }, { "epoch": 0.5755478105666839, "grad_norm": 0.8576977252960205, "learning_rate": 0.0002953059295018977, "loss": 5.3981, "step": 95450 }, { "epoch": 0.575849302348018, "grad_norm": 1.3233672380447388, "learning_rate": 0.0002953008773440297, "loss": 5.3773, "step": 95500 }, { "epoch": 0.576150794129352, "grad_norm": 1.070857286453247, "learning_rate": 0.0002952958225121065, "loss": 5.5175, "step": 95550 }, { "epoch": 0.5764522859106861, "grad_norm": 0.660190999507904, "learning_rate": 0.00029529076500622094, "loss": 5.5603, "step": 95600 }, { "epoch": 0.5767537776920201, "grad_norm": 1.0487653017044067, "learning_rate": 0.00029528570482646624, "loss": 5.4333, "step": 95650 }, { "epoch": 0.5770552694733542, "grad_norm": 0.622514545917511, "learning_rate": 0.0002952806419729355, "loss": 5.1062, "step": 95700 }, { "epoch": 0.5773567612546882, "grad_norm": 0.6715002059936523, "learning_rate": 0.00029527557644572184, "loss": 5.1804, "step": 95750 }, { "epoch": 0.5776582530360223, "grad_norm": 1.1689550876617432, "learning_rate": 0.0002952705082449185, "loss": 5.7058, "step": 95800 }, { "epoch": 0.5779597448173562, "grad_norm": 0.4240676760673523, "learning_rate": 0.0002952654373706188, "loss": 5.7151, "step": 95850 }, { "epoch": 0.5782612365986903, "grad_norm": 0.8572235107421875, "learning_rate": 0.000295260363822916, "loss": 5.4224, "step": 95900 }, { "epoch": 0.5785627283800243, "grad_norm": 1.1103814840316772, "learning_rate": 0.0002952552876019036, "loss": 5.2321, "step": 95950 }, { "epoch": 0.5788642201613584, "grad_norm": 1.2996931076049805, "learning_rate": 0.0002952502087076748, "loss": 5.7763, "step": 96000 }, { "epoch": 0.5788642201613584, "eval_loss": 5.861026287078857, "eval_runtime": 39.0459, "eval_samples_per_second": 13.113, "eval_steps_per_second": 6.556, "eval_tts_loss": 6.967499278766865, "step": 96000 }, { "epoch": 0.5791657119426924, "grad_norm": 0.668563961982727, "learning_rate": 0.0002952451271403232, "loss": 5.2668, "step": 96050 }, { "epoch": 0.5794672037240265, "grad_norm": 1.2248414754867554, "learning_rate": 0.0002952400428999424, "loss": 5.5719, "step": 96100 }, { "epoch": 0.5797686955053605, "grad_norm": 1.0758379697799683, "learning_rate": 0.0002952349559866258, "loss": 5.932, "step": 96150 }, { "epoch": 0.5800701872866946, "grad_norm": 0.8841288089752197, "learning_rate": 0.0002952298664004671, "loss": 5.4878, "step": 96200 }, { "epoch": 0.5803716790680286, "grad_norm": 0.35963162779808044, "learning_rate": 0.00029522477414155996, "loss": 5.112, "step": 96250 }, { "epoch": 0.5806731708493627, "grad_norm": 1.2029433250427246, "learning_rate": 0.00029521967920999813, "loss": 5.3116, "step": 96300 }, { "epoch": 0.5809746626306966, "grad_norm": 0.9555790424346924, "learning_rate": 0.00029521458160587535, "loss": 5.803, "step": 96350 }, { "epoch": 0.5812761544120307, "grad_norm": 1.6402039527893066, "learning_rate": 0.0002952094813292854, "loss": 5.4063, "step": 96400 }, { "epoch": 0.5815776461933647, "grad_norm": 1.1883490085601807, "learning_rate": 0.0002952043783803222, "loss": 5.584, "step": 96450 }, { "epoch": 0.5818791379746988, "grad_norm": 0.846251368522644, "learning_rate": 0.0002951992727590796, "loss": 5.3569, "step": 96500 }, { "epoch": 0.5821806297560328, "grad_norm": 1.4803708791732788, "learning_rate": 0.00029519416446565163, "loss": 5.5011, "step": 96550 }, { "epoch": 0.5824821215373669, "grad_norm": 0.9556460380554199, "learning_rate": 0.00029518905350013223, "loss": 5.2851, "step": 96600 }, { "epoch": 0.5827836133187009, "grad_norm": 0.8582298755645752, "learning_rate": 0.00029518393986261555, "loss": 5.7105, "step": 96650 }, { "epoch": 0.583085105100035, "grad_norm": 0.5510904788970947, "learning_rate": 0.0002951788235531956, "loss": 5.0821, "step": 96700 }, { "epoch": 0.5833865968813691, "grad_norm": 0.7597860097885132, "learning_rate": 0.0002951737045719666, "loss": 5.4697, "step": 96750 }, { "epoch": 0.583688088662703, "grad_norm": 0.883868396282196, "learning_rate": 0.00029516858291902273, "loss": 5.1684, "step": 96800 }, { "epoch": 0.5839895804440371, "grad_norm": 0.8884049654006958, "learning_rate": 0.0002951634585944583, "loss": 5.4475, "step": 96850 }, { "epoch": 0.5842910722253711, "grad_norm": 1.349727749824524, "learning_rate": 0.00029515833159836753, "loss": 5.3183, "step": 96900 }, { "epoch": 0.5845925640067052, "grad_norm": 0.9678090810775757, "learning_rate": 0.0002951532019308448, "loss": 5.5043, "step": 96950 }, { "epoch": 0.5848940557880392, "grad_norm": 0.5846478343009949, "learning_rate": 0.0002951480695919846, "loss": 5.4603, "step": 97000 }, { "epoch": 0.5851955475693733, "grad_norm": 1.352579116821289, "learning_rate": 0.00029514293458188126, "loss": 5.267, "step": 97050 }, { "epoch": 0.5854970393507073, "grad_norm": 1.3360036611557007, "learning_rate": 0.00029513779690062936, "loss": 5.3711, "step": 97100 }, { "epoch": 0.5857985311320414, "grad_norm": 1.1508197784423828, "learning_rate": 0.00029513265654832344, "loss": 5.7283, "step": 97150 }, { "epoch": 0.5861000229133754, "grad_norm": 1.1349819898605347, "learning_rate": 0.00029512751352505804, "loss": 5.9288, "step": 97200 }, { "epoch": 0.5864015146947095, "grad_norm": 0.28436875343322754, "learning_rate": 0.0002951223678309279, "loss": 5.1823, "step": 97250 }, { "epoch": 0.5867030064760435, "grad_norm": 0.8319147229194641, "learning_rate": 0.0002951172194660277, "loss": 5.1999, "step": 97300 }, { "epoch": 0.5870044982573775, "grad_norm": 1.7307416200637817, "learning_rate": 0.00029511206843045216, "loss": 5.2025, "step": 97350 }, { "epoch": 0.5873059900387115, "grad_norm": 0.8515149354934692, "learning_rate": 0.00029510691472429604, "loss": 5.0153, "step": 97400 }, { "epoch": 0.5876074818200456, "grad_norm": 1.5850592851638794, "learning_rate": 0.0002951017583476543, "loss": 5.4888, "step": 97450 }, { "epoch": 0.5879089736013796, "grad_norm": 1.1616911888122559, "learning_rate": 0.0002950965993006217, "loss": 5.511, "step": 97500 }, { "epoch": 0.5882104653827137, "grad_norm": 1.0337568521499634, "learning_rate": 0.0002950914375832933, "loss": 5.0654, "step": 97550 }, { "epoch": 0.5885119571640477, "grad_norm": 0.6623184680938721, "learning_rate": 0.0002950862731957641, "loss": 5.5122, "step": 97600 }, { "epoch": 0.5888134489453818, "grad_norm": 1.450376033782959, "learning_rate": 0.000295081106138129, "loss": 5.6072, "step": 97650 }, { "epoch": 0.5891149407267158, "grad_norm": 0.9526558518409729, "learning_rate": 0.00029507593641048323, "loss": 5.9595, "step": 97700 }, { "epoch": 0.5894164325080499, "grad_norm": 1.1840243339538574, "learning_rate": 0.00029507076401292186, "loss": 5.4585, "step": 97750 }, { "epoch": 0.5897179242893839, "grad_norm": 1.454665184020996, "learning_rate": 0.0002950655889455401, "loss": 5.2741, "step": 97800 }, { "epoch": 0.590019416070718, "grad_norm": 1.129980444908142, "learning_rate": 0.00029506041120843323, "loss": 5.3889, "step": 97850 }, { "epoch": 0.5903209078520519, "grad_norm": 1.3016072511672974, "learning_rate": 0.00029505523080169645, "loss": 5.824, "step": 97900 }, { "epoch": 0.590622399633386, "grad_norm": 1.2617602348327637, "learning_rate": 0.0002950500477254252, "loss": 4.8803, "step": 97950 }, { "epoch": 0.59092389141472, "grad_norm": 1.5184402465820312, "learning_rate": 0.0002950448619797148, "loss": 5.2136, "step": 98000 }, { "epoch": 0.5912253831960541, "grad_norm": 0.8517768383026123, "learning_rate": 0.00029503967356466073, "loss": 5.3775, "step": 98050 }, { "epoch": 0.5915268749773881, "grad_norm": 1.3449987173080444, "learning_rate": 0.0002950344824803584, "loss": 5.5908, "step": 98100 }, { "epoch": 0.5918283667587222, "grad_norm": 0.39074525237083435, "learning_rate": 0.0002950292887269035, "loss": 5.1833, "step": 98150 }, { "epoch": 0.5921298585400562, "grad_norm": 1.10807466506958, "learning_rate": 0.0002950240923043914, "loss": 5.8932, "step": 98200 }, { "epoch": 0.5924313503213903, "grad_norm": 0.8729186058044434, "learning_rate": 0.0002950188932129179, "loss": 5.4595, "step": 98250 }, { "epoch": 0.5927328421027243, "grad_norm": 1.0102006196975708, "learning_rate": 0.00029501369145257863, "loss": 4.9219, "step": 98300 }, { "epoch": 0.5930343338840584, "grad_norm": 1.8745781183242798, "learning_rate": 0.0002950084870234692, "loss": 4.7782, "step": 98350 }, { "epoch": 0.5933358256653923, "grad_norm": 0.886248767375946, "learning_rate": 0.00029500327992568567, "loss": 5.1367, "step": 98400 }, { "epoch": 0.5936373174467264, "grad_norm": 1.531010627746582, "learning_rate": 0.0002949980701593236, "loss": 5.4026, "step": 98450 }, { "epoch": 0.5939388092280604, "grad_norm": 0.8846261501312256, "learning_rate": 0.000294992857724479, "loss": 5.0067, "step": 98500 }, { "epoch": 0.5942403010093945, "grad_norm": 1.2546260356903076, "learning_rate": 0.00029498764262124775, "loss": 5.0671, "step": 98550 }, { "epoch": 0.5945417927907285, "grad_norm": 1.0031101703643799, "learning_rate": 0.0002949824248497259, "loss": 5.6237, "step": 98600 }, { "epoch": 0.5948432845720626, "grad_norm": 1.1203680038452148, "learning_rate": 0.00029497720441000935, "loss": 5.5893, "step": 98650 }, { "epoch": 0.5951447763533966, "grad_norm": 1.7577764987945557, "learning_rate": 0.0002949719813021943, "loss": 5.0397, "step": 98700 }, { "epoch": 0.5954462681347307, "grad_norm": 1.0268598794937134, "learning_rate": 0.0002949667555263768, "loss": 5.4616, "step": 98750 }, { "epoch": 0.5957477599160647, "grad_norm": 1.4763633012771606, "learning_rate": 0.000294961527082653, "loss": 5.7123, "step": 98800 }, { "epoch": 0.5960492516973988, "grad_norm": 0.11092682927846909, "learning_rate": 0.0002949562959711193, "loss": 5.7177, "step": 98850 }, { "epoch": 0.5963507434787327, "grad_norm": 1.015729308128357, "learning_rate": 0.00029495106219187166, "loss": 5.3132, "step": 98900 }, { "epoch": 0.5966522352600668, "grad_norm": 1.5098681449890137, "learning_rate": 0.0002949458257450067, "loss": 5.2639, "step": 98950 }, { "epoch": 0.5969537270414008, "grad_norm": 1.1715636253356934, "learning_rate": 0.0002949405866306206, "loss": 5.1183, "step": 99000 }, { "epoch": 0.5969537270414008, "eval_loss": 5.871153831481934, "eval_runtime": 39.2084, "eval_samples_per_second": 13.058, "eval_steps_per_second": 6.529, "eval_tts_loss": 6.910036413944982, "step": 99000 }, { "epoch": 0.5972552188227349, "grad_norm": 1.3624037504196167, "learning_rate": 0.00029493534484880985, "loss": 5.6392, "step": 99050 }, { "epoch": 0.5975567106040689, "grad_norm": 0.9961543679237366, "learning_rate": 0.000294930100399671, "loss": 5.3164, "step": 99100 }, { "epoch": 0.597858202385403, "grad_norm": 0.44448190927505493, "learning_rate": 0.00029492485328330036, "loss": 5.33, "step": 99150 }, { "epoch": 0.598159694166737, "grad_norm": 1.2374356985092163, "learning_rate": 0.00029491960349979467, "loss": 4.92, "step": 99200 }, { "epoch": 0.5984611859480711, "grad_norm": 0.9439703226089478, "learning_rate": 0.0002949143510492505, "loss": 5.5498, "step": 99250 }, { "epoch": 0.5987626777294051, "grad_norm": 1.0528745651245117, "learning_rate": 0.0002949090959317645, "loss": 5.2471, "step": 99300 }, { "epoch": 0.5990641695107392, "grad_norm": 0.9613762497901917, "learning_rate": 0.00029490383814743335, "loss": 5.6521, "step": 99350 }, { "epoch": 0.5993656612920731, "grad_norm": 1.0806288719177246, "learning_rate": 0.0002948985776963539, "loss": 5.2021, "step": 99400 }, { "epoch": 0.5996671530734072, "grad_norm": 1.09367835521698, "learning_rate": 0.0002948933145786228, "loss": 5.0539, "step": 99450 }, { "epoch": 0.5999686448547412, "grad_norm": 1.438496470451355, "learning_rate": 0.00029488804879433715, "loss": 5.2454, "step": 99500 }, { "epoch": 0.6002701366360753, "grad_norm": 1.5350620746612549, "learning_rate": 0.00029488278034359367, "loss": 5.5769, "step": 99550 }, { "epoch": 0.6005716284174093, "grad_norm": 0.9130475521087646, "learning_rate": 0.0002948775092264894, "loss": 5.6568, "step": 99600 }, { "epoch": 0.6008731201987434, "grad_norm": 1.1140917539596558, "learning_rate": 0.00029487223544312133, "loss": 5.3362, "step": 99650 }, { "epoch": 0.6011746119800774, "grad_norm": 1.2920349836349487, "learning_rate": 0.00029486695899358644, "loss": 5.5111, "step": 99700 }, { "epoch": 0.6014761037614115, "grad_norm": 1.4775487184524536, "learning_rate": 0.000294861679877982, "loss": 5.5127, "step": 99750 }, { "epoch": 0.6017775955427455, "grad_norm": 1.384597659111023, "learning_rate": 0.00029485639809640504, "loss": 5.2766, "step": 99800 }, { "epoch": 0.6020790873240796, "grad_norm": 0.8422510027885437, "learning_rate": 0.0002948511136489527, "loss": 5.4348, "step": 99850 }, { "epoch": 0.6023805791054135, "grad_norm": 1.4434524774551392, "learning_rate": 0.0002948458265357224, "loss": 5.6105, "step": 99900 }, { "epoch": 0.6026820708867476, "grad_norm": 1.9398729801177979, "learning_rate": 0.00029484053675681146, "loss": 5.1671, "step": 99950 }, { "epoch": 0.6029835626680816, "grad_norm": 1.071327805519104, "learning_rate": 0.00029483524431231703, "loss": 5.0803, "step": 100000 }, { "epoch": 0.6032850544494157, "grad_norm": 0.8959625363349915, "learning_rate": 0.00029482994920233666, "loss": 5.3816, "step": 100050 }, { "epoch": 0.6035865462307497, "grad_norm": 1.1253230571746826, "learning_rate": 0.0002948246514269677, "loss": 5.7544, "step": 100100 }, { "epoch": 0.6038880380120838, "grad_norm": 0.8450540900230408, "learning_rate": 0.00029481935098630776, "loss": 5.5111, "step": 100150 }, { "epoch": 0.6041895297934178, "grad_norm": 1.1447417736053467, "learning_rate": 0.00029481404788045433, "loss": 5.2279, "step": 100200 }, { "epoch": 0.6044910215747519, "grad_norm": 0.8895000219345093, "learning_rate": 0.000294808742109505, "loss": 5.343, "step": 100250 }, { "epoch": 0.6047925133560859, "grad_norm": 1.2948123216629028, "learning_rate": 0.0002948034336735574, "loss": 5.5942, "step": 100300 }, { "epoch": 0.60509400513742, "grad_norm": 0.5827965140342712, "learning_rate": 0.00029479812257270934, "loss": 5.501, "step": 100350 }, { "epoch": 0.605395496918754, "grad_norm": 0.8438893556594849, "learning_rate": 0.0002947928088070584, "loss": 5.269, "step": 100400 }, { "epoch": 0.605696988700088, "grad_norm": 0.9604769349098206, "learning_rate": 0.0002947874923767024, "loss": 5.531, "step": 100450 }, { "epoch": 0.6059984804814221, "grad_norm": 1.1620982885360718, "learning_rate": 0.0002947821732817393, "loss": 5.256, "step": 100500 }, { "epoch": 0.6062999722627561, "grad_norm": 1.3153866529464722, "learning_rate": 0.0002947768515222669, "loss": 5.4103, "step": 100550 }, { "epoch": 0.6066014640440902, "grad_norm": 0.4437163770198822, "learning_rate": 0.00029477152709838316, "loss": 5.413, "step": 100600 }, { "epoch": 0.6069029558254242, "grad_norm": 0.9449986815452576, "learning_rate": 0.00029476620001018606, "loss": 5.5683, "step": 100650 }, { "epoch": 0.6072044476067583, "grad_norm": 0.9543599486351013, "learning_rate": 0.00029476087025777364, "loss": 5.3406, "step": 100700 }, { "epoch": 0.6075059393880923, "grad_norm": 0.8560019135475159, "learning_rate": 0.000294755537841244, "loss": 5.3653, "step": 100750 }, { "epoch": 0.6078074311694264, "grad_norm": 1.4897048473358154, "learning_rate": 0.0002947502027606952, "loss": 5.7776, "step": 100800 }, { "epoch": 0.6081089229507604, "grad_norm": 1.2098082304000854, "learning_rate": 0.0002947448650162256, "loss": 5.2769, "step": 100850 }, { "epoch": 0.6084104147320945, "grad_norm": 1.4028221368789673, "learning_rate": 0.0002947395246079332, "loss": 5.1368, "step": 100900 }, { "epoch": 0.6087119065134284, "grad_norm": 1.0315817594528198, "learning_rate": 0.00029473418153591645, "loss": 5.357, "step": 100950 }, { "epoch": 0.6090133982947625, "grad_norm": 1.24613618850708, "learning_rate": 0.0002947288358002737, "loss": 5.3474, "step": 101000 }, { "epoch": 0.6093148900760965, "grad_norm": 0.8200045228004456, "learning_rate": 0.0002947234874011032, "loss": 5.3183, "step": 101050 }, { "epoch": 0.6096163818574306, "grad_norm": 0.8820110559463501, "learning_rate": 0.00029471813633850344, "loss": 5.5388, "step": 101100 }, { "epoch": 0.6099178736387646, "grad_norm": 0.9125785827636719, "learning_rate": 0.0002947127826125729, "loss": 5.4085, "step": 101150 }, { "epoch": 0.6102193654200987, "grad_norm": 0.42069241404533386, "learning_rate": 0.00029470742622341015, "loss": 5.473, "step": 101200 }, { "epoch": 0.6105208572014327, "grad_norm": 0.4369208812713623, "learning_rate": 0.0002947020671711137, "loss": 5.2152, "step": 101250 }, { "epoch": 0.6108223489827668, "grad_norm": 1.3665826320648193, "learning_rate": 0.0002946967054557822, "loss": 5.6071, "step": 101300 }, { "epoch": 0.6111238407641008, "grad_norm": 1.0695289373397827, "learning_rate": 0.00029469134107751435, "loss": 5.0256, "step": 101350 }, { "epoch": 0.6114253325454349, "grad_norm": 1.3678010702133179, "learning_rate": 0.00029468597403640886, "loss": 5.5868, "step": 101400 }, { "epoch": 0.6117268243267688, "grad_norm": 1.2144875526428223, "learning_rate": 0.0002946806043325645, "loss": 5.422, "step": 101450 }, { "epoch": 0.6120283161081029, "grad_norm": 0.6514288187026978, "learning_rate": 0.00029467523196608007, "loss": 5.6502, "step": 101500 }, { "epoch": 0.6123298078894369, "grad_norm": 0.5082715749740601, "learning_rate": 0.00029466985693705447, "loss": 5.6279, "step": 101550 }, { "epoch": 0.612631299670771, "grad_norm": 2.1525208950042725, "learning_rate": 0.0002946644792455866, "loss": 5.0351, "step": 101600 }, { "epoch": 0.612932791452105, "grad_norm": 0.935299813747406, "learning_rate": 0.0002946590988917754, "loss": 5.4467, "step": 101650 }, { "epoch": 0.6132342832334391, "grad_norm": 2.1440415382385254, "learning_rate": 0.00029465371587571995, "loss": 5.0503, "step": 101700 }, { "epoch": 0.6135357750147731, "grad_norm": 0.924377977848053, "learning_rate": 0.00029464833019751934, "loss": 4.9739, "step": 101750 }, { "epoch": 0.6138372667961072, "grad_norm": 0.978360116481781, "learning_rate": 0.0002946429418572726, "loss": 5.3097, "step": 101800 }, { "epoch": 0.6141387585774412, "grad_norm": 1.1055479049682617, "learning_rate": 0.0002946375508550789, "loss": 5.3424, "step": 101850 }, { "epoch": 0.6144402503587753, "grad_norm": 0.979363203048706, "learning_rate": 0.00029463215719103755, "loss": 5.6337, "step": 101900 }, { "epoch": 0.6147417421401092, "grad_norm": 0.9989897608757019, "learning_rate": 0.0002946267608652477, "loss": 5.3701, "step": 101950 }, { "epoch": 0.6150432339214433, "grad_norm": 0.9485950469970703, "learning_rate": 0.0002946213618778087, "loss": 5.1576, "step": 102000 }, { "epoch": 0.6150432339214433, "eval_loss": 5.837615966796875, "eval_runtime": 39.0997, "eval_samples_per_second": 13.095, "eval_steps_per_second": 6.547, "eval_tts_loss": 7.049216372017636, "step": 102000 }, { "epoch": 0.6153447257027773, "grad_norm": 1.4530264139175415, "learning_rate": 0.00029461596022882, "loss": 5.2668, "step": 102050 }, { "epoch": 0.6156462174841114, "grad_norm": 1.072216510772705, "learning_rate": 0.00029461055591838083, "loss": 5.633, "step": 102100 }, { "epoch": 0.6159477092654454, "grad_norm": 1.6508305072784424, "learning_rate": 0.0002946051489465908, "loss": 4.9155, "step": 102150 }, { "epoch": 0.6162492010467795, "grad_norm": 1.3397786617279053, "learning_rate": 0.00029459973931354933, "loss": 5.5308, "step": 102200 }, { "epoch": 0.6165506928281135, "grad_norm": 1.463889718055725, "learning_rate": 0.00029459432701935606, "loss": 5.1642, "step": 102250 }, { "epoch": 0.6168521846094476, "grad_norm": 0.9066433310508728, "learning_rate": 0.0002945889120641105, "loss": 5.2799, "step": 102300 }, { "epoch": 0.6171536763907816, "grad_norm": 1.3196994066238403, "learning_rate": 0.0002945834944479124, "loss": 5.2964, "step": 102350 }, { "epoch": 0.6174551681721157, "grad_norm": 1.0029774904251099, "learning_rate": 0.0002945780741708614, "loss": 5.1561, "step": 102400 }, { "epoch": 0.6177566599534496, "grad_norm": 0.9316608905792236, "learning_rate": 0.0002945726512330573, "loss": 5.4109, "step": 102450 }, { "epoch": 0.6180581517347837, "grad_norm": 1.5138548612594604, "learning_rate": 0.00029456722563459984, "loss": 5.1587, "step": 102500 }, { "epoch": 0.6183596435161177, "grad_norm": 0.19019411504268646, "learning_rate": 0.0002945617973755889, "loss": 5.4309, "step": 102550 }, { "epoch": 0.6186611352974518, "grad_norm": 1.113816499710083, "learning_rate": 0.0002945563664561244, "loss": 5.0372, "step": 102600 }, { "epoch": 0.6189626270787858, "grad_norm": 1.3126543760299683, "learning_rate": 0.00029455093287630624, "loss": 5.2173, "step": 102650 }, { "epoch": 0.6192641188601199, "grad_norm": 1.0093809366226196, "learning_rate": 0.0002945454966362344, "loss": 5.386, "step": 102700 }, { "epoch": 0.6195656106414539, "grad_norm": 1.0988709926605225, "learning_rate": 0.00029454005773600904, "loss": 5.6448, "step": 102750 }, { "epoch": 0.619867102422788, "grad_norm": 1.121958613395691, "learning_rate": 0.00029453461617573017, "loss": 4.8971, "step": 102800 }, { "epoch": 0.620168594204122, "grad_norm": 0.779065728187561, "learning_rate": 0.000294529171955498, "loss": 5.6916, "step": 102850 }, { "epoch": 0.620470085985456, "grad_norm": 1.4158509969711304, "learning_rate": 0.00029452372507541257, "loss": 4.7291, "step": 102900 }, { "epoch": 0.62077157776679, "grad_norm": 1.08955717086792, "learning_rate": 0.0002945182755355743, "loss": 4.7432, "step": 102950 }, { "epoch": 0.6210730695481241, "grad_norm": 1.1722379922866821, "learning_rate": 0.0002945128233360834, "loss": 5.3773, "step": 103000 }, { "epoch": 0.6213745613294581, "grad_norm": 1.1684001684188843, "learning_rate": 0.0002945073684770402, "loss": 5.3329, "step": 103050 }, { "epoch": 0.6216760531107922, "grad_norm": 1.3709434270858765, "learning_rate": 0.0002945019109585451, "loss": 5.575, "step": 103100 }, { "epoch": 0.6219775448921262, "grad_norm": 1.6339690685272217, "learning_rate": 0.00029449645078069857, "loss": 5.4434, "step": 103150 }, { "epoch": 0.6222790366734603, "grad_norm": 0.8125072717666626, "learning_rate": 0.00029449098794360107, "loss": 4.824, "step": 103200 }, { "epoch": 0.6225805284547943, "grad_norm": 1.1323151588439941, "learning_rate": 0.0002944855224473531, "loss": 5.343, "step": 103250 }, { "epoch": 0.6228820202361284, "grad_norm": 0.9417325854301453, "learning_rate": 0.00029448005429205535, "loss": 5.2653, "step": 103300 }, { "epoch": 0.6231835120174624, "grad_norm": 1.1227824687957764, "learning_rate": 0.00029447458347780836, "loss": 5.0693, "step": 103350 }, { "epoch": 0.6234850037987965, "grad_norm": 0.8182197213172913, "learning_rate": 0.0002944691100047128, "loss": 5.0026, "step": 103400 }, { "epoch": 0.6237864955801304, "grad_norm": 2.514007329940796, "learning_rate": 0.00029446363387286947, "loss": 5.4643, "step": 103450 }, { "epoch": 0.6240879873614645, "grad_norm": 1.0504732131958008, "learning_rate": 0.0002944581550823791, "loss": 4.9889, "step": 103500 }, { "epoch": 0.6243894791427985, "grad_norm": 0.9030991196632385, "learning_rate": 0.0002944526736333425, "loss": 4.7657, "step": 103550 }, { "epoch": 0.6246909709241326, "grad_norm": 1.7366719245910645, "learning_rate": 0.0002944471895258607, "loss": 5.7515, "step": 103600 }, { "epoch": 0.6249924627054666, "grad_norm": 1.6547585725784302, "learning_rate": 0.0002944417027600344, "loss": 5.2645, "step": 103650 }, { "epoch": 0.6252939544868007, "grad_norm": 1.4568120241165161, "learning_rate": 0.0002944362133359648, "loss": 5.8972, "step": 103700 }, { "epoch": 0.6255954462681347, "grad_norm": 0.7333816289901733, "learning_rate": 0.0002944307212537527, "loss": 5.8376, "step": 103750 }, { "epoch": 0.6258969380494688, "grad_norm": 0.5207763910293579, "learning_rate": 0.0002944252265134994, "loss": 5.5283, "step": 103800 }, { "epoch": 0.6261984298308028, "grad_norm": 1.546959400177002, "learning_rate": 0.0002944197291153058, "loss": 4.9932, "step": 103850 }, { "epoch": 0.6264999216121369, "grad_norm": 0.47081446647644043, "learning_rate": 0.00029441422905927323, "loss": 4.9509, "step": 103900 }, { "epoch": 0.6268014133934708, "grad_norm": 0.8691707253456116, "learning_rate": 0.0002944087263455029, "loss": 5.5963, "step": 103950 }, { "epoch": 0.6271029051748049, "grad_norm": 1.1598635911941528, "learning_rate": 0.00029440322097409606, "loss": 5.4763, "step": 104000 }, { "epoch": 0.6274043969561389, "grad_norm": 1.035084843635559, "learning_rate": 0.00029439771294515393, "loss": 5.5004, "step": 104050 }, { "epoch": 0.627705888737473, "grad_norm": 1.3556524515151978, "learning_rate": 0.000294392202258778, "loss": 5.7922, "step": 104100 }, { "epoch": 0.6280073805188071, "grad_norm": 1.4400566816329956, "learning_rate": 0.00029438668891506963, "loss": 5.5855, "step": 104150 }, { "epoch": 0.6283088723001411, "grad_norm": 1.1551868915557861, "learning_rate": 0.0002943811729141303, "loss": 5.4994, "step": 104200 }, { "epoch": 0.6286103640814752, "grad_norm": 0.8799368143081665, "learning_rate": 0.00029437565425606153, "loss": 5.2443, "step": 104250 }, { "epoch": 0.6289118558628092, "grad_norm": 1.2564563751220703, "learning_rate": 0.0002943701329409649, "loss": 4.9632, "step": 104300 }, { "epoch": 0.6292133476441433, "grad_norm": 0.9393047094345093, "learning_rate": 0.00029436460896894203, "loss": 5.2234, "step": 104350 }, { "epoch": 0.6295148394254773, "grad_norm": 0.36020079255104065, "learning_rate": 0.0002943590823400945, "loss": 5.1485, "step": 104400 }, { "epoch": 0.6298163312068114, "grad_norm": 1.504439353942871, "learning_rate": 0.00029435355305452407, "loss": 5.5962, "step": 104450 }, { "epoch": 0.6301178229881453, "grad_norm": 0.8549618721008301, "learning_rate": 0.0002943480211123325, "loss": 5.4114, "step": 104500 }, { "epoch": 0.6304193147694794, "grad_norm": 0.9423875212669373, "learning_rate": 0.0002943424865136216, "loss": 5.3426, "step": 104550 }, { "epoch": 0.6307208065508134, "grad_norm": 0.9536952376365662, "learning_rate": 0.00029433694925849323, "loss": 5.2532, "step": 104600 }, { "epoch": 0.6310222983321475, "grad_norm": 0.9064618945121765, "learning_rate": 0.0002943314093470493, "loss": 5.502, "step": 104650 }, { "epoch": 0.6313237901134815, "grad_norm": 0.9883925914764404, "learning_rate": 0.00029432586677939176, "loss": 5.4579, "step": 104700 }, { "epoch": 0.6316252818948156, "grad_norm": 1.3707066774368286, "learning_rate": 0.0002943203215556226, "loss": 5.4076, "step": 104750 }, { "epoch": 0.6319267736761496, "grad_norm": 1.3787860870361328, "learning_rate": 0.0002943147736758439, "loss": 5.519, "step": 104800 }, { "epoch": 0.6322282654574837, "grad_norm": 1.8293393850326538, "learning_rate": 0.0002943092231401577, "loss": 4.9517, "step": 104850 }, { "epoch": 0.6325297572388177, "grad_norm": 0.8567297458648682, "learning_rate": 0.00029430366994866614, "loss": 5.5951, "step": 104900 }, { "epoch": 0.6328312490201518, "grad_norm": 0.8741886019706726, "learning_rate": 0.0002942981141014715, "loss": 5.2581, "step": 104950 }, { "epoch": 0.6331327408014857, "grad_norm": 1.0748531818389893, "learning_rate": 0.0002942925555986761, "loss": 4.9761, "step": 105000 }, { "epoch": 0.6331327408014857, "eval_loss": 5.838661193847656, "eval_runtime": 39.1408, "eval_samples_per_second": 13.081, "eval_steps_per_second": 6.54, "eval_tts_loss": 7.008884166485142, "step": 105000 }, { "epoch": 0.6334342325828198, "grad_norm": 1.4654735326766968, "learning_rate": 0.00029428699444038203, "loss": 5.3065, "step": 105050 }, { "epoch": 0.6337357243641538, "grad_norm": 0.644093930721283, "learning_rate": 0.0002942814306266917, "loss": 5.4537, "step": 105100 }, { "epoch": 0.6340372161454879, "grad_norm": 1.2096362113952637, "learning_rate": 0.0002942758641577076, "loss": 5.1066, "step": 105150 }, { "epoch": 0.6343387079268219, "grad_norm": 0.5434061288833618, "learning_rate": 0.0002942702950335321, "loss": 5.7229, "step": 105200 }, { "epoch": 0.634640199708156, "grad_norm": 1.2738076448440552, "learning_rate": 0.00029426472325426776, "loss": 5.6068, "step": 105250 }, { "epoch": 0.63494169148949, "grad_norm": 0.9230553507804871, "learning_rate": 0.000294259148820017, "loss": 5.2462, "step": 105300 }, { "epoch": 0.6352431832708241, "grad_norm": 0.8964812755584717, "learning_rate": 0.0002942535717308825, "loss": 5.352, "step": 105350 }, { "epoch": 0.6355446750521581, "grad_norm": 0.7256632447242737, "learning_rate": 0.0002942479919869669, "loss": 4.8217, "step": 105400 }, { "epoch": 0.6358461668334922, "grad_norm": 1.2034554481506348, "learning_rate": 0.0002942424095883728, "loss": 5.5499, "step": 105450 }, { "epoch": 0.6361476586148261, "grad_norm": 0.849948525428772, "learning_rate": 0.00029423682453520305, "loss": 5.5398, "step": 105500 }, { "epoch": 0.6364491503961602, "grad_norm": 0.9974413514137268, "learning_rate": 0.0002942312368275603, "loss": 5.3788, "step": 105550 }, { "epoch": 0.6367506421774942, "grad_norm": 0.9187911748886108, "learning_rate": 0.00029422564646554755, "loss": 5.2656, "step": 105600 }, { "epoch": 0.6370521339588283, "grad_norm": 0.9361616969108582, "learning_rate": 0.00029422005344926757, "loss": 5.3933, "step": 105650 }, { "epoch": 0.6373536257401623, "grad_norm": 0.7526306509971619, "learning_rate": 0.0002942144577788233, "loss": 5.5599, "step": 105700 }, { "epoch": 0.6376551175214964, "grad_norm": 1.2729439735412598, "learning_rate": 0.00029420885945431773, "loss": 5.6225, "step": 105750 }, { "epoch": 0.6379566093028304, "grad_norm": 1.5185015201568604, "learning_rate": 0.00029420325847585394, "loss": 5.4151, "step": 105800 }, { "epoch": 0.6382581010841645, "grad_norm": 0.9427931308746338, "learning_rate": 0.0002941976548435349, "loss": 5.7009, "step": 105850 }, { "epoch": 0.6385595928654985, "grad_norm": 1.2810614109039307, "learning_rate": 0.0002941920485574639, "loss": 5.2965, "step": 105900 }, { "epoch": 0.6388610846468326, "grad_norm": 1.4346070289611816, "learning_rate": 0.000294186439617744, "loss": 5.4989, "step": 105950 }, { "epoch": 0.6391625764281665, "grad_norm": 1.1879334449768066, "learning_rate": 0.0002941808280244784, "loss": 5.5375, "step": 106000 }, { "epoch": 0.6394640682095006, "grad_norm": 1.1056243181228638, "learning_rate": 0.0002941752137777704, "loss": 5.4197, "step": 106050 }, { "epoch": 0.6397655599908346, "grad_norm": 2.254589319229126, "learning_rate": 0.0002941695968777234, "loss": 5.1458, "step": 106100 }, { "epoch": 0.6400670517721687, "grad_norm": 1.7328230142593384, "learning_rate": 0.0002941639773244407, "loss": 5.2438, "step": 106150 }, { "epoch": 0.6403685435535027, "grad_norm": 1.412644624710083, "learning_rate": 0.00029415835511802574, "loss": 5.6739, "step": 106200 }, { "epoch": 0.6406700353348368, "grad_norm": 1.702065110206604, "learning_rate": 0.0002941527302585819, "loss": 5.4644, "step": 106250 }, { "epoch": 0.6409715271161708, "grad_norm": 0.9631524085998535, "learning_rate": 0.0002941471027462129, "loss": 5.8213, "step": 106300 }, { "epoch": 0.6412730188975049, "grad_norm": 2.0379741191864014, "learning_rate": 0.0002941414725810221, "loss": 5.7705, "step": 106350 }, { "epoch": 0.6415745106788389, "grad_norm": 1.151562213897705, "learning_rate": 0.00029413583976311316, "loss": 5.2191, "step": 106400 }, { "epoch": 0.641876002460173, "grad_norm": 0.9275096654891968, "learning_rate": 0.0002941302042925899, "loss": 5.7824, "step": 106450 }, { "epoch": 0.6421774942415069, "grad_norm": 0.4315049350261688, "learning_rate": 0.00029412456616955584, "loss": 5.3684, "step": 106500 }, { "epoch": 0.642478986022841, "grad_norm": 1.2003899812698364, "learning_rate": 0.0002941189253941148, "loss": 5.6833, "step": 106550 }, { "epoch": 0.642780477804175, "grad_norm": 1.141762137413025, "learning_rate": 0.00029411328196637064, "loss": 5.2051, "step": 106600 }, { "epoch": 0.6430819695855091, "grad_norm": 0.9615454077720642, "learning_rate": 0.00029410763588642717, "loss": 5.4162, "step": 106650 }, { "epoch": 0.6433834613668431, "grad_norm": 2.0407633781433105, "learning_rate": 0.00029410198715438835, "loss": 5.203, "step": 106700 }, { "epoch": 0.6436849531481772, "grad_norm": 1.2247713804244995, "learning_rate": 0.00029409633577035806, "loss": 5.3745, "step": 106750 }, { "epoch": 0.6439864449295112, "grad_norm": 1.2593811750411987, "learning_rate": 0.00029409068173444033, "loss": 5.3275, "step": 106800 }, { "epoch": 0.6442879367108453, "grad_norm": 1.374839186668396, "learning_rate": 0.0002940850250467393, "loss": 5.0872, "step": 106850 }, { "epoch": 0.6445894284921793, "grad_norm": 1.5794621706008911, "learning_rate": 0.00029407936570735897, "loss": 5.4068, "step": 106900 }, { "epoch": 0.6448909202735134, "grad_norm": 1.046600103378296, "learning_rate": 0.00029407370371640346, "loss": 5.4897, "step": 106950 }, { "epoch": 0.6451924120548473, "grad_norm": 1.697604775428772, "learning_rate": 0.00029406803907397716, "loss": 5.7591, "step": 107000 }, { "epoch": 0.6454939038361814, "grad_norm": 1.2174930572509766, "learning_rate": 0.0002940623717801841, "loss": 5.5966, "step": 107050 }, { "epoch": 0.6457953956175154, "grad_norm": 1.0921722650527954, "learning_rate": 0.0002940567018351287, "loss": 5.4037, "step": 107100 }, { "epoch": 0.6460968873988495, "grad_norm": 0.7467524409294128, "learning_rate": 0.0002940510292389153, "loss": 5.3104, "step": 107150 }, { "epoch": 0.6463983791801835, "grad_norm": 1.0404870510101318, "learning_rate": 0.00029404535399164827, "loss": 5.4557, "step": 107200 }, { "epoch": 0.6466998709615176, "grad_norm": 0.885989785194397, "learning_rate": 0.00029403967609343206, "loss": 4.8703, "step": 107250 }, { "epoch": 0.6470013627428516, "grad_norm": 1.3032124042510986, "learning_rate": 0.0002940339955443712, "loss": 5.5593, "step": 107300 }, { "epoch": 0.6473028545241857, "grad_norm": 0.9279430508613586, "learning_rate": 0.0002940283123445702, "loss": 5.8192, "step": 107350 }, { "epoch": 0.6476043463055197, "grad_norm": 1.3227601051330566, "learning_rate": 0.0002940226264941336, "loss": 5.1717, "step": 107400 }, { "epoch": 0.6479058380868538, "grad_norm": 0.9228917956352234, "learning_rate": 0.0002940169379931661, "loss": 5.0039, "step": 107450 }, { "epoch": 0.6482073298681877, "grad_norm": 0.8791333436965942, "learning_rate": 0.0002940112468417724, "loss": 5.3628, "step": 107500 }, { "epoch": 0.6485088216495218, "grad_norm": 1.3814678192138672, "learning_rate": 0.00029400555304005723, "loss": 5.56, "step": 107550 }, { "epoch": 0.6488103134308558, "grad_norm": 1.3440629243850708, "learning_rate": 0.00029399985658812533, "loss": 5.0327, "step": 107600 }, { "epoch": 0.6491118052121899, "grad_norm": 1.4035028219223022, "learning_rate": 0.0002939941574860816, "loss": 5.5498, "step": 107650 }, { "epoch": 0.6494132969935239, "grad_norm": 1.0796725749969482, "learning_rate": 0.0002939884557340309, "loss": 5.3931, "step": 107700 }, { "epoch": 0.649714788774858, "grad_norm": 1.5110042095184326, "learning_rate": 0.0002939827513320781, "loss": 5.0295, "step": 107750 }, { "epoch": 0.650016280556192, "grad_norm": 1.5627739429473877, "learning_rate": 0.0002939770442803282, "loss": 5.3163, "step": 107800 }, { "epoch": 0.6503177723375261, "grad_norm": 1.2136458158493042, "learning_rate": 0.0002939713345788864, "loss": 5.2764, "step": 107850 }, { "epoch": 0.6506192641188602, "grad_norm": 1.1079142093658447, "learning_rate": 0.00029396562222785754, "loss": 5.371, "step": 107900 }, { "epoch": 0.6509207559001942, "grad_norm": 1.3020681142807007, "learning_rate": 0.0002939599072273469, "loss": 5.4613, "step": 107950 }, { "epoch": 0.6512222476815283, "grad_norm": 0.8654335737228394, "learning_rate": 0.0002939541895774596, "loss": 5.1865, "step": 108000 }, { "epoch": 0.6512222476815283, "eval_loss": 5.802750587463379, "eval_runtime": 39.0001, "eval_samples_per_second": 13.128, "eval_steps_per_second": 6.564, "eval_tts_loss": 7.091845988247731, "step": 108000 }, { "epoch": 0.6515237394628622, "grad_norm": 1.5162659883499146, "learning_rate": 0.0002939484692783009, "loss": 5.327, "step": 108050 }, { "epoch": 0.6518252312441963, "grad_norm": 1.3262367248535156, "learning_rate": 0.000293942746329976, "loss": 5.2716, "step": 108100 }, { "epoch": 0.6521267230255303, "grad_norm": 1.5254184007644653, "learning_rate": 0.00029393702073259026, "loss": 5.3657, "step": 108150 }, { "epoch": 0.6524282148068644, "grad_norm": 1.3480945825576782, "learning_rate": 0.00029393129248624913, "loss": 5.4964, "step": 108200 }, { "epoch": 0.6527297065881984, "grad_norm": 1.1427503824234009, "learning_rate": 0.00029392556159105795, "loss": 5.2839, "step": 108250 }, { "epoch": 0.6530311983695325, "grad_norm": 0.9480688571929932, "learning_rate": 0.00029391982804712215, "loss": 5.2557, "step": 108300 }, { "epoch": 0.6533326901508665, "grad_norm": 1.2235829830169678, "learning_rate": 0.00029391409185454734, "loss": 5.3711, "step": 108350 }, { "epoch": 0.6536341819322006, "grad_norm": 0.5634505152702332, "learning_rate": 0.00029390835301343907, "loss": 5.2139, "step": 108400 }, { "epoch": 0.6539356737135346, "grad_norm": 1.0447509288787842, "learning_rate": 0.0002939026115239028, "loss": 5.3713, "step": 108450 }, { "epoch": 0.6542371654948687, "grad_norm": 1.2542771100997925, "learning_rate": 0.0002938968673860445, "loss": 5.5445, "step": 108500 }, { "epoch": 0.6545386572762026, "grad_norm": 1.3821617364883423, "learning_rate": 0.0002938911205999696, "loss": 5.2188, "step": 108550 }, { "epoch": 0.6548401490575367, "grad_norm": 0.9859095215797424, "learning_rate": 0.000293885371165784, "loss": 5.2196, "step": 108600 }, { "epoch": 0.6551416408388707, "grad_norm": 1.136653184890747, "learning_rate": 0.0002938796190835935, "loss": 5.619, "step": 108650 }, { "epoch": 0.6554431326202048, "grad_norm": 1.4189419746398926, "learning_rate": 0.0002938738643535039, "loss": 5.4932, "step": 108700 }, { "epoch": 0.6557446244015388, "grad_norm": 2.13820481300354, "learning_rate": 0.00029386810697562115, "loss": 5.3012, "step": 108750 }, { "epoch": 0.6560461161828729, "grad_norm": 1.102687954902649, "learning_rate": 0.0002938623469500512, "loss": 5.225, "step": 108800 }, { "epoch": 0.6563476079642069, "grad_norm": 1.4939265251159668, "learning_rate": 0.0002938565842769001, "loss": 5.188, "step": 108850 }, { "epoch": 0.656649099745541, "grad_norm": 0.6716603636741638, "learning_rate": 0.00029385081895627385, "loss": 4.9292, "step": 108900 }, { "epoch": 0.656950591526875, "grad_norm": 1.3351904153823853, "learning_rate": 0.00029384505098827853, "loss": 5.2462, "step": 108950 }, { "epoch": 0.6572520833082091, "grad_norm": 1.0172258615493774, "learning_rate": 0.0002938392803730203, "loss": 5.5237, "step": 109000 }, { "epoch": 0.657553575089543, "grad_norm": 1.87437105178833, "learning_rate": 0.0002938335071106054, "loss": 5.3415, "step": 109050 }, { "epoch": 0.6578550668708771, "grad_norm": 1.1995797157287598, "learning_rate": 0.00029382773120114014, "loss": 5.4389, "step": 109100 }, { "epoch": 0.6581565586522111, "grad_norm": 1.0125476121902466, "learning_rate": 0.00029382195264473067, "loss": 5.5286, "step": 109150 }, { "epoch": 0.6584580504335452, "grad_norm": 0.971403956413269, "learning_rate": 0.0002938161714414834, "loss": 5.236, "step": 109200 }, { "epoch": 0.6587595422148792, "grad_norm": 1.2878447771072388, "learning_rate": 0.0002938103875915048, "loss": 5.5706, "step": 109250 }, { "epoch": 0.6590610339962133, "grad_norm": 1.2402522563934326, "learning_rate": 0.0002938046010949012, "loss": 5.1986, "step": 109300 }, { "epoch": 0.6593625257775473, "grad_norm": 1.7663466930389404, "learning_rate": 0.00029379881195177915, "loss": 5.1929, "step": 109350 }, { "epoch": 0.6596640175588814, "grad_norm": 0.4332813024520874, "learning_rate": 0.0002937930201622452, "loss": 4.8852, "step": 109400 }, { "epoch": 0.6599655093402154, "grad_norm": 1.2577108144760132, "learning_rate": 0.0002937872257264058, "loss": 5.4894, "step": 109450 }, { "epoch": 0.6602670011215495, "grad_norm": 1.2766706943511963, "learning_rate": 0.0002937814286443678, "loss": 4.931, "step": 109500 }, { "epoch": 0.6605684929028834, "grad_norm": 1.6405013799667358, "learning_rate": 0.0002937756289162378, "loss": 5.3198, "step": 109550 }, { "epoch": 0.6608699846842175, "grad_norm": 1.143553614616394, "learning_rate": 0.0002937698265421225, "loss": 5.2752, "step": 109600 }, { "epoch": 0.6611714764655515, "grad_norm": 0.5196934342384338, "learning_rate": 0.0002937640215221287, "loss": 5.1965, "step": 109650 }, { "epoch": 0.6614729682468856, "grad_norm": 1.247558832168579, "learning_rate": 0.0002937582138563633, "loss": 5.3748, "step": 109700 }, { "epoch": 0.6617744600282196, "grad_norm": 1.4457368850708008, "learning_rate": 0.0002937524035449331, "loss": 5.4111, "step": 109750 }, { "epoch": 0.6620759518095537, "grad_norm": 0.9742588996887207, "learning_rate": 0.00029374659058794503, "loss": 5.6138, "step": 109800 }, { "epoch": 0.6623774435908877, "grad_norm": 1.1346439123153687, "learning_rate": 0.0002937407749855061, "loss": 5.0802, "step": 109850 }, { "epoch": 0.6626789353722218, "grad_norm": 1.8148008584976196, "learning_rate": 0.0002937349567377234, "loss": 5.2228, "step": 109900 }, { "epoch": 0.6629804271535558, "grad_norm": 1.436062216758728, "learning_rate": 0.0002937291358447039, "loss": 5.56, "step": 109950 }, { "epoch": 0.6632819189348899, "grad_norm": 0.8456804156303406, "learning_rate": 0.00029372331230655477, "loss": 5.0743, "step": 110000 }, { "epoch": 0.6635834107162238, "grad_norm": 0.9477581977844238, "learning_rate": 0.00029371748612338317, "loss": 5.3923, "step": 110050 }, { "epoch": 0.6638849024975579, "grad_norm": 0.8730908632278442, "learning_rate": 0.0002937116572952963, "loss": 4.908, "step": 110100 }, { "epoch": 0.6641863942788919, "grad_norm": 1.8831430673599243, "learning_rate": 0.00029370582582240157, "loss": 4.7373, "step": 110150 }, { "epoch": 0.664487886060226, "grad_norm": 0.9566383957862854, "learning_rate": 0.00029369999170480613, "loss": 4.8769, "step": 110200 }, { "epoch": 0.66478937784156, "grad_norm": 0.9587188363075256, "learning_rate": 0.0002936941549426174, "loss": 5.5774, "step": 110250 }, { "epoch": 0.6650908696228941, "grad_norm": 1.029618740081787, "learning_rate": 0.0002936883155359428, "loss": 5.1534, "step": 110300 }, { "epoch": 0.6653923614042281, "grad_norm": 1.053440809249878, "learning_rate": 0.00029368247348488983, "loss": 5.6768, "step": 110350 }, { "epoch": 0.6656938531855622, "grad_norm": 1.2937761545181274, "learning_rate": 0.00029367662878956597, "loss": 5.2018, "step": 110400 }, { "epoch": 0.6659953449668962, "grad_norm": 1.225879430770874, "learning_rate": 0.00029367078145007885, "loss": 5.7887, "step": 110450 }, { "epoch": 0.6662968367482303, "grad_norm": 0.5256618857383728, "learning_rate": 0.000293664931466536, "loss": 5.287, "step": 110500 }, { "epoch": 0.6665983285295642, "grad_norm": 1.609654188156128, "learning_rate": 0.0002936590788390451, "loss": 5.6439, "step": 110550 }, { "epoch": 0.6668998203108983, "grad_norm": 1.2024414539337158, "learning_rate": 0.0002936532235677138, "loss": 5.5027, "step": 110600 }, { "epoch": 0.6672013120922323, "grad_norm": 1.0876377820968628, "learning_rate": 0.00029364736565265, "loss": 5.4104, "step": 110650 }, { "epoch": 0.6675028038735664, "grad_norm": 1.7916916608810425, "learning_rate": 0.0002936415050939614, "loss": 5.3501, "step": 110700 }, { "epoch": 0.6678042956549004, "grad_norm": 1.4585905075073242, "learning_rate": 0.00029363564189175594, "loss": 5.0417, "step": 110750 }, { "epoch": 0.6681057874362345, "grad_norm": 1.5228122472763062, "learning_rate": 0.0002936297760461414, "loss": 5.1934, "step": 110800 }, { "epoch": 0.6684072792175685, "grad_norm": 1.9883520603179932, "learning_rate": 0.0002936239075572259, "loss": 4.9673, "step": 110850 }, { "epoch": 0.6687087709989026, "grad_norm": 0.936138391494751, "learning_rate": 0.00029361803642511726, "loss": 5.2868, "step": 110900 }, { "epoch": 0.6690102627802366, "grad_norm": 1.2656031847000122, "learning_rate": 0.00029361216264992365, "loss": 5.2795, "step": 110950 }, { "epoch": 0.6693117545615707, "grad_norm": 1.0060670375823975, "learning_rate": 0.0002936062862317531, "loss": 5.4645, "step": 111000 }, { "epoch": 0.6693117545615707, "eval_loss": 5.800556182861328, "eval_runtime": 39.1606, "eval_samples_per_second": 13.074, "eval_steps_per_second": 6.537, "eval_tts_loss": 7.027178487819165, "step": 111000 }, { "epoch": 0.6696132463429046, "grad_norm": 0.81554114818573, "learning_rate": 0.0002936004071707138, "loss": 5.5546, "step": 111050 }, { "epoch": 0.6699147381242387, "grad_norm": 1.1335926055908203, "learning_rate": 0.0002935945254669139, "loss": 5.3478, "step": 111100 }, { "epoch": 0.6702162299055727, "grad_norm": 0.9967849850654602, "learning_rate": 0.00029358864112046174, "loss": 5.1235, "step": 111150 }, { "epoch": 0.6705177216869068, "grad_norm": 2.0477874279022217, "learning_rate": 0.0002935827541314655, "loss": 5.6216, "step": 111200 }, { "epoch": 0.6708192134682408, "grad_norm": 1.0827871561050415, "learning_rate": 0.0002935768645000336, "loss": 5.6094, "step": 111250 }, { "epoch": 0.6711207052495749, "grad_norm": 1.0348451137542725, "learning_rate": 0.0002935709722262744, "loss": 5.2988, "step": 111300 }, { "epoch": 0.6714221970309089, "grad_norm": 0.62703537940979, "learning_rate": 0.0002935650773102964, "loss": 5.0619, "step": 111350 }, { "epoch": 0.671723688812243, "grad_norm": 2.0079939365386963, "learning_rate": 0.00029355917975220795, "loss": 5.2988, "step": 111400 }, { "epoch": 0.672025180593577, "grad_norm": 1.2661044597625732, "learning_rate": 0.0002935532795521177, "loss": 5.1246, "step": 111450 }, { "epoch": 0.6723266723749111, "grad_norm": 1.3821232318878174, "learning_rate": 0.00029354737671013424, "loss": 5.0905, "step": 111500 }, { "epoch": 0.672628164156245, "grad_norm": 1.6021144390106201, "learning_rate": 0.00029354147122636606, "loss": 5.3363, "step": 111550 }, { "epoch": 0.6729296559375791, "grad_norm": 1.63002347946167, "learning_rate": 0.00029353556310092205, "loss": 5.319, "step": 111600 }, { "epoch": 0.6732311477189132, "grad_norm": 1.4814685583114624, "learning_rate": 0.0002935296523339108, "loss": 5.8393, "step": 111650 }, { "epoch": 0.6735326395002472, "grad_norm": 1.200495719909668, "learning_rate": 0.0002935237389254411, "loss": 5.0887, "step": 111700 }, { "epoch": 0.6738341312815813, "grad_norm": 1.286515712738037, "learning_rate": 0.00029351782287562184, "loss": 5.3638, "step": 111750 }, { "epoch": 0.6741356230629153, "grad_norm": 1.0357638597488403, "learning_rate": 0.00029351190418456185, "loss": 5.1623, "step": 111800 }, { "epoch": 0.6744371148442494, "grad_norm": 0.8146662712097168, "learning_rate": 0.0002935059828523701, "loss": 5.2128, "step": 111850 }, { "epoch": 0.6747386066255834, "grad_norm": 0.6003003716468811, "learning_rate": 0.00029350005887915544, "loss": 5.2836, "step": 111900 }, { "epoch": 0.6750400984069175, "grad_norm": 1.3227876424789429, "learning_rate": 0.00029349413226502705, "loss": 4.5207, "step": 111950 }, { "epoch": 0.6753415901882515, "grad_norm": 1.7049016952514648, "learning_rate": 0.00029348820301009393, "loss": 5.1755, "step": 112000 }, { "epoch": 0.6756430819695856, "grad_norm": 0.9242771863937378, "learning_rate": 0.0002934822711144652, "loss": 4.8982, "step": 112050 }, { "epoch": 0.6759445737509195, "grad_norm": 0.9852840304374695, "learning_rate": 0.00029347633657825, "loss": 5.0217, "step": 112100 }, { "epoch": 0.6762460655322536, "grad_norm": 1.7252291440963745, "learning_rate": 0.0002934703994015576, "loss": 5.68, "step": 112150 }, { "epoch": 0.6765475573135876, "grad_norm": 1.4996370077133179, "learning_rate": 0.00029346445958449726, "loss": 5.2827, "step": 112200 }, { "epoch": 0.6768490490949217, "grad_norm": 0.9417099952697754, "learning_rate": 0.00029345851712717823, "loss": 5.251, "step": 112250 }, { "epoch": 0.6771505408762557, "grad_norm": 1.0771760940551758, "learning_rate": 0.00029345257202970993, "loss": 5.3638, "step": 112300 }, { "epoch": 0.6774520326575898, "grad_norm": 0.28955525159835815, "learning_rate": 0.0002934466242922018, "loss": 5.3011, "step": 112350 }, { "epoch": 0.6777535244389238, "grad_norm": 1.0660260915756226, "learning_rate": 0.00029344067391476325, "loss": 5.6528, "step": 112400 }, { "epoch": 0.6780550162202579, "grad_norm": 1.7319215536117554, "learning_rate": 0.00029343472089750377, "loss": 5.316, "step": 112450 }, { "epoch": 0.6783565080015919, "grad_norm": 0.92625492811203, "learning_rate": 0.000293428765240533, "loss": 5.6859, "step": 112500 }, { "epoch": 0.678657999782926, "grad_norm": 1.6690372228622437, "learning_rate": 0.0002934228069439604, "loss": 4.9553, "step": 112550 }, { "epoch": 0.6789594915642599, "grad_norm": 1.9945374727249146, "learning_rate": 0.00029341684600789575, "loss": 5.5554, "step": 112600 }, { "epoch": 0.679260983345594, "grad_norm": 1.5053213834762573, "learning_rate": 0.00029341088243244875, "loss": 5.2553, "step": 112650 }, { "epoch": 0.679562475126928, "grad_norm": 1.4705780744552612, "learning_rate": 0.00029340491621772906, "loss": 4.8753, "step": 112700 }, { "epoch": 0.6798639669082621, "grad_norm": 1.0124378204345703, "learning_rate": 0.0002933989473638466, "loss": 5.3925, "step": 112750 }, { "epoch": 0.6801654586895961, "grad_norm": 1.0874732732772827, "learning_rate": 0.0002933929758709111, "loss": 5.4804, "step": 112800 }, { "epoch": 0.6804669504709302, "grad_norm": 0.9206704497337341, "learning_rate": 0.0002933870017390325, "loss": 5.2169, "step": 112850 }, { "epoch": 0.6807684422522642, "grad_norm": 1.2392936944961548, "learning_rate": 0.00029338102496832087, "loss": 5.6286, "step": 112900 }, { "epoch": 0.6810699340335983, "grad_norm": 1.227021336555481, "learning_rate": 0.00029337504555888594, "loss": 5.2868, "step": 112950 }, { "epoch": 0.6813714258149323, "grad_norm": 1.790723204612732, "learning_rate": 0.000293369063510838, "loss": 5.3106, "step": 113000 }, { "epoch": 0.6816729175962664, "grad_norm": 1.6556979417800903, "learning_rate": 0.00029336307882428704, "loss": 5.4295, "step": 113050 }, { "epoch": 0.6819744093776003, "grad_norm": 1.3921915292739868, "learning_rate": 0.0002933570914993432, "loss": 5.543, "step": 113100 }, { "epoch": 0.6822759011589344, "grad_norm": 0.9902196526527405, "learning_rate": 0.0002933511015361166, "loss": 4.7508, "step": 113150 }, { "epoch": 0.6825773929402684, "grad_norm": 1.2218148708343506, "learning_rate": 0.00029334510893471767, "loss": 5.5973, "step": 113200 }, { "epoch": 0.6828788847216025, "grad_norm": 1.0021979808807373, "learning_rate": 0.0002933391136952565, "loss": 5.3432, "step": 113250 }, { "epoch": 0.6831803765029365, "grad_norm": 0.971460223197937, "learning_rate": 0.0002933331158178435, "loss": 5.4183, "step": 113300 }, { "epoch": 0.6834818682842706, "grad_norm": 1.593544840812683, "learning_rate": 0.00029332711530258906, "loss": 5.4719, "step": 113350 }, { "epoch": 0.6837833600656046, "grad_norm": 1.0569508075714111, "learning_rate": 0.0002933211121496036, "loss": 5.5741, "step": 113400 }, { "epoch": 0.6840848518469387, "grad_norm": 0.9366718530654907, "learning_rate": 0.0002933151063589976, "loss": 5.3867, "step": 113450 }, { "epoch": 0.6843863436282727, "grad_norm": 0.9240245223045349, "learning_rate": 0.0002933090979308816, "loss": 5.1812, "step": 113500 }, { "epoch": 0.6846878354096068, "grad_norm": 1.1990158557891846, "learning_rate": 0.00029330308686536616, "loss": 5.4573, "step": 113550 }, { "epoch": 0.6849893271909407, "grad_norm": 1.0223842859268188, "learning_rate": 0.0002932970731625619, "loss": 5.4375, "step": 113600 }, { "epoch": 0.6852908189722748, "grad_norm": 1.1164838075637817, "learning_rate": 0.0002932910568225795, "loss": 5.1809, "step": 113650 }, { "epoch": 0.6855923107536088, "grad_norm": 0.9501457810401917, "learning_rate": 0.0002932850378455297, "loss": 5.343, "step": 113700 }, { "epoch": 0.6858938025349429, "grad_norm": 1.0440956354141235, "learning_rate": 0.0002932790162315233, "loss": 5.397, "step": 113750 }, { "epoch": 0.6861952943162769, "grad_norm": 1.0059024095535278, "learning_rate": 0.000293272991980671, "loss": 5.5456, "step": 113800 }, { "epoch": 0.686496786097611, "grad_norm": 1.5315430164337158, "learning_rate": 0.0002932669650930838, "loss": 5.309, "step": 113850 }, { "epoch": 0.686798277878945, "grad_norm": 1.041115641593933, "learning_rate": 0.00029326093556887253, "loss": 5.3371, "step": 113900 }, { "epoch": 0.6870997696602791, "grad_norm": 1.1217573881149292, "learning_rate": 0.0002932549034081482, "loss": 5.484, "step": 113950 }, { "epoch": 0.6874012614416131, "grad_norm": 1.1404939889907837, "learning_rate": 0.00029324886861102184, "loss": 5.3841, "step": 114000 }, { "epoch": 0.6874012614416131, "eval_loss": 5.782533645629883, "eval_runtime": 39.2581, "eval_samples_per_second": 13.042, "eval_steps_per_second": 6.521, "eval_tts_loss": 7.130208377832728, "step": 114000 }, { "epoch": 0.6877027532229472, "grad_norm": 0.9036880731582642, "learning_rate": 0.0002932428311776044, "loss": 5.5053, "step": 114050 }, { "epoch": 0.6880042450042811, "grad_norm": 1.1503283977508545, "learning_rate": 0.0002932367911080071, "loss": 5.0379, "step": 114100 }, { "epoch": 0.6883057367856152, "grad_norm": 0.9858829379081726, "learning_rate": 0.00029323074840234114, "loss": 5.5428, "step": 114150 }, { "epoch": 0.6886072285669492, "grad_norm": 1.039884328842163, "learning_rate": 0.0002932247030607176, "loss": 5.2919, "step": 114200 }, { "epoch": 0.6889087203482833, "grad_norm": 0.8371467590332031, "learning_rate": 0.0002932186550832478, "loss": 5.4491, "step": 114250 }, { "epoch": 0.6892102121296173, "grad_norm": 1.2852884531021118, "learning_rate": 0.00029321260447004303, "loss": 5.6381, "step": 114300 }, { "epoch": 0.6895117039109514, "grad_norm": 1.6477490663528442, "learning_rate": 0.0002932065512212146, "loss": 5.2851, "step": 114350 }, { "epoch": 0.6898131956922854, "grad_norm": 1.1860023736953735, "learning_rate": 0.00029320049533687403, "loss": 5.2323, "step": 114400 }, { "epoch": 0.6901146874736195, "grad_norm": 1.4838883876800537, "learning_rate": 0.0002931944368171327, "loss": 5.5523, "step": 114450 }, { "epoch": 0.6904161792549535, "grad_norm": 0.5623582005500793, "learning_rate": 0.00029318837566210206, "loss": 5.1553, "step": 114500 }, { "epoch": 0.6907176710362876, "grad_norm": 0.4229927361011505, "learning_rate": 0.0002931823118718938, "loss": 4.8332, "step": 114550 }, { "epoch": 0.6910191628176215, "grad_norm": 1.4133483171463013, "learning_rate": 0.00029317624544661933, "loss": 5.4856, "step": 114600 }, { "epoch": 0.6913206545989556, "grad_norm": 1.3514316082000732, "learning_rate": 0.0002931701763863904, "loss": 5.3272, "step": 114650 }, { "epoch": 0.6916221463802896, "grad_norm": 0.9428803324699402, "learning_rate": 0.0002931641046913187, "loss": 5.3615, "step": 114700 }, { "epoch": 0.6919236381616237, "grad_norm": 1.0300064086914062, "learning_rate": 0.000293158030361516, "loss": 5.5303, "step": 114750 }, { "epoch": 0.6922251299429577, "grad_norm": 1.4272966384887695, "learning_rate": 0.00029315195339709403, "loss": 5.3765, "step": 114800 }, { "epoch": 0.6925266217242918, "grad_norm": 2.2103679180145264, "learning_rate": 0.0002931458737981646, "loss": 5.373, "step": 114850 }, { "epoch": 0.6928281135056258, "grad_norm": 0.9749315977096558, "learning_rate": 0.00029313979156483965, "loss": 5.5091, "step": 114900 }, { "epoch": 0.6931296052869599, "grad_norm": 6.330508232116699, "learning_rate": 0.00029313370669723117, "loss": 4.8557, "step": 114950 }, { "epoch": 0.6934310970682939, "grad_norm": 1.1209157705307007, "learning_rate": 0.000293127619195451, "loss": 5.0921, "step": 115000 }, { "epoch": 0.693732588849628, "grad_norm": 1.4632904529571533, "learning_rate": 0.0002931215290596113, "loss": 4.8866, "step": 115050 }, { "epoch": 0.694034080630962, "grad_norm": 1.4568432569503784, "learning_rate": 0.0002931154362898241, "loss": 4.9966, "step": 115100 }, { "epoch": 0.694335572412296, "grad_norm": 1.2250001430511475, "learning_rate": 0.00029310934088620154, "loss": 5.3318, "step": 115150 }, { "epoch": 0.69463706419363, "grad_norm": 1.5111027956008911, "learning_rate": 0.0002931032428488558, "loss": 5.5335, "step": 115200 }, { "epoch": 0.6949385559749641, "grad_norm": 1.5642250776290894, "learning_rate": 0.0002930971421778991, "loss": 5.6352, "step": 115250 }, { "epoch": 0.6952400477562982, "grad_norm": 1.7766767740249634, "learning_rate": 0.0002930910388734437, "loss": 5.5406, "step": 115300 }, { "epoch": 0.6955415395376322, "grad_norm": 1.5745385885238647, "learning_rate": 0.00029308493293560186, "loss": 5.3709, "step": 115350 }, { "epoch": 0.6958430313189663, "grad_norm": 0.6794758439064026, "learning_rate": 0.00029307882436448607, "loss": 5.5158, "step": 115400 }, { "epoch": 0.6961445231003003, "grad_norm": 1.2570713758468628, "learning_rate": 0.0002930727131602087, "loss": 5.357, "step": 115450 }, { "epoch": 0.6964460148816344, "grad_norm": 1.1869735717773438, "learning_rate": 0.00029306659932288226, "loss": 4.853, "step": 115500 }, { "epoch": 0.6967475066629684, "grad_norm": 1.4238816499710083, "learning_rate": 0.00029306048285261915, "loss": 4.77, "step": 115550 }, { "epoch": 0.6970489984443025, "grad_norm": 1.5082730054855347, "learning_rate": 0.00029305436374953207, "loss": 5.1886, "step": 115600 }, { "epoch": 0.6973504902256364, "grad_norm": 0.9561188220977783, "learning_rate": 0.00029304824201373356, "loss": 5.2097, "step": 115650 }, { "epoch": 0.6976519820069705, "grad_norm": 1.1249417066574097, "learning_rate": 0.0002930421176453363, "loss": 5.5906, "step": 115700 }, { "epoch": 0.6979534737883045, "grad_norm": 1.592322826385498, "learning_rate": 0.000293035990644453, "loss": 4.9561, "step": 115750 }, { "epoch": 0.6982549655696386, "grad_norm": 1.9610008001327515, "learning_rate": 0.0002930298610111964, "loss": 5.394, "step": 115800 }, { "epoch": 0.6985564573509726, "grad_norm": 1.7756117582321167, "learning_rate": 0.00029302372874567936, "loss": 5.0673, "step": 115850 }, { "epoch": 0.6988579491323067, "grad_norm": 2.1594724655151367, "learning_rate": 0.00029301759384801467, "loss": 5.1505, "step": 115900 }, { "epoch": 0.6991594409136407, "grad_norm": 0.9788317084312439, "learning_rate": 0.00029301145631831525, "loss": 5.2351, "step": 115950 }, { "epoch": 0.6994609326949748, "grad_norm": 1.3208895921707153, "learning_rate": 0.0002930053161566941, "loss": 5.3973, "step": 116000 }, { "epoch": 0.6997624244763088, "grad_norm": 1.46024751663208, "learning_rate": 0.0002929991733632641, "loss": 5.3509, "step": 116050 }, { "epoch": 0.7000639162576429, "grad_norm": 0.9031099677085876, "learning_rate": 0.00029299302793813853, "loss": 5.4839, "step": 116100 }, { "epoch": 0.7003654080389768, "grad_norm": 1.3940480947494507, "learning_rate": 0.0002929868798814302, "loss": 5.2559, "step": 116150 }, { "epoch": 0.7006668998203109, "grad_norm": 1.0044900178909302, "learning_rate": 0.00029298072919325247, "loss": 5.3677, "step": 116200 }, { "epoch": 0.7009683916016449, "grad_norm": 1.3397228717803955, "learning_rate": 0.0002929745758737185, "loss": 5.3004, "step": 116250 }, { "epoch": 0.701269883382979, "grad_norm": 1.785234808921814, "learning_rate": 0.0002929684199229414, "loss": 5.4043, "step": 116300 }, { "epoch": 0.701571375164313, "grad_norm": 0.9393771290779114, "learning_rate": 0.00029296226134103466, "loss": 5.6023, "step": 116350 }, { "epoch": 0.7018728669456471, "grad_norm": 1.1164199113845825, "learning_rate": 0.00029295610012811146, "loss": 5.3488, "step": 116400 }, { "epoch": 0.7021743587269811, "grad_norm": 1.3793373107910156, "learning_rate": 0.00029294993628428527, "loss": 5.0373, "step": 116450 }, { "epoch": 0.7024758505083152, "grad_norm": 1.2243084907531738, "learning_rate": 0.00029294376980966945, "loss": 5.0238, "step": 116500 }, { "epoch": 0.7027773422896492, "grad_norm": 1.9029377698898315, "learning_rate": 0.0002929376007043776, "loss": 5.3004, "step": 116550 }, { "epoch": 0.7030788340709833, "grad_norm": 0.8713666200637817, "learning_rate": 0.00029293142896852316, "loss": 5.042, "step": 116600 }, { "epoch": 0.7033803258523172, "grad_norm": 1.4589729309082031, "learning_rate": 0.00029292525460221976, "loss": 5.1952, "step": 116650 }, { "epoch": 0.7036818176336513, "grad_norm": 1.7465118169784546, "learning_rate": 0.00029291907760558097, "loss": 5.4491, "step": 116700 }, { "epoch": 0.7039833094149853, "grad_norm": 1.237581491470337, "learning_rate": 0.00029291289797872056, "loss": 5.4008, "step": 116750 }, { "epoch": 0.7042848011963194, "grad_norm": 0.8625478148460388, "learning_rate": 0.0002929067157217522, "loss": 5.2514, "step": 116800 }, { "epoch": 0.7045862929776534, "grad_norm": 1.2371209859848022, "learning_rate": 0.00029290053083478966, "loss": 4.9914, "step": 116850 }, { "epoch": 0.7048877847589875, "grad_norm": 1.3941978216171265, "learning_rate": 0.0002928943433179468, "loss": 5.5146, "step": 116900 }, { "epoch": 0.7051892765403215, "grad_norm": 1.36220121383667, "learning_rate": 0.0002928881531713375, "loss": 5.192, "step": 116950 }, { "epoch": 0.7054907683216556, "grad_norm": 1.1150002479553223, "learning_rate": 0.0002928819603950756, "loss": 5.3363, "step": 117000 }, { "epoch": 0.7054907683216556, "eval_loss": 5.783914566040039, "eval_runtime": 38.9921, "eval_samples_per_second": 13.131, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.106632197712685, "step": 117000 }, { "epoch": 0.7057922601029896, "grad_norm": 0.6230214834213257, "learning_rate": 0.0002928757649892751, "loss": 4.9384, "step": 117050 }, { "epoch": 0.7060937518843237, "grad_norm": 0.868767261505127, "learning_rate": 0.00029286956695405013, "loss": 5.4761, "step": 117100 }, { "epoch": 0.7063952436656576, "grad_norm": 1.5441124439239502, "learning_rate": 0.00029286336628951465, "loss": 5.7226, "step": 117150 }, { "epoch": 0.7066967354469917, "grad_norm": 1.0519450902938843, "learning_rate": 0.00029285716299578277, "loss": 5.7016, "step": 117200 }, { "epoch": 0.7069982272283257, "grad_norm": 1.2417021989822388, "learning_rate": 0.0002928509570729686, "loss": 5.6327, "step": 117250 }, { "epoch": 0.7072997190096598, "grad_norm": 1.1467745304107666, "learning_rate": 0.0002928447485211865, "loss": 5.2334, "step": 117300 }, { "epoch": 0.7076012107909938, "grad_norm": 0.9844264984130859, "learning_rate": 0.00029283853734055065, "loss": 5.5488, "step": 117350 }, { "epoch": 0.7079027025723279, "grad_norm": 1.1669011116027832, "learning_rate": 0.00029283232353117535, "loss": 5.3389, "step": 117400 }, { "epoch": 0.7082041943536619, "grad_norm": 1.2191689014434814, "learning_rate": 0.000292826107093175, "loss": 5.4005, "step": 117450 }, { "epoch": 0.708505686134996, "grad_norm": 1.3950763940811157, "learning_rate": 0.0002928198880266639, "loss": 5.2771, "step": 117500 }, { "epoch": 0.70880717791633, "grad_norm": 1.0194323062896729, "learning_rate": 0.00029281366633175665, "loss": 5.6006, "step": 117550 }, { "epoch": 0.7091086696976641, "grad_norm": 2.369945764541626, "learning_rate": 0.0002928074420085676, "loss": 5.1446, "step": 117600 }, { "epoch": 0.709410161478998, "grad_norm": 0.7221342325210571, "learning_rate": 0.00029280121505721144, "loss": 5.3335, "step": 117650 }, { "epoch": 0.7097116532603321, "grad_norm": 0.2488994151353836, "learning_rate": 0.0002927949854778027, "loss": 5.2018, "step": 117700 }, { "epoch": 0.7100131450416661, "grad_norm": 1.5907676219940186, "learning_rate": 0.00029278875327045604, "loss": 5.7848, "step": 117750 }, { "epoch": 0.7103146368230002, "grad_norm": 1.4025719165802002, "learning_rate": 0.0002927825184352861, "loss": 5.367, "step": 117800 }, { "epoch": 0.7106161286043342, "grad_norm": 0.8956763744354248, "learning_rate": 0.0002927762809724077, "loss": 5.5135, "step": 117850 }, { "epoch": 0.7109176203856683, "grad_norm": 1.0475412607192993, "learning_rate": 0.0002927700408819356, "loss": 5.3057, "step": 117900 }, { "epoch": 0.7112191121670023, "grad_norm": 1.1489057540893555, "learning_rate": 0.00029276379816398464, "loss": 5.3156, "step": 117950 }, { "epoch": 0.7115206039483364, "grad_norm": 1.275122046470642, "learning_rate": 0.00029275755281866974, "loss": 5.375, "step": 118000 }, { "epoch": 0.7118220957296704, "grad_norm": 0.9895021915435791, "learning_rate": 0.0002927513048461057, "loss": 5.1514, "step": 118050 }, { "epoch": 0.7121235875110045, "grad_norm": 0.8610686659812927, "learning_rate": 0.00029274505424640773, "loss": 5.5892, "step": 118100 }, { "epoch": 0.7124250792923384, "grad_norm": 1.1309982538223267, "learning_rate": 0.0002927388010196907, "loss": 5.4563, "step": 118150 }, { "epoch": 0.7127265710736725, "grad_norm": 1.7910879850387573, "learning_rate": 0.00029273254516606976, "loss": 5.4142, "step": 118200 }, { "epoch": 0.7130280628550065, "grad_norm": 1.0980029106140137, "learning_rate": 0.00029272628668565996, "loss": 5.2533, "step": 118250 }, { "epoch": 0.7133295546363406, "grad_norm": 1.292504906654358, "learning_rate": 0.0002927200255785766, "loss": 5.4312, "step": 118300 }, { "epoch": 0.7136310464176746, "grad_norm": 1.3145530223846436, "learning_rate": 0.0002927137618449348, "loss": 5.4348, "step": 118350 }, { "epoch": 0.7139325381990087, "grad_norm": 1.0059853792190552, "learning_rate": 0.0002927074954848499, "loss": 5.6554, "step": 118400 }, { "epoch": 0.7142340299803427, "grad_norm": 0.3631850779056549, "learning_rate": 0.00029270122649843717, "loss": 5.2738, "step": 118450 }, { "epoch": 0.7145355217616768, "grad_norm": 1.5122346878051758, "learning_rate": 0.00029269495488581205, "loss": 5.2164, "step": 118500 }, { "epoch": 0.7148370135430108, "grad_norm": 0.949828565120697, "learning_rate": 0.00029268868064708984, "loss": 5.2808, "step": 118550 }, { "epoch": 0.7151385053243449, "grad_norm": 1.813415765762329, "learning_rate": 0.00029268240378238616, "loss": 5.3443, "step": 118600 }, { "epoch": 0.7154399971056788, "grad_norm": 0.5495845675468445, "learning_rate": 0.00029267612429181644, "loss": 4.9306, "step": 118650 }, { "epoch": 0.7157414888870129, "grad_norm": 2.214021682739258, "learning_rate": 0.0002926698421754963, "loss": 5.1842, "step": 118700 }, { "epoch": 0.7160429806683469, "grad_norm": 1.2394436597824097, "learning_rate": 0.0002926635574335413, "loss": 5.034, "step": 118750 }, { "epoch": 0.716344472449681, "grad_norm": 1.7565866708755493, "learning_rate": 0.0002926572700660671, "loss": 5.6064, "step": 118800 }, { "epoch": 0.716645964231015, "grad_norm": 1.6694667339324951, "learning_rate": 0.0002926509800731894, "loss": 5.0659, "step": 118850 }, { "epoch": 0.7169474560123491, "grad_norm": 1.1073501110076904, "learning_rate": 0.00029264468745502406, "loss": 5.2194, "step": 118900 }, { "epoch": 0.7172489477936831, "grad_norm": 0.9244844317436218, "learning_rate": 0.00029263839221168675, "loss": 5.2011, "step": 118950 }, { "epoch": 0.7175504395750172, "grad_norm": 1.7583943605422974, "learning_rate": 0.0002926320943432934, "loss": 5.6111, "step": 119000 }, { "epoch": 0.7178519313563513, "grad_norm": 1.2643414735794067, "learning_rate": 0.0002926257938499599, "loss": 5.1711, "step": 119050 }, { "epoch": 0.7181534231376853, "grad_norm": 1.2832531929016113, "learning_rate": 0.00029261949073180224, "loss": 4.8972, "step": 119100 }, { "epoch": 0.7184549149190194, "grad_norm": 0.8706262111663818, "learning_rate": 0.00029261318498893633, "loss": 5.2849, "step": 119150 }, { "epoch": 0.7187564067003533, "grad_norm": 0.9810001254081726, "learning_rate": 0.0002926068766214783, "loss": 5.5236, "step": 119200 }, { "epoch": 0.7190578984816874, "grad_norm": 1.555525779724121, "learning_rate": 0.0002926005656295442, "loss": 5.0459, "step": 119250 }, { "epoch": 0.7193593902630214, "grad_norm": 2.04788875579834, "learning_rate": 0.00029259425201325025, "loss": 5.4046, "step": 119300 }, { "epoch": 0.7196608820443555, "grad_norm": 1.3085627555847168, "learning_rate": 0.0002925879357727125, "loss": 5.3075, "step": 119350 }, { "epoch": 0.7199623738256895, "grad_norm": 1.0839691162109375, "learning_rate": 0.0002925816169080473, "loss": 5.9432, "step": 119400 }, { "epoch": 0.7202638656070236, "grad_norm": 1.289135456085205, "learning_rate": 0.00029257529541937093, "loss": 5.2531, "step": 119450 }, { "epoch": 0.7205653573883576, "grad_norm": 1.2894569635391235, "learning_rate": 0.0002925689713067997, "loss": 5.8182, "step": 119500 }, { "epoch": 0.7208668491696917, "grad_norm": 0.9531562924385071, "learning_rate": 0.00029256264457045003, "loss": 5.3695, "step": 119550 }, { "epoch": 0.7211683409510257, "grad_norm": 1.0052783489227295, "learning_rate": 0.00029255631521043834, "loss": 5.4616, "step": 119600 }, { "epoch": 0.7214698327323598, "grad_norm": 1.346319556236267, "learning_rate": 0.00029254998322688106, "loss": 5.603, "step": 119650 }, { "epoch": 0.7217713245136937, "grad_norm": 0.5765519738197327, "learning_rate": 0.0002925436486198948, "loss": 4.9614, "step": 119700 }, { "epoch": 0.7220728162950278, "grad_norm": 1.7217450141906738, "learning_rate": 0.0002925373113895961, "loss": 5.0632, "step": 119750 }, { "epoch": 0.7223743080763618, "grad_norm": 1.5318177938461304, "learning_rate": 0.0002925309715361016, "loss": 5.6192, "step": 119800 }, { "epoch": 0.7226757998576959, "grad_norm": 1.342105507850647, "learning_rate": 0.00029252462905952805, "loss": 5.4232, "step": 119850 }, { "epoch": 0.7229772916390299, "grad_norm": 0.389747679233551, "learning_rate": 0.000292518283959992, "loss": 5.5112, "step": 119900 }, { "epoch": 0.723278783420364, "grad_norm": 0.9597907066345215, "learning_rate": 0.0002925119362376103, "loss": 5.192, "step": 119950 }, { "epoch": 0.723580275201698, "grad_norm": 1.6688579320907593, "learning_rate": 0.0002925055858924998, "loss": 5.0781, "step": 120000 }, { "epoch": 0.723580275201698, "eval_loss": 5.760333061218262, "eval_runtime": 38.8406, "eval_samples_per_second": 13.182, "eval_steps_per_second": 6.591, "eval_tts_loss": 7.022141989513127, "step": 120000 }, { "epoch": 0.7238817669830321, "grad_norm": 0.5470027923583984, "learning_rate": 0.00029249923292477744, "loss": 4.718, "step": 120050 }, { "epoch": 0.7241832587643661, "grad_norm": 1.0218124389648438, "learning_rate": 0.00029249287733455997, "loss": 5.5185, "step": 120100 }, { "epoch": 0.7244847505457002, "grad_norm": 1.641676664352417, "learning_rate": 0.0002924865191219645, "loss": 4.9969, "step": 120150 }, { "epoch": 0.7247862423270341, "grad_norm": 1.3578218221664429, "learning_rate": 0.00029248015828710794, "loss": 5.2854, "step": 120200 }, { "epoch": 0.7250877341083682, "grad_norm": 1.7651782035827637, "learning_rate": 0.0002924737948301074, "loss": 5.1834, "step": 120250 }, { "epoch": 0.7253892258897022, "grad_norm": 1.5566399097442627, "learning_rate": 0.00029246742875108, "loss": 4.8992, "step": 120300 }, { "epoch": 0.7256907176710363, "grad_norm": 1.537219524383545, "learning_rate": 0.0002924610600501429, "loss": 5.2877, "step": 120350 }, { "epoch": 0.7259922094523703, "grad_norm": 1.2298606634140015, "learning_rate": 0.00029245468872741325, "loss": 5.3516, "step": 120400 }, { "epoch": 0.7262937012337044, "grad_norm": 1.2801377773284912, "learning_rate": 0.0002924483147830084, "loss": 5.2718, "step": 120450 }, { "epoch": 0.7265951930150384, "grad_norm": 1.173919439315796, "learning_rate": 0.0002924419382170456, "loss": 5.5481, "step": 120500 }, { "epoch": 0.7268966847963725, "grad_norm": 1.1231608390808105, "learning_rate": 0.00029243555902964215, "loss": 5.3057, "step": 120550 }, { "epoch": 0.7271981765777065, "grad_norm": 2.3533570766448975, "learning_rate": 0.00029242917722091554, "loss": 5.4691, "step": 120600 }, { "epoch": 0.7274996683590406, "grad_norm": 1.5801122188568115, "learning_rate": 0.00029242279279098317, "loss": 5.3372, "step": 120650 }, { "epoch": 0.7278011601403745, "grad_norm": 1.55729079246521, "learning_rate": 0.00029241640573996256, "loss": 5.0834, "step": 120700 }, { "epoch": 0.7281026519217086, "grad_norm": 1.9559659957885742, "learning_rate": 0.00029241001606797123, "loss": 5.3382, "step": 120750 }, { "epoch": 0.7284041437030426, "grad_norm": 1.3904250860214233, "learning_rate": 0.00029240362377512676, "loss": 5.0861, "step": 120800 }, { "epoch": 0.7287056354843767, "grad_norm": 0.6401329636573792, "learning_rate": 0.00029239722886154686, "loss": 5.4659, "step": 120850 }, { "epoch": 0.7290071272657107, "grad_norm": 0.8864406943321228, "learning_rate": 0.00029239083132734915, "loss": 5.4195, "step": 120900 }, { "epoch": 0.7293086190470448, "grad_norm": 1.1812316179275513, "learning_rate": 0.00029238443117265143, "loss": 5.2834, "step": 120950 }, { "epoch": 0.7296101108283788, "grad_norm": 1.2644054889678955, "learning_rate": 0.00029237802839757144, "loss": 5.0572, "step": 121000 }, { "epoch": 0.7299116026097129, "grad_norm": 1.0684508085250854, "learning_rate": 0.00029237162300222694, "loss": 4.8714, "step": 121050 }, { "epoch": 0.7302130943910469, "grad_norm": 1.4618738889694214, "learning_rate": 0.000292365214986736, "loss": 5.3295, "step": 121100 }, { "epoch": 0.730514586172381, "grad_norm": 1.043054223060608, "learning_rate": 0.0002923588043512164, "loss": 5.3645, "step": 121150 }, { "epoch": 0.730816077953715, "grad_norm": 1.5056662559509277, "learning_rate": 0.00029235239109578614, "loss": 4.9449, "step": 121200 }, { "epoch": 0.731117569735049, "grad_norm": 1.1273994445800781, "learning_rate": 0.0002923459752205633, "loss": 4.9209, "step": 121250 }, { "epoch": 0.731419061516383, "grad_norm": 1.831263780593872, "learning_rate": 0.0002923395567256659, "loss": 4.8101, "step": 121300 }, { "epoch": 0.7317205532977171, "grad_norm": 1.7517116069793701, "learning_rate": 0.0002923331356112121, "loss": 5.572, "step": 121350 }, { "epoch": 0.7320220450790511, "grad_norm": 1.1787099838256836, "learning_rate": 0.00029232671187732004, "loss": 5.5009, "step": 121400 }, { "epoch": 0.7323235368603852, "grad_norm": 1.439056158065796, "learning_rate": 0.000292320285524108, "loss": 5.7855, "step": 121450 }, { "epoch": 0.7326250286417192, "grad_norm": 2.1086690425872803, "learning_rate": 0.0002923138565516942, "loss": 5.4969, "step": 121500 }, { "epoch": 0.7329265204230533, "grad_norm": 1.110904574394226, "learning_rate": 0.0002923074249601969, "loss": 5.2162, "step": 121550 }, { "epoch": 0.7332280122043873, "grad_norm": 0.9344980716705322, "learning_rate": 0.0002923009907497346, "loss": 5.2448, "step": 121600 }, { "epoch": 0.7335295039857214, "grad_norm": 0.542750358581543, "learning_rate": 0.00029229455392042566, "loss": 5.5246, "step": 121650 }, { "epoch": 0.7338309957670553, "grad_norm": 1.618652105331421, "learning_rate": 0.00029228811447238847, "loss": 5.346, "step": 121700 }, { "epoch": 0.7341324875483894, "grad_norm": 1.589255452156067, "learning_rate": 0.0002922816724057416, "loss": 5.6454, "step": 121750 }, { "epoch": 0.7344339793297234, "grad_norm": 1.6753673553466797, "learning_rate": 0.00029227522772060354, "loss": 5.3957, "step": 121800 }, { "epoch": 0.7347354711110575, "grad_norm": 1.535239815711975, "learning_rate": 0.000292268780417093, "loss": 5.1479, "step": 121850 }, { "epoch": 0.7350369628923915, "grad_norm": 1.0299286842346191, "learning_rate": 0.0002922623304953286, "loss": 5.2183, "step": 121900 }, { "epoch": 0.7353384546737256, "grad_norm": 1.3087469339370728, "learning_rate": 0.000292255877955429, "loss": 5.0778, "step": 121950 }, { "epoch": 0.7356399464550596, "grad_norm": 2.403214454650879, "learning_rate": 0.000292249422797513, "loss": 5.0104, "step": 122000 }, { "epoch": 0.7359414382363937, "grad_norm": 1.3395968675613403, "learning_rate": 0.0002922429650216994, "loss": 5.3602, "step": 122050 }, { "epoch": 0.7362429300177277, "grad_norm": 1.3669610023498535, "learning_rate": 0.00029223650462810696, "loss": 5.4484, "step": 122100 }, { "epoch": 0.7365444217990618, "grad_norm": 1.4591608047485352, "learning_rate": 0.00029223004161685464, "loss": 5.1405, "step": 122150 }, { "epoch": 0.7368459135803958, "grad_norm": 1.1382830142974854, "learning_rate": 0.00029222357598806143, "loss": 5.6789, "step": 122200 }, { "epoch": 0.7371474053617298, "grad_norm": 1.622544765472412, "learning_rate": 0.0002922171077418462, "loss": 5.4517, "step": 122250 }, { "epoch": 0.7374488971430638, "grad_norm": 1.435128927230835, "learning_rate": 0.0002922106368783281, "loss": 5.3826, "step": 122300 }, { "epoch": 0.7377503889243979, "grad_norm": 0.4929655194282532, "learning_rate": 0.00029220416339762617, "loss": 5.1342, "step": 122350 }, { "epoch": 0.7380518807057319, "grad_norm": 1.1508455276489258, "learning_rate": 0.00029219768729985956, "loss": 5.4486, "step": 122400 }, { "epoch": 0.738353372487066, "grad_norm": 1.0503361225128174, "learning_rate": 0.00029219120858514744, "loss": 5.31, "step": 122450 }, { "epoch": 0.7386548642684, "grad_norm": 1.9129258394241333, "learning_rate": 0.00029218472725360897, "loss": 4.8151, "step": 122500 }, { "epoch": 0.7389563560497341, "grad_norm": 1.345146894454956, "learning_rate": 0.00029217824330536363, "loss": 5.1519, "step": 122550 }, { "epoch": 0.7392578478310681, "grad_norm": 0.9836500883102417, "learning_rate": 0.00029217175674053057, "loss": 5.3818, "step": 122600 }, { "epoch": 0.7395593396124022, "grad_norm": 1.8823966979980469, "learning_rate": 0.00029216526755922916, "loss": 5.9518, "step": 122650 }, { "epoch": 0.7398608313937362, "grad_norm": 1.1576098203659058, "learning_rate": 0.00029215877576157893, "loss": 4.6624, "step": 122700 }, { "epoch": 0.7401623231750702, "grad_norm": 1.7133835554122925, "learning_rate": 0.00029215228134769927, "loss": 5.2619, "step": 122750 }, { "epoch": 0.7404638149564043, "grad_norm": 0.9869288802146912, "learning_rate": 0.0002921457843177098, "loss": 5.0355, "step": 122800 }, { "epoch": 0.7407653067377383, "grad_norm": 1.1489150524139404, "learning_rate": 0.00029213928467172994, "loss": 5.733, "step": 122850 }, { "epoch": 0.7410667985190724, "grad_norm": 0.9827554225921631, "learning_rate": 0.00029213278240987947, "loss": 5.2208, "step": 122900 }, { "epoch": 0.7413682903004064, "grad_norm": 0.9666951894760132, "learning_rate": 0.0002921262775322779, "loss": 4.8856, "step": 122950 }, { "epoch": 0.7416697820817405, "grad_norm": 0.9676090478897095, "learning_rate": 0.0002921197700390451, "loss": 5.1799, "step": 123000 }, { "epoch": 0.7416697820817405, "eval_loss": 5.774966716766357, "eval_runtime": 39.2971, "eval_samples_per_second": 13.029, "eval_steps_per_second": 6.514, "eval_tts_loss": 7.027016280174803, "step": 123000 }, { "epoch": 0.7419712738630745, "grad_norm": 0.7219750285148621, "learning_rate": 0.0002921132599303007, "loss": 5.6956, "step": 123050 }, { "epoch": 0.7422727656444086, "grad_norm": 0.8646533489227295, "learning_rate": 0.00029210674720616455, "loss": 5.4489, "step": 123100 }, { "epoch": 0.7425742574257426, "grad_norm": 0.9746789932250977, "learning_rate": 0.00029210023186675653, "loss": 5.6143, "step": 123150 }, { "epoch": 0.7428757492070767, "grad_norm": 1.099870204925537, "learning_rate": 0.0002920937139121965, "loss": 5.0018, "step": 123200 }, { "epoch": 0.7431772409884106, "grad_norm": 0.32792726159095764, "learning_rate": 0.00029208719334260447, "loss": 5.1299, "step": 123250 }, { "epoch": 0.7434787327697447, "grad_norm": 1.176655650138855, "learning_rate": 0.0002920806701581004, "loss": 5.4675, "step": 123300 }, { "epoch": 0.7437802245510787, "grad_norm": 0.9635154604911804, "learning_rate": 0.0002920741443588044, "loss": 5.078, "step": 123350 }, { "epoch": 0.7440817163324128, "grad_norm": 1.2510781288146973, "learning_rate": 0.0002920676159448365, "loss": 5.0563, "step": 123400 }, { "epoch": 0.7443832081137468, "grad_norm": 1.1285841464996338, "learning_rate": 0.00029206108491631685, "loss": 5.096, "step": 123450 }, { "epoch": 0.7446846998950809, "grad_norm": 2.035322904586792, "learning_rate": 0.0002920545512733657, "loss": 5.6302, "step": 123500 }, { "epoch": 0.7449861916764149, "grad_norm": 1.3202792406082153, "learning_rate": 0.0002920480150161032, "loss": 5.2528, "step": 123550 }, { "epoch": 0.745287683457749, "grad_norm": 1.3965545892715454, "learning_rate": 0.00029204147614464974, "loss": 5.261, "step": 123600 }, { "epoch": 0.745589175239083, "grad_norm": 1.9478204250335693, "learning_rate": 0.00029203493465912564, "loss": 5.1094, "step": 123650 }, { "epoch": 0.7458906670204171, "grad_norm": 0.9324188828468323, "learning_rate": 0.0002920283905596512, "loss": 5.2683, "step": 123700 }, { "epoch": 0.746192158801751, "grad_norm": 1.5904738903045654, "learning_rate": 0.00029202184384634697, "loss": 5.0612, "step": 123750 }, { "epoch": 0.7464936505830851, "grad_norm": 1.185761570930481, "learning_rate": 0.00029201529451933336, "loss": 5.5702, "step": 123800 }, { "epoch": 0.7467951423644191, "grad_norm": 1.607657790184021, "learning_rate": 0.00029200874257873094, "loss": 5.2374, "step": 123850 }, { "epoch": 0.7470966341457532, "grad_norm": 1.403889536857605, "learning_rate": 0.00029200218802466027, "loss": 5.4972, "step": 123900 }, { "epoch": 0.7473981259270872, "grad_norm": 0.8438761234283447, "learning_rate": 0.0002919956308572419, "loss": 5.4499, "step": 123950 }, { "epoch": 0.7476996177084213, "grad_norm": 0.6774864792823792, "learning_rate": 0.00029198907107659663, "loss": 5.3589, "step": 124000 }, { "epoch": 0.7480011094897553, "grad_norm": 1.0508304834365845, "learning_rate": 0.0002919825086828452, "loss": 5.3439, "step": 124050 }, { "epoch": 0.7483026012710894, "grad_norm": 3.2416152954101562, "learning_rate": 0.0002919759436761082, "loss": 5.3205, "step": 124100 }, { "epoch": 0.7486040930524234, "grad_norm": 0.37650370597839355, "learning_rate": 0.0002919693760565067, "loss": 4.9319, "step": 124150 }, { "epoch": 0.7489055848337575, "grad_norm": 1.6140141487121582, "learning_rate": 0.00029196280582416135, "loss": 5.299, "step": 124200 }, { "epoch": 0.7492070766150914, "grad_norm": 1.7330018281936646, "learning_rate": 0.00029195623297919313, "loss": 5.7205, "step": 124250 }, { "epoch": 0.7495085683964255, "grad_norm": 0.9707470536231995, "learning_rate": 0.00029194965752172306, "loss": 5.1062, "step": 124300 }, { "epoch": 0.7498100601777595, "grad_norm": 1.6527899503707886, "learning_rate": 0.0002919430794518721, "loss": 5.7336, "step": 124350 }, { "epoch": 0.7501115519590936, "grad_norm": 0.7092000842094421, "learning_rate": 0.00029193649876976135, "loss": 4.953, "step": 124400 }, { "epoch": 0.7504130437404276, "grad_norm": 1.7770415544509888, "learning_rate": 0.00029192991547551193, "loss": 5.3551, "step": 124450 }, { "epoch": 0.7507145355217617, "grad_norm": 1.2117023468017578, "learning_rate": 0.00029192332956924486, "loss": 5.4654, "step": 124500 }, { "epoch": 0.7510160273030957, "grad_norm": 0.43218451738357544, "learning_rate": 0.00029191674105108153, "loss": 5.1216, "step": 124550 }, { "epoch": 0.7513175190844298, "grad_norm": 1.0458258390426636, "learning_rate": 0.0002919101499211431, "loss": 5.2722, "step": 124600 }, { "epoch": 0.7516190108657638, "grad_norm": 1.2059657573699951, "learning_rate": 0.00029190355617955086, "loss": 5.5041, "step": 124650 }, { "epoch": 0.7519205026470979, "grad_norm": 1.1412699222564697, "learning_rate": 0.00029189695982642617, "loss": 4.9606, "step": 124700 }, { "epoch": 0.7522219944284319, "grad_norm": 0.8800144195556641, "learning_rate": 0.0002918903608618904, "loss": 4.941, "step": 124750 }, { "epoch": 0.752523486209766, "grad_norm": 0.984514057636261, "learning_rate": 0.00029188375928606506, "loss": 4.8174, "step": 124800 }, { "epoch": 0.7528249779910999, "grad_norm": 0.46969595551490784, "learning_rate": 0.0002918771550990716, "loss": 5.154, "step": 124850 }, { "epoch": 0.753126469772434, "grad_norm": 1.3456989526748657, "learning_rate": 0.0002918705483010316, "loss": 5.2071, "step": 124900 }, { "epoch": 0.753427961553768, "grad_norm": 1.5329527854919434, "learning_rate": 0.0002918639388920666, "loss": 5.258, "step": 124950 }, { "epoch": 0.7537294533351021, "grad_norm": 1.3210960626602173, "learning_rate": 0.00029185732687229824, "loss": 4.8819, "step": 125000 }, { "epoch": 0.7540309451164361, "grad_norm": 1.9118067026138306, "learning_rate": 0.00029185071224184826, "loss": 5.3357, "step": 125050 }, { "epoch": 0.7543324368977702, "grad_norm": 1.2141294479370117, "learning_rate": 0.0002918440950008383, "loss": 5.201, "step": 125100 }, { "epoch": 0.7546339286791042, "grad_norm": 1.5039736032485962, "learning_rate": 0.00029183747514939024, "loss": 5.0262, "step": 125150 }, { "epoch": 0.7549354204604383, "grad_norm": 1.5043437480926514, "learning_rate": 0.00029183085268762583, "loss": 5.3049, "step": 125200 }, { "epoch": 0.7552369122417723, "grad_norm": 0.8196192383766174, "learning_rate": 0.000291824227615667, "loss": 5.2109, "step": 125250 }, { "epoch": 0.7555384040231063, "grad_norm": 0.9377695322036743, "learning_rate": 0.00029181759993363564, "loss": 5.1562, "step": 125300 }, { "epoch": 0.7558398958044403, "grad_norm": 1.4830445051193237, "learning_rate": 0.0002918109696416538, "loss": 5.2183, "step": 125350 }, { "epoch": 0.7561413875857744, "grad_norm": 1.002656102180481, "learning_rate": 0.00029180433673984337, "loss": 5.1446, "step": 125400 }, { "epoch": 0.7564428793671084, "grad_norm": 1.6984821557998657, "learning_rate": 0.0002917977012283265, "loss": 5.5358, "step": 125450 }, { "epoch": 0.7567443711484425, "grad_norm": 1.4692113399505615, "learning_rate": 0.00029179106310722524, "loss": 5.3548, "step": 125500 }, { "epoch": 0.7570458629297765, "grad_norm": 0.5311010479927063, "learning_rate": 0.00029178442237666187, "loss": 5.2753, "step": 125550 }, { "epoch": 0.7573473547111106, "grad_norm": 0.7508612871170044, "learning_rate": 0.00029177777903675856, "loss": 5.445, "step": 125600 }, { "epoch": 0.7576488464924446, "grad_norm": 1.2792001962661743, "learning_rate": 0.00029177113308763756, "loss": 5.3766, "step": 125650 }, { "epoch": 0.7579503382737787, "grad_norm": 1.107495903968811, "learning_rate": 0.00029176448452942116, "loss": 5.0232, "step": 125700 }, { "epoch": 0.7582518300551127, "grad_norm": 1.294872760772705, "learning_rate": 0.00029175783336223167, "loss": 5.3896, "step": 125750 }, { "epoch": 0.7585533218364467, "grad_norm": 1.3931705951690674, "learning_rate": 0.00029175117958619165, "loss": 5.1499, "step": 125800 }, { "epoch": 0.7588548136177807, "grad_norm": 2.103543758392334, "learning_rate": 0.0002917445232014234, "loss": 5.2231, "step": 125850 }, { "epoch": 0.7591563053991148, "grad_norm": 1.0781652927398682, "learning_rate": 0.0002917378642080495, "loss": 5.4591, "step": 125900 }, { "epoch": 0.7594577971804488, "grad_norm": 1.0784703493118286, "learning_rate": 0.0002917312026061925, "loss": 5.4488, "step": 125950 }, { "epoch": 0.7597592889617829, "grad_norm": 0.984164834022522, "learning_rate": 0.0002917245383959749, "loss": 5.0514, "step": 126000 }, { "epoch": 0.7597592889617829, "eval_loss": 5.72927188873291, "eval_runtime": 38.9772, "eval_samples_per_second": 13.136, "eval_steps_per_second": 6.568, "eval_tts_loss": 7.138930143774754, "step": 126000 }, { "epoch": 0.7600607807431169, "grad_norm": 1.4688537120819092, "learning_rate": 0.0002917178715775195, "loss": 5.3647, "step": 126050 }, { "epoch": 0.760362272524451, "grad_norm": 1.0113446712493896, "learning_rate": 0.0002917112021509489, "loss": 5.3305, "step": 126100 }, { "epoch": 0.760663764305785, "grad_norm": 1.5530301332473755, "learning_rate": 0.00029170453011638584, "loss": 5.2656, "step": 126150 }, { "epoch": 0.7609652560871191, "grad_norm": 1.617897391319275, "learning_rate": 0.0002916978554739531, "loss": 5.6926, "step": 126200 }, { "epoch": 0.761266747868453, "grad_norm": 0.980926513671875, "learning_rate": 0.00029169117822377364, "loss": 4.9468, "step": 126250 }, { "epoch": 0.7615682396497871, "grad_norm": 1.0847673416137695, "learning_rate": 0.0002916844983659702, "loss": 5.1175, "step": 126300 }, { "epoch": 0.7618697314311211, "grad_norm": 1.702433705329895, "learning_rate": 0.0002916778159006658, "loss": 5.2531, "step": 126350 }, { "epoch": 0.7621712232124552, "grad_norm": 1.155092477798462, "learning_rate": 0.0002916711308279833, "loss": 5.4704, "step": 126400 }, { "epoch": 0.7624727149937893, "grad_norm": 1.2400072813034058, "learning_rate": 0.00029166444314804584, "loss": 5.4349, "step": 126450 }, { "epoch": 0.7627742067751233, "grad_norm": 0.44214341044425964, "learning_rate": 0.0002916577528609765, "loss": 5.0577, "step": 126500 }, { "epoch": 0.7630756985564574, "grad_norm": 1.1296730041503906, "learning_rate": 0.00029165105996689834, "loss": 4.9239, "step": 126550 }, { "epoch": 0.7633771903377914, "grad_norm": 0.3359615206718445, "learning_rate": 0.00029164436446593457, "loss": 5.1963, "step": 126600 }, { "epoch": 0.7636786821191255, "grad_norm": 0.9969423413276672, "learning_rate": 0.00029163766635820844, "loss": 5.472, "step": 126650 }, { "epoch": 0.7639801739004595, "grad_norm": 1.9051196575164795, "learning_rate": 0.0002916309656438431, "loss": 5.0403, "step": 126700 }, { "epoch": 0.7642816656817936, "grad_norm": 1.597663402557373, "learning_rate": 0.00029162426232296206, "loss": 5.2867, "step": 126750 }, { "epoch": 0.7645831574631275, "grad_norm": 1.1656110286712646, "learning_rate": 0.00029161755639568857, "loss": 5.6162, "step": 126800 }, { "epoch": 0.7648846492444616, "grad_norm": 1.2983297109603882, "learning_rate": 0.00029161084786214603, "loss": 5.5462, "step": 126850 }, { "epoch": 0.7651861410257956, "grad_norm": 2.0357604026794434, "learning_rate": 0.0002916041367224579, "loss": 4.9236, "step": 126900 }, { "epoch": 0.7654876328071297, "grad_norm": 1.7150014638900757, "learning_rate": 0.00029159742297674775, "loss": 5.1338, "step": 126950 }, { "epoch": 0.7657891245884637, "grad_norm": 1.4085897207260132, "learning_rate": 0.0002915907066251391, "loss": 5.4817, "step": 127000 }, { "epoch": 0.7660906163697978, "grad_norm": 3.301474094390869, "learning_rate": 0.0002915839876677555, "loss": 5.2152, "step": 127050 }, { "epoch": 0.7663921081511318, "grad_norm": 2.489518880844116, "learning_rate": 0.00029157726610472073, "loss": 5.2265, "step": 127100 }, { "epoch": 0.7666935999324659, "grad_norm": 1.6111459732055664, "learning_rate": 0.0002915705419361584, "loss": 5.3042, "step": 127150 }, { "epoch": 0.7669950917137999, "grad_norm": 1.199538230895996, "learning_rate": 0.00029156381516219225, "loss": 5.3519, "step": 127200 }, { "epoch": 0.767296583495134, "grad_norm": 0.791989803314209, "learning_rate": 0.00029155708578294614, "loss": 5.2962, "step": 127250 }, { "epoch": 0.767598075276468, "grad_norm": 1.134133219718933, "learning_rate": 0.00029155035379854386, "loss": 5.7006, "step": 127300 }, { "epoch": 0.767899567057802, "grad_norm": 0.9864721298217773, "learning_rate": 0.0002915436192091093, "loss": 5.0769, "step": 127350 }, { "epoch": 0.768201058839136, "grad_norm": 0.8204807043075562, "learning_rate": 0.00029153688201476646, "loss": 5.1544, "step": 127400 }, { "epoch": 0.7685025506204701, "grad_norm": 0.7032056450843811, "learning_rate": 0.0002915301422156393, "loss": 4.7647, "step": 127450 }, { "epoch": 0.7688040424018041, "grad_norm": 1.0878989696502686, "learning_rate": 0.00029152339981185183, "loss": 5.3766, "step": 127500 }, { "epoch": 0.7691055341831382, "grad_norm": 1.7651170492172241, "learning_rate": 0.0002915166548035281, "loss": 5.2157, "step": 127550 }, { "epoch": 0.7694070259644722, "grad_norm": 1.1877442598342896, "learning_rate": 0.00029150990719079234, "loss": 5.0233, "step": 127600 }, { "epoch": 0.7697085177458063, "grad_norm": 1.0799394845962524, "learning_rate": 0.0002915031569737687, "loss": 5.1534, "step": 127650 }, { "epoch": 0.7700100095271403, "grad_norm": 1.0322502851486206, "learning_rate": 0.00029149640415258133, "loss": 5.4336, "step": 127700 }, { "epoch": 0.7703115013084744, "grad_norm": 1.2913941144943237, "learning_rate": 0.0002914896487273547, "loss": 5.5685, "step": 127750 }, { "epoch": 0.7706129930898084, "grad_norm": 0.9388161301612854, "learning_rate": 0.00029148289069821286, "loss": 5.0909, "step": 127800 }, { "epoch": 0.7709144848711424, "grad_norm": 1.4356236457824707, "learning_rate": 0.00029147613006528036, "loss": 5.0673, "step": 127850 }, { "epoch": 0.7712159766524764, "grad_norm": 1.354750156402588, "learning_rate": 0.0002914693668286816, "loss": 5.3286, "step": 127900 }, { "epoch": 0.7715174684338105, "grad_norm": 1.6089489459991455, "learning_rate": 0.0002914626009885411, "loss": 5.6909, "step": 127950 }, { "epoch": 0.7718189602151445, "grad_norm": 1.08358895778656, "learning_rate": 0.0002914558325449832, "loss": 5.1119, "step": 128000 }, { "epoch": 0.7721204519964786, "grad_norm": 0.6483991742134094, "learning_rate": 0.0002914490614981326, "loss": 4.9396, "step": 128050 }, { "epoch": 0.7724219437778126, "grad_norm": 1.0997370481491089, "learning_rate": 0.0002914422878481139, "loss": 5.1302, "step": 128100 }, { "epoch": 0.7727234355591467, "grad_norm": 1.0813915729522705, "learning_rate": 0.0002914355115950517, "loss": 5.5819, "step": 128150 }, { "epoch": 0.7730249273404807, "grad_norm": 0.4717705249786377, "learning_rate": 0.0002914287327390708, "loss": 5.2839, "step": 128200 }, { "epoch": 0.7733264191218148, "grad_norm": 0.8676457405090332, "learning_rate": 0.00029142195128029584, "loss": 4.856, "step": 128250 }, { "epoch": 0.7736279109031488, "grad_norm": 1.327278733253479, "learning_rate": 0.0002914151672188518, "loss": 5.5552, "step": 128300 }, { "epoch": 0.7739294026844828, "grad_norm": 2.3012444972991943, "learning_rate": 0.00029140838055486334, "loss": 4.8988, "step": 128350 }, { "epoch": 0.7742308944658168, "grad_norm": 1.6060945987701416, "learning_rate": 0.00029140159128845537, "loss": 5.1431, "step": 128400 }, { "epoch": 0.7745323862471509, "grad_norm": 2.5434019565582275, "learning_rate": 0.000291394799419753, "loss": 5.4016, "step": 128450 }, { "epoch": 0.7748338780284849, "grad_norm": 1.4859672784805298, "learning_rate": 0.0002913880049488811, "loss": 5.3116, "step": 128500 }, { "epoch": 0.775135369809819, "grad_norm": 1.3162564039230347, "learning_rate": 0.00029138120787596477, "loss": 5.6796, "step": 128550 }, { "epoch": 0.775436861591153, "grad_norm": 1.1491605043411255, "learning_rate": 0.00029137440820112905, "loss": 5.2505, "step": 128600 }, { "epoch": 0.7757383533724871, "grad_norm": 1.652840256690979, "learning_rate": 0.00029136760592449906, "loss": 5.0269, "step": 128650 }, { "epoch": 0.7760398451538211, "grad_norm": 0.467025488615036, "learning_rate": 0.0002913608010462001, "loss": 4.9883, "step": 128700 }, { "epoch": 0.7763413369351552, "grad_norm": 1.0938507318496704, "learning_rate": 0.0002913539935663573, "loss": 5.13, "step": 128750 }, { "epoch": 0.7766428287164892, "grad_norm": 1.0418701171875, "learning_rate": 0.00029134718348509595, "loss": 5.4046, "step": 128800 }, { "epoch": 0.7769443204978232, "grad_norm": 1.2349730730056763, "learning_rate": 0.0002913403708025414, "loss": 5.2674, "step": 128850 }, { "epoch": 0.7772458122791572, "grad_norm": 1.2522149085998535, "learning_rate": 0.00029133355551881905, "loss": 5.346, "step": 128900 }, { "epoch": 0.7775473040604913, "grad_norm": 1.3279451131820679, "learning_rate": 0.00029132673763405427, "loss": 4.7155, "step": 128950 }, { "epoch": 0.7778487958418253, "grad_norm": 1.0035914182662964, "learning_rate": 0.0002913199171483726, "loss": 5.3299, "step": 129000 }, { "epoch": 0.7778487958418253, "eval_loss": 5.72955846786499, "eval_runtime": 38.913, "eval_samples_per_second": 13.158, "eval_steps_per_second": 6.579, "eval_tts_loss": 7.191275175602289, "step": 129000 }, { "epoch": 0.7781502876231594, "grad_norm": 1.26189386844635, "learning_rate": 0.0002913130940618995, "loss": 5.1615, "step": 129050 }, { "epoch": 0.7784517794044934, "grad_norm": 1.653571367263794, "learning_rate": 0.00029130626837476064, "loss": 5.1091, "step": 129100 }, { "epoch": 0.7787532711858275, "grad_norm": 1.5955101251602173, "learning_rate": 0.0002912994400870815, "loss": 5.2944, "step": 129150 }, { "epoch": 0.7790547629671615, "grad_norm": 1.237913966178894, "learning_rate": 0.0002912926091989878, "loss": 5.3179, "step": 129200 }, { "epoch": 0.7793562547484956, "grad_norm": 1.2084652185440063, "learning_rate": 0.0002912857757106053, "loss": 5.2347, "step": 129250 }, { "epoch": 0.7796577465298296, "grad_norm": 0.7200645208358765, "learning_rate": 0.00029127893962205975, "loss": 5.2878, "step": 129300 }, { "epoch": 0.7799592383111636, "grad_norm": 0.549976110458374, "learning_rate": 0.0002912721009334769, "loss": 5.4506, "step": 129350 }, { "epoch": 0.7802607300924976, "grad_norm": 1.06051504611969, "learning_rate": 0.0002912652596449826, "loss": 5.0471, "step": 129400 }, { "epoch": 0.7805622218738317, "grad_norm": 1.6894397735595703, "learning_rate": 0.00029125841575670285, "loss": 5.0032, "step": 129450 }, { "epoch": 0.7808637136551657, "grad_norm": 1.2265739440917969, "learning_rate": 0.0002912515692687635, "loss": 4.8853, "step": 129500 }, { "epoch": 0.7811652054364998, "grad_norm": 1.4489303827285767, "learning_rate": 0.00029124472018129065, "loss": 5.4616, "step": 129550 }, { "epoch": 0.7814666972178338, "grad_norm": 1.2831724882125854, "learning_rate": 0.00029123786849441025, "loss": 5.0738, "step": 129600 }, { "epoch": 0.7817681889991679, "grad_norm": 1.0674221515655518, "learning_rate": 0.00029123101420824843, "loss": 5.1506, "step": 129650 }, { "epoch": 0.7820696807805019, "grad_norm": 1.1283519268035889, "learning_rate": 0.00029122415732293135, "loss": 5.8391, "step": 129700 }, { "epoch": 0.782371172561836, "grad_norm": 1.4831706285476685, "learning_rate": 0.0002912172978385852, "loss": 5.3661, "step": 129750 }, { "epoch": 0.78267266434317, "grad_norm": 1.2333991527557373, "learning_rate": 0.00029121043575533625, "loss": 5.5395, "step": 129800 }, { "epoch": 0.782974156124504, "grad_norm": 0.7553555369377136, "learning_rate": 0.0002912035710733107, "loss": 4.9995, "step": 129850 }, { "epoch": 0.783275647905838, "grad_norm": 0.6679754853248596, "learning_rate": 0.00029119670379263487, "loss": 4.9285, "step": 129900 }, { "epoch": 0.7835771396871721, "grad_norm": 1.484755516052246, "learning_rate": 0.0002911898339134353, "loss": 5.0903, "step": 129950 }, { "epoch": 0.7838786314685061, "grad_norm": 1.6234513521194458, "learning_rate": 0.0002911829614358383, "loss": 5.4174, "step": 130000 }, { "epoch": 0.7841801232498402, "grad_norm": 1.0271718502044678, "learning_rate": 0.00029117608635997037, "loss": 5.0735, "step": 130050 }, { "epoch": 0.7844816150311742, "grad_norm": 2.1793086528778076, "learning_rate": 0.00029116920868595803, "loss": 5.2472, "step": 130100 }, { "epoch": 0.7847831068125083, "grad_norm": 1.14033842086792, "learning_rate": 0.0002911623284139279, "loss": 4.7287, "step": 130150 }, { "epoch": 0.7850845985938424, "grad_norm": 0.9402980804443359, "learning_rate": 0.00029115544554400656, "loss": 4.866, "step": 130200 }, { "epoch": 0.7853860903751764, "grad_norm": 1.6027016639709473, "learning_rate": 0.00029114856007632063, "loss": 5.1294, "step": 130250 }, { "epoch": 0.7856875821565105, "grad_norm": 1.6558880805969238, "learning_rate": 0.00029114167201099695, "loss": 4.9992, "step": 130300 }, { "epoch": 0.7859890739378445, "grad_norm": 1.1551703214645386, "learning_rate": 0.0002911347813481622, "loss": 5.1842, "step": 130350 }, { "epoch": 0.7862905657191785, "grad_norm": 1.0531498193740845, "learning_rate": 0.00029112788808794316, "loss": 5.4695, "step": 130400 }, { "epoch": 0.7865920575005125, "grad_norm": 1.5055956840515137, "learning_rate": 0.0002911209922304668, "loss": 5.4729, "step": 130450 }, { "epoch": 0.7868935492818466, "grad_norm": 2.272383451461792, "learning_rate": 0.00029111409377585994, "loss": 5.6828, "step": 130500 }, { "epoch": 0.7871950410631806, "grad_norm": 1.1125491857528687, "learning_rate": 0.00029110719272424957, "loss": 5.0704, "step": 130550 }, { "epoch": 0.7874965328445147, "grad_norm": 1.0772309303283691, "learning_rate": 0.0002911002890757627, "loss": 5.0604, "step": 130600 }, { "epoch": 0.7877980246258487, "grad_norm": 0.47305476665496826, "learning_rate": 0.0002910933828305263, "loss": 4.8532, "step": 130650 }, { "epoch": 0.7880995164071828, "grad_norm": 1.5033586025238037, "learning_rate": 0.0002910864739886676, "loss": 5.4404, "step": 130700 }, { "epoch": 0.7884010081885168, "grad_norm": 1.0064964294433594, "learning_rate": 0.0002910795625503137, "loss": 5.1318, "step": 130750 }, { "epoch": 0.7887024999698509, "grad_norm": 1.088860034942627, "learning_rate": 0.0002910726485155918, "loss": 5.3599, "step": 130800 }, { "epoch": 0.7890039917511849, "grad_norm": 1.0234642028808594, "learning_rate": 0.0002910657318846291, "loss": 5.6683, "step": 130850 }, { "epoch": 0.789305483532519, "grad_norm": 1.7070342302322388, "learning_rate": 0.0002910588126575529, "loss": 5.5912, "step": 130900 }, { "epoch": 0.7896069753138529, "grad_norm": 1.4788031578063965, "learning_rate": 0.00029105189083449055, "loss": 5.451, "step": 130950 }, { "epoch": 0.789908467095187, "grad_norm": 1.113247036933899, "learning_rate": 0.00029104496641556944, "loss": 4.9678, "step": 131000 }, { "epoch": 0.790209958876521, "grad_norm": 1.716799259185791, "learning_rate": 0.000291038039400917, "loss": 5.5417, "step": 131050 }, { "epoch": 0.7905114506578551, "grad_norm": 1.166069507598877, "learning_rate": 0.0002910311097906608, "loss": 5.5586, "step": 131100 }, { "epoch": 0.7908129424391891, "grad_norm": 1.3591129779815674, "learning_rate": 0.00029102417758492823, "loss": 5.7086, "step": 131150 }, { "epoch": 0.7911144342205232, "grad_norm": 1.3314286470413208, "learning_rate": 0.0002910172427838469, "loss": 5.4554, "step": 131200 }, { "epoch": 0.7914159260018572, "grad_norm": 1.4486899375915527, "learning_rate": 0.0002910103053875445, "loss": 5.1596, "step": 131250 }, { "epoch": 0.7917174177831913, "grad_norm": 0.3195931017398834, "learning_rate": 0.0002910033653961486, "loss": 5.1114, "step": 131300 }, { "epoch": 0.7920189095645253, "grad_norm": 1.8505327701568604, "learning_rate": 0.00029099642280978704, "loss": 5.2594, "step": 131350 }, { "epoch": 0.7923204013458593, "grad_norm": 1.0982904434204102, "learning_rate": 0.0002909894776285875, "loss": 5.0452, "step": 131400 }, { "epoch": 0.7926218931271933, "grad_norm": 1.3152421712875366, "learning_rate": 0.00029098252985267786, "loss": 5.24, "step": 131450 }, { "epoch": 0.7929233849085274, "grad_norm": 1.4293696880340576, "learning_rate": 0.0002909755794821859, "loss": 5.1326, "step": 131500 }, { "epoch": 0.7932248766898614, "grad_norm": 1.4812448024749756, "learning_rate": 0.00029096862651723963, "loss": 5.141, "step": 131550 }, { "epoch": 0.7935263684711955, "grad_norm": 2.1798994541168213, "learning_rate": 0.00029096167095796693, "loss": 5.5675, "step": 131600 }, { "epoch": 0.7938278602525295, "grad_norm": 1.8055225610733032, "learning_rate": 0.00029095471280449583, "loss": 5.1372, "step": 131650 }, { "epoch": 0.7941293520338636, "grad_norm": 1.4698280096054077, "learning_rate": 0.0002909477520569545, "loss": 5.1335, "step": 131700 }, { "epoch": 0.7944308438151976, "grad_norm": 1.4793106317520142, "learning_rate": 0.00029094078871547076, "loss": 5.2619, "step": 131750 }, { "epoch": 0.7947323355965317, "grad_norm": 0.9476176500320435, "learning_rate": 0.00029093382278017304, "loss": 5.533, "step": 131800 }, { "epoch": 0.7950338273778657, "grad_norm": 1.4069280624389648, "learning_rate": 0.0002909268542511895, "loss": 5.0333, "step": 131850 }, { "epoch": 0.7953353191591997, "grad_norm": 1.4570660591125488, "learning_rate": 0.00029091988312864817, "loss": 5.2432, "step": 131900 }, { "epoch": 0.7956368109405337, "grad_norm": 1.496341347694397, "learning_rate": 0.00029091290941267765, "loss": 5.6826, "step": 131950 }, { "epoch": 0.7959383027218678, "grad_norm": 1.1498674154281616, "learning_rate": 0.000290905933103406, "loss": 5.331, "step": 132000 }, { "epoch": 0.7959383027218678, "eval_loss": 5.731607437133789, "eval_runtime": 39.1678, "eval_samples_per_second": 13.072, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.085883951918039, "step": 132000 }, { "epoch": 0.7962397945032018, "grad_norm": 1.0113195180892944, "learning_rate": 0.0002908989542009618, "loss": 5.1622, "step": 132050 }, { "epoch": 0.7965412862845359, "grad_norm": 0.9689655900001526, "learning_rate": 0.0002908919727054734, "loss": 5.2534, "step": 132100 }, { "epoch": 0.7968427780658699, "grad_norm": 1.4978599548339844, "learning_rate": 0.0002908849886170693, "loss": 5.5787, "step": 132150 }, { "epoch": 0.797144269847204, "grad_norm": 1.17929208278656, "learning_rate": 0.000290878001935878, "loss": 5.3787, "step": 132200 }, { "epoch": 0.797445761628538, "grad_norm": 1.8078290224075317, "learning_rate": 0.0002908710126620282, "loss": 5.1117, "step": 132250 }, { "epoch": 0.7977472534098721, "grad_norm": 0.48583680391311646, "learning_rate": 0.0002908640207956484, "loss": 5.0094, "step": 132300 }, { "epoch": 0.798048745191206, "grad_norm": 1.3890587091445923, "learning_rate": 0.00029085702633686736, "loss": 5.2921, "step": 132350 }, { "epoch": 0.7983502369725402, "grad_norm": 1.3957005739212036, "learning_rate": 0.00029085002928581373, "loss": 5.0687, "step": 132400 }, { "epoch": 0.7986517287538741, "grad_norm": 1.0296698808670044, "learning_rate": 0.0002908430296426163, "loss": 5.791, "step": 132450 }, { "epoch": 0.7989532205352082, "grad_norm": 0.4877823293209076, "learning_rate": 0.0002908360274074039, "loss": 5.0943, "step": 132500 }, { "epoch": 0.7992547123165422, "grad_norm": 1.1243330240249634, "learning_rate": 0.00029082902258030545, "loss": 5.2508, "step": 132550 }, { "epoch": 0.7995562040978763, "grad_norm": 0.727188766002655, "learning_rate": 0.0002908220151614498, "loss": 5.2494, "step": 132600 }, { "epoch": 0.7998576958792103, "grad_norm": 0.5567173957824707, "learning_rate": 0.0002908150051509659, "loss": 5.1313, "step": 132650 }, { "epoch": 0.8001591876605444, "grad_norm": 1.1945198774337769, "learning_rate": 0.0002908079925489828, "loss": 5.473, "step": 132700 }, { "epoch": 0.8004606794418784, "grad_norm": 1.5984001159667969, "learning_rate": 0.00029080097735562955, "loss": 5.5238, "step": 132750 }, { "epoch": 0.8007621712232125, "grad_norm": 0.9499383568763733, "learning_rate": 0.0002907939595710352, "loss": 5.2525, "step": 132800 }, { "epoch": 0.8010636630045465, "grad_norm": 1.430915117263794, "learning_rate": 0.000290786939195329, "loss": 5.0468, "step": 132850 }, { "epoch": 0.8013651547858806, "grad_norm": 1.0832710266113281, "learning_rate": 0.00029077991622864007, "loss": 5.2031, "step": 132900 }, { "epoch": 0.8016666465672145, "grad_norm": 0.8721439242362976, "learning_rate": 0.0002907728906710977, "loss": 5.4549, "step": 132950 }, { "epoch": 0.8019681383485486, "grad_norm": 1.0068916082382202, "learning_rate": 0.00029076586252283116, "loss": 5.0993, "step": 133000 }, { "epoch": 0.8022696301298826, "grad_norm": 1.2152971029281616, "learning_rate": 0.00029075883178396984, "loss": 4.9362, "step": 133050 }, { "epoch": 0.8025711219112167, "grad_norm": 1.1411411762237549, "learning_rate": 0.000290751798454643, "loss": 5.2864, "step": 133100 }, { "epoch": 0.8028726136925507, "grad_norm": 2.0171194076538086, "learning_rate": 0.00029074476253498026, "loss": 4.884, "step": 133150 }, { "epoch": 0.8031741054738848, "grad_norm": 1.4290567636489868, "learning_rate": 0.00029073772402511103, "loss": 5.4507, "step": 133200 }, { "epoch": 0.8034755972552188, "grad_norm": 2.5161948204040527, "learning_rate": 0.00029073068292516477, "loss": 5.2834, "step": 133250 }, { "epoch": 0.8037770890365529, "grad_norm": 1.2897961139678955, "learning_rate": 0.00029072363923527116, "loss": 5.1984, "step": 133300 }, { "epoch": 0.8040785808178869, "grad_norm": 1.430635929107666, "learning_rate": 0.0002907165929555598, "loss": 5.4525, "step": 133350 }, { "epoch": 0.804380072599221, "grad_norm": 1.004004955291748, "learning_rate": 0.0002907095440861603, "loss": 5.2849, "step": 133400 }, { "epoch": 0.8046815643805549, "grad_norm": 1.0255727767944336, "learning_rate": 0.00029070249262720254, "loss": 5.2947, "step": 133450 }, { "epoch": 0.804983056161889, "grad_norm": 1.0059518814086914, "learning_rate": 0.0002906954385788161, "loss": 5.5342, "step": 133500 }, { "epoch": 0.805284547943223, "grad_norm": 0.806026816368103, "learning_rate": 0.00029068838194113095, "loss": 5.1692, "step": 133550 }, { "epoch": 0.8055860397245571, "grad_norm": 1.3268852233886719, "learning_rate": 0.0002906813227142769, "loss": 5.2304, "step": 133600 }, { "epoch": 0.8058875315058911, "grad_norm": 0.3662572205066681, "learning_rate": 0.0002906742608983839, "loss": 5.1418, "step": 133650 }, { "epoch": 0.8061890232872252, "grad_norm": 0.6420709490776062, "learning_rate": 0.0002906671964935818, "loss": 5.3279, "step": 133700 }, { "epoch": 0.8064905150685592, "grad_norm": 1.0476487874984741, "learning_rate": 0.00029066012950000077, "loss": 5.3742, "step": 133750 }, { "epoch": 0.8067920068498933, "grad_norm": 1.1019951105117798, "learning_rate": 0.00029065305991777076, "loss": 5.2101, "step": 133800 }, { "epoch": 0.8070934986312273, "grad_norm": 1.1313132047653198, "learning_rate": 0.0002906459877470219, "loss": 5.3877, "step": 133850 }, { "epoch": 0.8073949904125614, "grad_norm": 1.4470409154891968, "learning_rate": 0.00029063891298788437, "loss": 5.6247, "step": 133900 }, { "epoch": 0.8076964821938954, "grad_norm": 0.8944329023361206, "learning_rate": 0.0002906318356404883, "loss": 4.91, "step": 133950 }, { "epoch": 0.8079979739752294, "grad_norm": 0.7046508193016052, "learning_rate": 0.000290624755704964, "loss": 5.2891, "step": 134000 }, { "epoch": 0.8082994657565635, "grad_norm": 1.273919939994812, "learning_rate": 0.00029061767318144176, "loss": 4.8763, "step": 134050 }, { "epoch": 0.8086009575378975, "grad_norm": 0.9677875638008118, "learning_rate": 0.0002906105880700519, "loss": 4.9179, "step": 134100 }, { "epoch": 0.8089024493192316, "grad_norm": 1.888319492340088, "learning_rate": 0.00029060350037092493, "loss": 5.1481, "step": 134150 }, { "epoch": 0.8092039411005656, "grad_norm": 1.5735918283462524, "learning_rate": 0.0002905964100841911, "loss": 4.6803, "step": 134200 }, { "epoch": 0.8095054328818997, "grad_norm": 1.546261191368103, "learning_rate": 0.000290589317209981, "loss": 5.4477, "step": 134250 }, { "epoch": 0.8098069246632337, "grad_norm": 1.2878342866897583, "learning_rate": 0.00029058222174842513, "loss": 4.9911, "step": 134300 }, { "epoch": 0.8101084164445678, "grad_norm": 1.3530205488204956, "learning_rate": 0.0002905751236996541, "loss": 4.9283, "step": 134350 }, { "epoch": 0.8104099082259018, "grad_norm": 1.233054757118225, "learning_rate": 0.00029056802306379854, "loss": 5.2291, "step": 134400 }, { "epoch": 0.8107114000072358, "grad_norm": 1.013548493385315, "learning_rate": 0.0002905609198409891, "loss": 5.2447, "step": 134450 }, { "epoch": 0.8110128917885698, "grad_norm": 0.5203688144683838, "learning_rate": 0.00029055381403135657, "loss": 5.0001, "step": 134500 }, { "epoch": 0.8113143835699039, "grad_norm": 1.1649212837219238, "learning_rate": 0.0002905467056350316, "loss": 5.3562, "step": 134550 }, { "epoch": 0.8116158753512379, "grad_norm": 0.8767015337944031, "learning_rate": 0.00029053959465214516, "loss": 5.4628, "step": 134600 }, { "epoch": 0.811917367132572, "grad_norm": 0.919330358505249, "learning_rate": 0.000290532481082828, "loss": 5.4758, "step": 134650 }, { "epoch": 0.812218858913906, "grad_norm": 2.2194764614105225, "learning_rate": 0.0002905253649272111, "loss": 5.2241, "step": 134700 }, { "epoch": 0.8125203506952401, "grad_norm": 0.5944370627403259, "learning_rate": 0.0002905182461854253, "loss": 5.2093, "step": 134750 }, { "epoch": 0.8128218424765741, "grad_norm": 1.203338861465454, "learning_rate": 0.0002905111248576018, "loss": 5.3515, "step": 134800 }, { "epoch": 0.8131233342579082, "grad_norm": 1.078797459602356, "learning_rate": 0.0002905040009438716, "loss": 5.1608, "step": 134850 }, { "epoch": 0.8134248260392422, "grad_norm": 1.010256290435791, "learning_rate": 0.0002904968744443657, "loss": 5.0244, "step": 134900 }, { "epoch": 0.8137263178205763, "grad_norm": 1.4708383083343506, "learning_rate": 0.0002904897453592154, "loss": 5.1812, "step": 134950 }, { "epoch": 0.8140278096019102, "grad_norm": 1.4647935628890991, "learning_rate": 0.00029048261368855175, "loss": 5.1744, "step": 135000 }, { "epoch": 0.8140278096019102, "eval_loss": 5.734951019287109, "eval_runtime": 39.1246, "eval_samples_per_second": 13.086, "eval_steps_per_second": 6.543, "eval_tts_loss": 7.207141410498426, "step": 135000 }, { "epoch": 0.8143293013832443, "grad_norm": 1.298025369644165, "learning_rate": 0.00029047547943250606, "loss": 5.5353, "step": 135050 }, { "epoch": 0.8146307931645783, "grad_norm": 2.1945571899414062, "learning_rate": 0.00029046834259120974, "loss": 5.5714, "step": 135100 }, { "epoch": 0.8149322849459124, "grad_norm": 1.1502937078475952, "learning_rate": 0.00029046120316479396, "loss": 5.3362, "step": 135150 }, { "epoch": 0.8152337767272464, "grad_norm": 2.2068865299224854, "learning_rate": 0.0002904540611533902, "loss": 5.6077, "step": 135200 }, { "epoch": 0.8155352685085805, "grad_norm": 1.3854870796203613, "learning_rate": 0.00029044691655712987, "loss": 5.5115, "step": 135250 }, { "epoch": 0.8158367602899145, "grad_norm": 1.322706699371338, "learning_rate": 0.0002904397693761445, "loss": 4.7865, "step": 135300 }, { "epoch": 0.8161382520712486, "grad_norm": 1.6164238452911377, "learning_rate": 0.0002904326196105656, "loss": 4.7772, "step": 135350 }, { "epoch": 0.8164397438525826, "grad_norm": 1.725279688835144, "learning_rate": 0.0002904254672605247, "loss": 5.5071, "step": 135400 }, { "epoch": 0.8167412356339167, "grad_norm": 1.431560754776001, "learning_rate": 0.0002904183123261535, "loss": 5.0042, "step": 135450 }, { "epoch": 0.8170427274152506, "grad_norm": 1.7441141605377197, "learning_rate": 0.00029041115480758366, "loss": 5.1221, "step": 135500 }, { "epoch": 0.8173442191965847, "grad_norm": 1.6067993640899658, "learning_rate": 0.0002904039947049469, "loss": 5.574, "step": 135550 }, { "epoch": 0.8176457109779187, "grad_norm": 1.756365418434143, "learning_rate": 0.00029039683201837493, "loss": 5.3249, "step": 135600 }, { "epoch": 0.8179472027592528, "grad_norm": 1.6851232051849365, "learning_rate": 0.00029038966674799964, "loss": 5.3797, "step": 135650 }, { "epoch": 0.8182486945405868, "grad_norm": 1.0889946222305298, "learning_rate": 0.0002903824988939529, "loss": 5.1505, "step": 135700 }, { "epoch": 0.8185501863219209, "grad_norm": 1.2214865684509277, "learning_rate": 0.0002903753284563666, "loss": 5.2356, "step": 135750 }, { "epoch": 0.8188516781032549, "grad_norm": 0.7702077031135559, "learning_rate": 0.0002903681554353727, "loss": 4.5853, "step": 135800 }, { "epoch": 0.819153169884589, "grad_norm": 0.9945457577705383, "learning_rate": 0.0002903609798311032, "loss": 5.4454, "step": 135850 }, { "epoch": 0.819454661665923, "grad_norm": 1.3169077634811401, "learning_rate": 0.0002903538016436902, "loss": 4.9104, "step": 135900 }, { "epoch": 0.819756153447257, "grad_norm": 1.563934087753296, "learning_rate": 0.0002903466208732658, "loss": 5.0575, "step": 135950 }, { "epoch": 0.820057645228591, "grad_norm": 1.3695234060287476, "learning_rate": 0.00029033943751996207, "loss": 5.252, "step": 136000 }, { "epoch": 0.8203591370099251, "grad_norm": 1.4205106496810913, "learning_rate": 0.0002903322515839113, "loss": 5.173, "step": 136050 }, { "epoch": 0.8206606287912591, "grad_norm": 1.237276554107666, "learning_rate": 0.0002903250630652457, "loss": 4.7158, "step": 136100 }, { "epoch": 0.8209621205725932, "grad_norm": 1.6901017427444458, "learning_rate": 0.00029031787196409756, "loss": 4.9621, "step": 136150 }, { "epoch": 0.8212636123539272, "grad_norm": 1.370245099067688, "learning_rate": 0.0002903106782805993, "loss": 5.5585, "step": 136200 }, { "epoch": 0.8215651041352613, "grad_norm": 1.2136391401290894, "learning_rate": 0.00029030348201488314, "loss": 5.0139, "step": 136250 }, { "epoch": 0.8218665959165953, "grad_norm": 1.6570115089416504, "learning_rate": 0.0002902962831670817, "loss": 5.3841, "step": 136300 }, { "epoch": 0.8221680876979294, "grad_norm": 0.9758792519569397, "learning_rate": 0.0002902890817373273, "loss": 5.7096, "step": 136350 }, { "epoch": 0.8224695794792634, "grad_norm": 1.3399864435195923, "learning_rate": 0.0002902818777257526, "loss": 5.221, "step": 136400 }, { "epoch": 0.8227710712605975, "grad_norm": 0.9215564131736755, "learning_rate": 0.0002902746711324901, "loss": 4.8022, "step": 136450 }, { "epoch": 0.8230725630419314, "grad_norm": 0.9870257377624512, "learning_rate": 0.0002902674619576725, "loss": 5.2323, "step": 136500 }, { "epoch": 0.8233740548232655, "grad_norm": 0.9883350729942322, "learning_rate": 0.0002902602502014324, "loss": 4.5745, "step": 136550 }, { "epoch": 0.8236755466045995, "grad_norm": 1.1198054552078247, "learning_rate": 0.00029025303586390264, "loss": 5.1727, "step": 136600 }, { "epoch": 0.8239770383859336, "grad_norm": 1.1063677072525024, "learning_rate": 0.00029024581894521585, "loss": 5.7097, "step": 136650 }, { "epoch": 0.8242785301672676, "grad_norm": 1.546543002128601, "learning_rate": 0.00029023859944550485, "loss": 5.4237, "step": 136700 }, { "epoch": 0.8245800219486017, "grad_norm": 1.445135474205017, "learning_rate": 0.0002902313773649026, "loss": 4.9806, "step": 136750 }, { "epoch": 0.8248815137299357, "grad_norm": 1.0399121046066284, "learning_rate": 0.00029022415270354204, "loss": 5.0672, "step": 136800 }, { "epoch": 0.8251830055112698, "grad_norm": 0.7156456708908081, "learning_rate": 0.00029021692546155596, "loss": 5.2311, "step": 136850 }, { "epoch": 0.8254844972926038, "grad_norm": 1.310280680656433, "learning_rate": 0.0002902096956390775, "loss": 5.1707, "step": 136900 }, { "epoch": 0.8257859890739379, "grad_norm": 1.374313235282898, "learning_rate": 0.0002902024632362397, "loss": 5.3866, "step": 136950 }, { "epoch": 0.8260874808552718, "grad_norm": 1.4426546096801758, "learning_rate": 0.0002901952282531757, "loss": 5.5413, "step": 137000 }, { "epoch": 0.8263889726366059, "grad_norm": 1.7661269903182983, "learning_rate": 0.0002901879906900185, "loss": 4.851, "step": 137050 }, { "epoch": 0.8266904644179399, "grad_norm": 1.0356687307357788, "learning_rate": 0.00029018075054690134, "loss": 5.1058, "step": 137100 }, { "epoch": 0.826991956199274, "grad_norm": 1.2628909349441528, "learning_rate": 0.00029017350782395765, "loss": 5.4554, "step": 137150 }, { "epoch": 0.827293447980608, "grad_norm": 1.2172057628631592, "learning_rate": 0.0002901662625213205, "loss": 4.5391, "step": 137200 }, { "epoch": 0.8275949397619421, "grad_norm": 1.2102835178375244, "learning_rate": 0.00029015901463912337, "loss": 5.33, "step": 137250 }, { "epoch": 0.8278964315432761, "grad_norm": 1.5023066997528076, "learning_rate": 0.00029015176417749954, "loss": 5.9391, "step": 137300 }, { "epoch": 0.8281979233246102, "grad_norm": 1.4498045444488525, "learning_rate": 0.00029014451113658254, "loss": 5.2582, "step": 137350 }, { "epoch": 0.8284994151059442, "grad_norm": 1.0980250835418701, "learning_rate": 0.0002901372555165058, "loss": 5.2118, "step": 137400 }, { "epoch": 0.8288009068872783, "grad_norm": 1.1943756341934204, "learning_rate": 0.00029012999731740284, "loss": 5.6381, "step": 137450 }, { "epoch": 0.8291023986686122, "grad_norm": 0.9890260696411133, "learning_rate": 0.0002901227365394073, "loss": 5.3471, "step": 137500 }, { "epoch": 0.8294038904499463, "grad_norm": 1.535514235496521, "learning_rate": 0.0002901154731826527, "loss": 4.8665, "step": 137550 }, { "epoch": 0.8297053822312804, "grad_norm": 1.7396818399429321, "learning_rate": 0.0002901082072472728, "loss": 5.5806, "step": 137600 }, { "epoch": 0.8300068740126144, "grad_norm": 1.230008602142334, "learning_rate": 0.0002901009387334013, "loss": 4.9607, "step": 137650 }, { "epoch": 0.8303083657939485, "grad_norm": 1.1701226234436035, "learning_rate": 0.00029009366764117195, "loss": 5.1837, "step": 137700 }, { "epoch": 0.8306098575752825, "grad_norm": 1.526203989982605, "learning_rate": 0.00029008639397071857, "loss": 5.326, "step": 137750 }, { "epoch": 0.8309113493566166, "grad_norm": 1.1228090524673462, "learning_rate": 0.000290079117722175, "loss": 5.2447, "step": 137800 }, { "epoch": 0.8312128411379506, "grad_norm": 1.2271183729171753, "learning_rate": 0.0002900718388956752, "loss": 4.7942, "step": 137850 }, { "epoch": 0.8315143329192847, "grad_norm": 0.4828726351261139, "learning_rate": 0.00029006455749135314, "loss": 4.8728, "step": 137900 }, { "epoch": 0.8318158247006187, "grad_norm": 0.45825764536857605, "learning_rate": 0.0002900572735093427, "loss": 5.5739, "step": 137950 }, { "epoch": 0.8321173164819528, "grad_norm": 1.4814683198928833, "learning_rate": 0.00029004998694977805, "loss": 5.0511, "step": 138000 }, { "epoch": 0.8321173164819528, "eval_loss": 5.719982147216797, "eval_runtime": 38.9019, "eval_samples_per_second": 13.161, "eval_steps_per_second": 6.581, "eval_tts_loss": 7.142721306703367, "step": 138000 }, { "epoch": 0.8324188082632867, "grad_norm": 0.9169418811798096, "learning_rate": 0.00029004269781279324, "loss": 5.2821, "step": 138050 }, { "epoch": 0.8327203000446208, "grad_norm": 1.0362788438796997, "learning_rate": 0.00029003540609852244, "loss": 5.3366, "step": 138100 }, { "epoch": 0.8330217918259548, "grad_norm": 1.1905378103256226, "learning_rate": 0.00029002811180709983, "loss": 4.9973, "step": 138150 }, { "epoch": 0.8333232836072889, "grad_norm": 1.0489524602890015, "learning_rate": 0.0002900208149386597, "loss": 4.9442, "step": 138200 }, { "epoch": 0.8336247753886229, "grad_norm": 0.9766947627067566, "learning_rate": 0.00029001351549333616, "loss": 5.19, "step": 138250 }, { "epoch": 0.833926267169957, "grad_norm": 2.118494749069214, "learning_rate": 0.00029000621347126377, "loss": 5.5512, "step": 138300 }, { "epoch": 0.834227758951291, "grad_norm": 1.1266542673110962, "learning_rate": 0.0002899989088725769, "loss": 5.2928, "step": 138350 }, { "epoch": 0.8345292507326251, "grad_norm": 1.2166049480438232, "learning_rate": 0.0002899916016974098, "loss": 5.0954, "step": 138400 }, { "epoch": 0.8348307425139591, "grad_norm": 1.292631983757019, "learning_rate": 0.000289984291945897, "loss": 5.2401, "step": 138450 }, { "epoch": 0.8351322342952932, "grad_norm": 1.0331097841262817, "learning_rate": 0.00028997697961817315, "loss": 4.724, "step": 138500 }, { "epoch": 0.8354337260766271, "grad_norm": 0.8226186037063599, "learning_rate": 0.00028996966471437274, "loss": 4.8899, "step": 138550 }, { "epoch": 0.8357352178579612, "grad_norm": 1.9371695518493652, "learning_rate": 0.0002899623472346304, "loss": 5.2066, "step": 138600 }, { "epoch": 0.8360367096392952, "grad_norm": 1.604947566986084, "learning_rate": 0.0002899550271790808, "loss": 5.3121, "step": 138650 }, { "epoch": 0.8363382014206293, "grad_norm": 2.1797285079956055, "learning_rate": 0.0002899477045478587, "loss": 4.9232, "step": 138700 }, { "epoch": 0.8366396932019633, "grad_norm": 0.9705803394317627, "learning_rate": 0.0002899403793410987, "loss": 5.196, "step": 138750 }, { "epoch": 0.8369411849832974, "grad_norm": 1.3456295728683472, "learning_rate": 0.0002899330515589358, "loss": 5.0993, "step": 138800 }, { "epoch": 0.8372426767646314, "grad_norm": 1.3652032613754272, "learning_rate": 0.0002899257212015048, "loss": 5.4786, "step": 138850 }, { "epoch": 0.8375441685459655, "grad_norm": 1.1542319059371948, "learning_rate": 0.0002899183882689405, "loss": 5.4354, "step": 138900 }, { "epoch": 0.8378456603272995, "grad_norm": 1.0036309957504272, "learning_rate": 0.000289911052761378, "loss": 5.2018, "step": 138950 }, { "epoch": 0.8381471521086336, "grad_norm": 1.429843544960022, "learning_rate": 0.0002899037146789522, "loss": 5.2608, "step": 139000 }, { "epoch": 0.8384486438899675, "grad_norm": 1.537366271018982, "learning_rate": 0.0002898963740217982, "loss": 4.9725, "step": 139050 }, { "epoch": 0.8387501356713016, "grad_norm": 0.3024328649044037, "learning_rate": 0.0002898890307900511, "loss": 5.2008, "step": 139100 }, { "epoch": 0.8390516274526356, "grad_norm": 1.3279597759246826, "learning_rate": 0.000289881684983846, "loss": 5.0004, "step": 139150 }, { "epoch": 0.8393531192339697, "grad_norm": 1.2568728923797607, "learning_rate": 0.00028987433660331816, "loss": 5.3752, "step": 139200 }, { "epoch": 0.8396546110153037, "grad_norm": 1.0839368104934692, "learning_rate": 0.0002898669856486027, "loss": 5.3063, "step": 139250 }, { "epoch": 0.8399561027966378, "grad_norm": 1.0197902917861938, "learning_rate": 0.000289859632119835, "loss": 5.1843, "step": 139300 }, { "epoch": 0.8402575945779718, "grad_norm": 1.3094348907470703, "learning_rate": 0.0002898522760171504, "loss": 4.8718, "step": 139350 }, { "epoch": 0.8405590863593059, "grad_norm": 1.6153157949447632, "learning_rate": 0.00028984491734068414, "loss": 5.1926, "step": 139400 }, { "epoch": 0.8408605781406399, "grad_norm": 1.3705559968948364, "learning_rate": 0.0002898375560905718, "loss": 5.0178, "step": 139450 }, { "epoch": 0.841162069921974, "grad_norm": 1.0655553340911865, "learning_rate": 0.0002898301922669488, "loss": 5.1148, "step": 139500 }, { "epoch": 0.8414635617033079, "grad_norm": 0.8245575428009033, "learning_rate": 0.0002898228258699507, "loss": 5.2315, "step": 139550 }, { "epoch": 0.841765053484642, "grad_norm": 1.349256157875061, "learning_rate": 0.00028981545689971296, "loss": 5.325, "step": 139600 }, { "epoch": 0.842066545265976, "grad_norm": 0.5645418167114258, "learning_rate": 0.0002898080853563713, "loss": 5.4691, "step": 139650 }, { "epoch": 0.8423680370473101, "grad_norm": 1.070770263671875, "learning_rate": 0.00028980071124006136, "loss": 5.203, "step": 139700 }, { "epoch": 0.8426695288286441, "grad_norm": 1.1078985929489136, "learning_rate": 0.0002897933345509188, "loss": 5.1963, "step": 139750 }, { "epoch": 0.8429710206099782, "grad_norm": 1.23484206199646, "learning_rate": 0.0002897859552890794, "loss": 5.5551, "step": 139800 }, { "epoch": 0.8432725123913122, "grad_norm": 2.0887680053710938, "learning_rate": 0.000289778573454679, "loss": 5.0615, "step": 139850 }, { "epoch": 0.8435740041726463, "grad_norm": 1.7395751476287842, "learning_rate": 0.00028977118904785346, "loss": 5.5353, "step": 139900 }, { "epoch": 0.8438754959539803, "grad_norm": 1.269874930381775, "learning_rate": 0.00028976380206873864, "loss": 4.9776, "step": 139950 }, { "epoch": 0.8441769877353144, "grad_norm": 1.1180641651153564, "learning_rate": 0.00028975641251747044, "loss": 5.2481, "step": 140000 }, { "epoch": 0.8444784795166483, "grad_norm": 0.9120194315910339, "learning_rate": 0.000289749020394185, "loss": 5.0458, "step": 140050 }, { "epoch": 0.8447799712979824, "grad_norm": 0.3846522569656372, "learning_rate": 0.0002897416256990182, "loss": 5.1293, "step": 140100 }, { "epoch": 0.8450814630793164, "grad_norm": 1.4342693090438843, "learning_rate": 0.0002897342284321062, "loss": 5.9822, "step": 140150 }, { "epoch": 0.8453829548606505, "grad_norm": 1.4758929014205933, "learning_rate": 0.00028972682859358515, "loss": 4.5466, "step": 140200 }, { "epoch": 0.8456844466419845, "grad_norm": 0.9668847918510437, "learning_rate": 0.00028971942618359126, "loss": 5.621, "step": 140250 }, { "epoch": 0.8459859384233186, "grad_norm": 1.4154738187789917, "learning_rate": 0.0002897120212022607, "loss": 5.201, "step": 140300 }, { "epoch": 0.8462874302046526, "grad_norm": 0.5626274943351746, "learning_rate": 0.0002897046136497298, "loss": 5.5809, "step": 140350 }, { "epoch": 0.8465889219859867, "grad_norm": 1.2079267501831055, "learning_rate": 0.00028969720352613475, "loss": 4.9091, "step": 140400 }, { "epoch": 0.8468904137673207, "grad_norm": 1.069636344909668, "learning_rate": 0.00028968979083161215, "loss": 5.0017, "step": 140450 }, { "epoch": 0.8471919055486548, "grad_norm": 1.6458073854446411, "learning_rate": 0.0002896823755662982, "loss": 5.4289, "step": 140500 }, { "epoch": 0.8474933973299887, "grad_norm": 2.0044679641723633, "learning_rate": 0.0002896749577303295, "loss": 5.2769, "step": 140550 }, { "epoch": 0.8477948891113228, "grad_norm": 1.331874132156372, "learning_rate": 0.00028966753732384255, "loss": 5.4758, "step": 140600 }, { "epoch": 0.8480963808926568, "grad_norm": 1.1950173377990723, "learning_rate": 0.0002896601143469739, "loss": 4.8988, "step": 140650 }, { "epoch": 0.8483978726739909, "grad_norm": 0.7479628920555115, "learning_rate": 0.00028965268879986014, "loss": 5.1374, "step": 140700 }, { "epoch": 0.8486993644553249, "grad_norm": 1.1420049667358398, "learning_rate": 0.0002896452606826379, "loss": 5.1542, "step": 140750 }, { "epoch": 0.849000856236659, "grad_norm": 0.9401954412460327, "learning_rate": 0.000289637829995444, "loss": 5.0899, "step": 140800 }, { "epoch": 0.849302348017993, "grad_norm": 1.7032604217529297, "learning_rate": 0.00028963039673841505, "loss": 5.0185, "step": 140850 }, { "epoch": 0.8496038397993271, "grad_norm": 1.6602299213409424, "learning_rate": 0.0002896229609116879, "loss": 5.1008, "step": 140900 }, { "epoch": 0.8499053315806611, "grad_norm": 1.071987509727478, "learning_rate": 0.00028961552251539945, "loss": 5.305, "step": 140950 }, { "epoch": 0.8502068233619952, "grad_norm": 1.7509183883666992, "learning_rate": 0.0002896080815496865, "loss": 5.0555, "step": 141000 }, { "epoch": 0.8502068233619952, "eval_loss": 5.721113204956055, "eval_runtime": 39.0306, "eval_samples_per_second": 13.118, "eval_steps_per_second": 6.559, "eval_tts_loss": 7.172998202296283, "step": 141000 }, { "epoch": 0.8505083151433291, "grad_norm": 0.8998810052871704, "learning_rate": 0.0002896006380146861, "loss": 5.0322, "step": 141050 }, { "epoch": 0.8508098069246632, "grad_norm": 0.7591813802719116, "learning_rate": 0.00028959319191053514, "loss": 5.1708, "step": 141100 }, { "epoch": 0.8511112987059972, "grad_norm": 1.3330764770507812, "learning_rate": 0.00028958574323737076, "loss": 5.5178, "step": 141150 }, { "epoch": 0.8514127904873313, "grad_norm": 1.283432126045227, "learning_rate": 0.0002895782919953299, "loss": 5.3753, "step": 141200 }, { "epoch": 0.8517142822686653, "grad_norm": 1.060749888420105, "learning_rate": 0.00028957083818454977, "loss": 5.3678, "step": 141250 }, { "epoch": 0.8520157740499994, "grad_norm": 1.7078582048416138, "learning_rate": 0.00028956338180516757, "loss": 5.2598, "step": 141300 }, { "epoch": 0.8523172658313335, "grad_norm": 1.4696850776672363, "learning_rate": 0.00028955592285732044, "loss": 5.1048, "step": 141350 }, { "epoch": 0.8526187576126675, "grad_norm": 1.2513585090637207, "learning_rate": 0.0002895484613411458, "loss": 5.0103, "step": 141400 }, { "epoch": 0.8529202493940016, "grad_norm": 1.0984876155853271, "learning_rate": 0.0002895409972567808, "loss": 5.1968, "step": 141450 }, { "epoch": 0.8532217411753356, "grad_norm": 1.0055692195892334, "learning_rate": 0.00028953353060436293, "loss": 5.3455, "step": 141500 }, { "epoch": 0.8535232329566697, "grad_norm": 0.8197088837623596, "learning_rate": 0.0002895260613840295, "loss": 5.6551, "step": 141550 }, { "epoch": 0.8538247247380036, "grad_norm": 0.9842035174369812, "learning_rate": 0.0002895185895959181, "loss": 5.9079, "step": 141600 }, { "epoch": 0.8541262165193377, "grad_norm": 1.0068162679672241, "learning_rate": 0.0002895111152401661, "loss": 5.2717, "step": 141650 }, { "epoch": 0.8544277083006717, "grad_norm": 0.7818233370780945, "learning_rate": 0.00028950363831691114, "loss": 5.2633, "step": 141700 }, { "epoch": 0.8547292000820058, "grad_norm": 0.8947495222091675, "learning_rate": 0.00028949615882629075, "loss": 5.0456, "step": 141750 }, { "epoch": 0.8550306918633398, "grad_norm": 2.1282548904418945, "learning_rate": 0.0002894886767684427, "loss": 5.2774, "step": 141800 }, { "epoch": 0.8553321836446739, "grad_norm": 1.4944359064102173, "learning_rate": 0.0002894811921435046, "loss": 5.188, "step": 141850 }, { "epoch": 0.8556336754260079, "grad_norm": 1.688428521156311, "learning_rate": 0.00028947370495161417, "loss": 4.8903, "step": 141900 }, { "epoch": 0.855935167207342, "grad_norm": 1.891233205795288, "learning_rate": 0.0002894662151929093, "loss": 5.8669, "step": 141950 }, { "epoch": 0.856236658988676, "grad_norm": 1.3330365419387817, "learning_rate": 0.0002894587228675277, "loss": 5.4252, "step": 142000 }, { "epoch": 0.85653815077001, "grad_norm": 1.661759376525879, "learning_rate": 0.00028945122797560734, "loss": 5.3684, "step": 142050 }, { "epoch": 0.856839642551344, "grad_norm": 1.4836699962615967, "learning_rate": 0.00028944373051728616, "loss": 5.2115, "step": 142100 }, { "epoch": 0.8571411343326781, "grad_norm": 1.1644299030303955, "learning_rate": 0.0002894362304927021, "loss": 5.4161, "step": 142150 }, { "epoch": 0.8574426261140121, "grad_norm": 1.0555540323257446, "learning_rate": 0.00028942872790199323, "loss": 5.4441, "step": 142200 }, { "epoch": 0.8577441178953462, "grad_norm": 1.4620413780212402, "learning_rate": 0.00028942122274529757, "loss": 5.3689, "step": 142250 }, { "epoch": 0.8580456096766802, "grad_norm": 1.097014307975769, "learning_rate": 0.0002894137150227533, "loss": 5.5275, "step": 142300 }, { "epoch": 0.8583471014580143, "grad_norm": 1.9301772117614746, "learning_rate": 0.00028940620473449853, "loss": 5.357, "step": 142350 }, { "epoch": 0.8586485932393483, "grad_norm": 0.6344509124755859, "learning_rate": 0.0002893986918806715, "loss": 5.1223, "step": 142400 }, { "epoch": 0.8589500850206824, "grad_norm": 1.2937593460083008, "learning_rate": 0.0002893911764614105, "loss": 5.5271, "step": 142450 }, { "epoch": 0.8592515768020164, "grad_norm": 1.3789441585540771, "learning_rate": 0.0002893836584768538, "loss": 4.6476, "step": 142500 }, { "epoch": 0.8595530685833505, "grad_norm": 1.2555580139160156, "learning_rate": 0.0002893761379271398, "loss": 5.2554, "step": 142550 }, { "epoch": 0.8598545603646844, "grad_norm": 2.158155679702759, "learning_rate": 0.0002893686148124069, "loss": 5.0464, "step": 142600 }, { "epoch": 0.8601560521460185, "grad_norm": 1.3294527530670166, "learning_rate": 0.0002893610891327935, "loss": 5.4462, "step": 142650 }, { "epoch": 0.8604575439273525, "grad_norm": 1.1815835237503052, "learning_rate": 0.0002893535608884382, "loss": 5.2518, "step": 142700 }, { "epoch": 0.8607590357086866, "grad_norm": 1.6405402421951294, "learning_rate": 0.0002893460300794794, "loss": 5.0968, "step": 142750 }, { "epoch": 0.8610605274900206, "grad_norm": 1.3954404592514038, "learning_rate": 0.0002893384967060558, "loss": 4.9784, "step": 142800 }, { "epoch": 0.8613620192713547, "grad_norm": 1.0264015197753906, "learning_rate": 0.000289330960768306, "loss": 5.4572, "step": 142850 }, { "epoch": 0.8616635110526887, "grad_norm": 1.1157246828079224, "learning_rate": 0.0002893234222663687, "loss": 5.4934, "step": 142900 }, { "epoch": 0.8619650028340228, "grad_norm": 1.2574434280395508, "learning_rate": 0.00028931588120038263, "loss": 5.1827, "step": 142950 }, { "epoch": 0.8622664946153568, "grad_norm": 0.7771114706993103, "learning_rate": 0.00028930833757048665, "loss": 5.6248, "step": 143000 }, { "epoch": 0.8625679863966909, "grad_norm": 1.0829697847366333, "learning_rate": 0.0002893007913768194, "loss": 4.8531, "step": 143050 }, { "epoch": 0.8628694781780248, "grad_norm": 1.0231046676635742, "learning_rate": 0.00028929324261952, "loss": 5.4479, "step": 143100 }, { "epoch": 0.8631709699593589, "grad_norm": 1.1490973234176636, "learning_rate": 0.00028928569129872716, "loss": 4.905, "step": 143150 }, { "epoch": 0.8634724617406929, "grad_norm": 1.1765950918197632, "learning_rate": 0.00028927813741458, "loss": 5.5329, "step": 143200 }, { "epoch": 0.863773953522027, "grad_norm": 0.961167573928833, "learning_rate": 0.0002892705809672175, "loss": 5.5028, "step": 143250 }, { "epoch": 0.864075445303361, "grad_norm": 1.0280877351760864, "learning_rate": 0.0002892630219567786, "loss": 5.1943, "step": 143300 }, { "epoch": 0.8643769370846951, "grad_norm": 1.0833556652069092, "learning_rate": 0.00028925546038340266, "loss": 5.3187, "step": 143350 }, { "epoch": 0.8646784288660291, "grad_norm": 1.4595706462860107, "learning_rate": 0.00028924789624722863, "loss": 5.0926, "step": 143400 }, { "epoch": 0.8649799206473632, "grad_norm": 0.2801996171474457, "learning_rate": 0.0002892403295483958, "loss": 5.3522, "step": 143450 }, { "epoch": 0.8652814124286972, "grad_norm": 1.8865679502487183, "learning_rate": 0.0002892327602870434, "loss": 5.3158, "step": 143500 }, { "epoch": 0.8655829042100313, "grad_norm": 1.0316210985183716, "learning_rate": 0.00028922518846331076, "loss": 5.3236, "step": 143550 }, { "epoch": 0.8658843959913652, "grad_norm": 1.0600582361221313, "learning_rate": 0.0002892176140773372, "loss": 4.9825, "step": 143600 }, { "epoch": 0.8661858877726993, "grad_norm": 1.0276118516921997, "learning_rate": 0.0002892100371292621, "loss": 5.3475, "step": 143650 }, { "epoch": 0.8664873795540333, "grad_norm": 1.9541202783584595, "learning_rate": 0.00028920245761922497, "loss": 4.832, "step": 143700 }, { "epoch": 0.8667888713353674, "grad_norm": 1.0336891412734985, "learning_rate": 0.00028919487554736524, "loss": 4.7236, "step": 143750 }, { "epoch": 0.8670903631167014, "grad_norm": 2.048583984375, "learning_rate": 0.00028918729091382243, "loss": 5.0903, "step": 143800 }, { "epoch": 0.8673918548980355, "grad_norm": 0.7875738143920898, "learning_rate": 0.00028917970371873626, "loss": 5.3106, "step": 143850 }, { "epoch": 0.8676933466793695, "grad_norm": 2.4229555130004883, "learning_rate": 0.00028917211396224614, "loss": 5.3676, "step": 143900 }, { "epoch": 0.8679948384607036, "grad_norm": 1.7766834497451782, "learning_rate": 0.000289164521644492, "loss": 5.3828, "step": 143950 }, { "epoch": 0.8682963302420376, "grad_norm": 1.4160016775131226, "learning_rate": 0.0002891569267656133, "loss": 5.3118, "step": 144000 }, { "epoch": 0.8682963302420376, "eval_loss": 5.700840950012207, "eval_runtime": 39.2777, "eval_samples_per_second": 13.035, "eval_steps_per_second": 6.518, "eval_tts_loss": 7.140052350346035, "step": 144000 }, { "epoch": 0.8685978220233717, "grad_norm": 1.8698116540908813, "learning_rate": 0.0002891493293257501, "loss": 5.285, "step": 144050 }, { "epoch": 0.8688993138047056, "grad_norm": 1.2387173175811768, "learning_rate": 0.00028914172932504196, "loss": 5.4921, "step": 144100 }, { "epoch": 0.8692008055860397, "grad_norm": 1.5631221532821655, "learning_rate": 0.0002891341267636289, "loss": 5.4771, "step": 144150 }, { "epoch": 0.8695022973673737, "grad_norm": 1.0370080471038818, "learning_rate": 0.0002891265216416508, "loss": 5.5106, "step": 144200 }, { "epoch": 0.8698037891487078, "grad_norm": 0.42564305663108826, "learning_rate": 0.0002891189139592476, "loss": 4.8306, "step": 144250 }, { "epoch": 0.8701052809300418, "grad_norm": 0.569721519947052, "learning_rate": 0.0002891113037165593, "loss": 5.5031, "step": 144300 }, { "epoch": 0.8704067727113759, "grad_norm": 2.0639824867248535, "learning_rate": 0.00028910369091372605, "loss": 5.2487, "step": 144350 }, { "epoch": 0.8707082644927099, "grad_norm": 0.8978548049926758, "learning_rate": 0.0002890960755508878, "loss": 4.9657, "step": 144400 }, { "epoch": 0.871009756274044, "grad_norm": 2.2060816287994385, "learning_rate": 0.00028908845762818476, "loss": 5.1756, "step": 144450 }, { "epoch": 0.871311248055378, "grad_norm": 1.9720770120620728, "learning_rate": 0.0002890808371457572, "loss": 5.6068, "step": 144500 }, { "epoch": 0.8716127398367121, "grad_norm": 1.7249536514282227, "learning_rate": 0.00028907321410374535, "loss": 5.6245, "step": 144550 }, { "epoch": 0.871914231618046, "grad_norm": 1.3802748918533325, "learning_rate": 0.0002890655885022894, "loss": 5.6224, "step": 144600 }, { "epoch": 0.8722157233993801, "grad_norm": 0.9178386926651001, "learning_rate": 0.0002890579603415298, "loss": 5.3625, "step": 144650 }, { "epoch": 0.8725172151807141, "grad_norm": 1.497206211090088, "learning_rate": 0.00028905032962160687, "loss": 5.4499, "step": 144700 }, { "epoch": 0.8728187069620482, "grad_norm": 1.5105074644088745, "learning_rate": 0.000289042696342661, "loss": 5.3903, "step": 144750 }, { "epoch": 0.8731201987433822, "grad_norm": 1.5202513933181763, "learning_rate": 0.0002890350605048328, "loss": 5.2405, "step": 144800 }, { "epoch": 0.8734216905247163, "grad_norm": 0.7239273190498352, "learning_rate": 0.0002890274221082627, "loss": 4.8343, "step": 144850 }, { "epoch": 0.8737231823060503, "grad_norm": 0.6528559327125549, "learning_rate": 0.0002890197811530913, "loss": 5.2667, "step": 144900 }, { "epoch": 0.8740246740873844, "grad_norm": 1.8724254369735718, "learning_rate": 0.00028901213763945924, "loss": 5.6046, "step": 144950 }, { "epoch": 0.8743261658687184, "grad_norm": 1.1181082725524902, "learning_rate": 0.0002890044915675071, "loss": 5.0353, "step": 145000 }, { "epoch": 0.8746276576500525, "grad_norm": 1.1425760984420776, "learning_rate": 0.00028899684293737574, "loss": 4.9809, "step": 145050 }, { "epoch": 0.8749291494313866, "grad_norm": 1.0635451078414917, "learning_rate": 0.0002889891917492058, "loss": 5.2805, "step": 145100 }, { "epoch": 0.8752306412127205, "grad_norm": 1.432281255722046, "learning_rate": 0.00028898153800313814, "loss": 5.003, "step": 145150 }, { "epoch": 0.8755321329940546, "grad_norm": 1.3861051797866821, "learning_rate": 0.00028897388169931364, "loss": 5.3747, "step": 145200 }, { "epoch": 0.8758336247753886, "grad_norm": 1.2200214862823486, "learning_rate": 0.00028896622283787314, "loss": 5.068, "step": 145250 }, { "epoch": 0.8761351165567227, "grad_norm": 1.8978623151779175, "learning_rate": 0.0002889585614189576, "loss": 5.0147, "step": 145300 }, { "epoch": 0.8764366083380567, "grad_norm": 1.120331048965454, "learning_rate": 0.0002889508974427081, "loss": 5.0078, "step": 145350 }, { "epoch": 0.8767381001193908, "grad_norm": 0.49541449546813965, "learning_rate": 0.0002889432309092656, "loss": 5.2024, "step": 145400 }, { "epoch": 0.8770395919007248, "grad_norm": 1.074135422706604, "learning_rate": 0.00028893556181877123, "loss": 5.1865, "step": 145450 }, { "epoch": 0.8773410836820589, "grad_norm": 1.2677806615829468, "learning_rate": 0.0002889278901713661, "loss": 5.0302, "step": 145500 }, { "epoch": 0.8776425754633929, "grad_norm": 0.5982608199119568, "learning_rate": 0.0002889202159671914, "loss": 4.9282, "step": 145550 }, { "epoch": 0.877944067244727, "grad_norm": 2.285259485244751, "learning_rate": 0.0002889125392063884, "loss": 5.2175, "step": 145600 }, { "epoch": 0.8782455590260609, "grad_norm": 1.7274301052093506, "learning_rate": 0.0002889048598890984, "loss": 5.0644, "step": 145650 }, { "epoch": 0.878547050807395, "grad_norm": 1.0626628398895264, "learning_rate": 0.0002888971780154626, "loss": 5.318, "step": 145700 }, { "epoch": 0.878848542588729, "grad_norm": 1.5790388584136963, "learning_rate": 0.0002888894935856225, "loss": 4.9114, "step": 145750 }, { "epoch": 0.8791500343700631, "grad_norm": 1.34740149974823, "learning_rate": 0.00028888180659971947, "loss": 4.9403, "step": 145800 }, { "epoch": 0.8794515261513971, "grad_norm": 1.6845463514328003, "learning_rate": 0.00028887411705789495, "loss": 5.5667, "step": 145850 }, { "epoch": 0.8797530179327312, "grad_norm": 1.1356867551803589, "learning_rate": 0.0002888664249602905, "loss": 5.2518, "step": 145900 }, { "epoch": 0.8800545097140652, "grad_norm": 1.532355785369873, "learning_rate": 0.0002888587303070477, "loss": 5.7311, "step": 145950 }, { "epoch": 0.8803560014953993, "grad_norm": 1.219971776008606, "learning_rate": 0.00028885103309830813, "loss": 5.4707, "step": 146000 }, { "epoch": 0.8806574932767333, "grad_norm": 1.0771408081054688, "learning_rate": 0.0002888433333342135, "loss": 5.6877, "step": 146050 }, { "epoch": 0.8809589850580674, "grad_norm": 1.6511188745498657, "learning_rate": 0.00028883563101490535, "loss": 4.9964, "step": 146100 }, { "epoch": 0.8812604768394013, "grad_norm": 1.3231632709503174, "learning_rate": 0.0002888279261405256, "loss": 5.6081, "step": 146150 }, { "epoch": 0.8815619686207354, "grad_norm": 1.4807732105255127, "learning_rate": 0.00028882021871121593, "loss": 4.8884, "step": 146200 }, { "epoch": 0.8818634604020694, "grad_norm": 1.2028992176055908, "learning_rate": 0.0002888125087271183, "loss": 5.4431, "step": 146250 }, { "epoch": 0.8821649521834035, "grad_norm": 1.4945834875106812, "learning_rate": 0.00028880479618837454, "loss": 4.9943, "step": 146300 }, { "epoch": 0.8824664439647375, "grad_norm": 1.2146224975585938, "learning_rate": 0.0002887970810951266, "loss": 5.3737, "step": 146350 }, { "epoch": 0.8827679357460716, "grad_norm": 2.1729624271392822, "learning_rate": 0.0002887893634475164, "loss": 5.155, "step": 146400 }, { "epoch": 0.8830694275274056, "grad_norm": 1.0776455402374268, "learning_rate": 0.00028878164324568604, "loss": 5.4343, "step": 146450 }, { "epoch": 0.8833709193087397, "grad_norm": 2.0133771896362305, "learning_rate": 0.00028877392048977764, "loss": 4.81, "step": 146500 }, { "epoch": 0.8836724110900737, "grad_norm": 2.227557420730591, "learning_rate": 0.0002887661951799333, "loss": 5.1318, "step": 146550 }, { "epoch": 0.8839739028714078, "grad_norm": 1.0963597297668457, "learning_rate": 0.00028875846731629505, "loss": 5.2111, "step": 146600 }, { "epoch": 0.8842753946527417, "grad_norm": 1.1338919401168823, "learning_rate": 0.00028875073689900536, "loss": 5.0425, "step": 146650 }, { "epoch": 0.8845768864340758, "grad_norm": 1.4744070768356323, "learning_rate": 0.0002887430039282063, "loss": 5.2033, "step": 146700 }, { "epoch": 0.8848783782154098, "grad_norm": 1.444624662399292, "learning_rate": 0.0002887352684040402, "loss": 4.9597, "step": 146750 }, { "epoch": 0.8851798699967439, "grad_norm": 1.3371951580047607, "learning_rate": 0.0002887275303266495, "loss": 5.5268, "step": 146800 }, { "epoch": 0.8854813617780779, "grad_norm": 1.2945256233215332, "learning_rate": 0.0002887197896961766, "loss": 5.4534, "step": 146850 }, { "epoch": 0.885782853559412, "grad_norm": 1.5551913976669312, "learning_rate": 0.00028871204651276396, "loss": 5.1696, "step": 146900 }, { "epoch": 0.886084345340746, "grad_norm": 0.9080455899238586, "learning_rate": 0.00028870430077655403, "loss": 4.957, "step": 146950 }, { "epoch": 0.8863858371220801, "grad_norm": 1.4167227745056152, "learning_rate": 0.0002886965524876894, "loss": 5.1981, "step": 147000 }, { "epoch": 0.8863858371220801, "eval_loss": 5.689861297607422, "eval_runtime": 39.0728, "eval_samples_per_second": 13.104, "eval_steps_per_second": 6.552, "eval_tts_loss": 7.150050787608922, "step": 147000 }, { "epoch": 0.8866873289034141, "grad_norm": 1.6451678276062012, "learning_rate": 0.0002886888016463126, "loss": 4.9504, "step": 147050 }, { "epoch": 0.8869888206847482, "grad_norm": 0.973886251449585, "learning_rate": 0.0002886810482525664, "loss": 5.2724, "step": 147100 }, { "epoch": 0.8872903124660821, "grad_norm": 1.709571361541748, "learning_rate": 0.00028867329230659335, "loss": 4.968, "step": 147150 }, { "epoch": 0.8875918042474162, "grad_norm": 1.291527271270752, "learning_rate": 0.0002886655338085363, "loss": 5.0132, "step": 147200 }, { "epoch": 0.8878932960287502, "grad_norm": 1.3879674673080444, "learning_rate": 0.000288657772758538, "loss": 4.8643, "step": 147250 }, { "epoch": 0.8881947878100843, "grad_norm": 1.1990844011306763, "learning_rate": 0.00028865000915674117, "loss": 5.3349, "step": 147300 }, { "epoch": 0.8884962795914183, "grad_norm": 0.9934024810791016, "learning_rate": 0.00028864224300328886, "loss": 5.1846, "step": 147350 }, { "epoch": 0.8887977713727524, "grad_norm": 1.166325569152832, "learning_rate": 0.00028863447429832394, "loss": 5.336, "step": 147400 }, { "epoch": 0.8890992631540864, "grad_norm": 1.8771528005599976, "learning_rate": 0.0002886267030419893, "loss": 5.0954, "step": 147450 }, { "epoch": 0.8894007549354205, "grad_norm": 1.6549595594406128, "learning_rate": 0.0002886189292344281, "loss": 4.9214, "step": 147500 }, { "epoch": 0.8897022467167545, "grad_norm": 1.1076807975769043, "learning_rate": 0.00028861115287578325, "loss": 5.4934, "step": 147550 }, { "epoch": 0.8900037384980886, "grad_norm": 1.2061291933059692, "learning_rate": 0.000288603373966198, "loss": 4.9398, "step": 147600 }, { "epoch": 0.8903052302794225, "grad_norm": 1.4357142448425293, "learning_rate": 0.0002885955925058154, "loss": 5.5384, "step": 147650 }, { "epoch": 0.8906067220607566, "grad_norm": 1.214878797531128, "learning_rate": 0.00028858780849477876, "loss": 5.2878, "step": 147700 }, { "epoch": 0.8909082138420906, "grad_norm": 1.0492652654647827, "learning_rate": 0.00028858002193323125, "loss": 5.543, "step": 147750 }, { "epoch": 0.8912097056234247, "grad_norm": 1.4032264947891235, "learning_rate": 0.00028857223282131617, "loss": 5.3723, "step": 147800 }, { "epoch": 0.8915111974047587, "grad_norm": 1.607499122619629, "learning_rate": 0.0002885644411591769, "loss": 4.6451, "step": 147850 }, { "epoch": 0.8918126891860928, "grad_norm": 1.0932668447494507, "learning_rate": 0.00028855664694695687, "loss": 5.267, "step": 147900 }, { "epoch": 0.8921141809674268, "grad_norm": 2.039634943008423, "learning_rate": 0.00028854885018479953, "loss": 5.1173, "step": 147950 }, { "epoch": 0.8924156727487609, "grad_norm": 1.0473285913467407, "learning_rate": 0.00028854105087284827, "loss": 5.3178, "step": 148000 }, { "epoch": 0.8927171645300949, "grad_norm": 2.36649489402771, "learning_rate": 0.0002885332490112466, "loss": 5.6175, "step": 148050 }, { "epoch": 0.893018656311429, "grad_norm": 0.6960721015930176, "learning_rate": 0.0002885254446001383, "loss": 5.1344, "step": 148100 }, { "epoch": 0.893320148092763, "grad_norm": 1.3285318613052368, "learning_rate": 0.00028851763763966683, "loss": 4.9779, "step": 148150 }, { "epoch": 0.893621639874097, "grad_norm": 1.0953854322433472, "learning_rate": 0.0002885098281299759, "loss": 5.1414, "step": 148200 }, { "epoch": 0.893923131655431, "grad_norm": 1.4582884311676025, "learning_rate": 0.0002885020160712093, "loss": 5.4941, "step": 148250 }, { "epoch": 0.8942246234367651, "grad_norm": 1.4002840518951416, "learning_rate": 0.0002884942014635107, "loss": 5.2728, "step": 148300 }, { "epoch": 0.8945261152180991, "grad_norm": 1.8111584186553955, "learning_rate": 0.00028848638430702394, "loss": 5.5195, "step": 148350 }, { "epoch": 0.8948276069994332, "grad_norm": 0.54244065284729, "learning_rate": 0.00028847856460189296, "loss": 5.0024, "step": 148400 }, { "epoch": 0.8951290987807672, "grad_norm": 1.5836035013198853, "learning_rate": 0.0002884707423482616, "loss": 5.1892, "step": 148450 }, { "epoch": 0.8954305905621013, "grad_norm": 1.0972826480865479, "learning_rate": 0.00028846291754627384, "loss": 5.3102, "step": 148500 }, { "epoch": 0.8957320823434353, "grad_norm": 1.1006559133529663, "learning_rate": 0.0002884550901960737, "loss": 5.1692, "step": 148550 }, { "epoch": 0.8960335741247694, "grad_norm": 1.3687934875488281, "learning_rate": 0.00028844726029780517, "loss": 5.0361, "step": 148600 }, { "epoch": 0.8963350659061033, "grad_norm": 0.9689916968345642, "learning_rate": 0.00028843942785161236, "loss": 5.2588, "step": 148650 }, { "epoch": 0.8966365576874374, "grad_norm": 0.9383699297904968, "learning_rate": 0.0002884315928576395, "loss": 5.3072, "step": 148700 }, { "epoch": 0.8969380494687715, "grad_norm": 1.2714022397994995, "learning_rate": 0.00028842375531603075, "loss": 5.1175, "step": 148750 }, { "epoch": 0.8972395412501055, "grad_norm": 0.9359409809112549, "learning_rate": 0.00028841591522693025, "loss": 5.3417, "step": 148800 }, { "epoch": 0.8975410330314396, "grad_norm": 2.226539134979248, "learning_rate": 0.00028840807259048243, "loss": 5.1649, "step": 148850 }, { "epoch": 0.8978425248127736, "grad_norm": 1.1385458707809448, "learning_rate": 0.00028840022740683156, "loss": 5.2808, "step": 148900 }, { "epoch": 0.8981440165941077, "grad_norm": 1.0463316440582275, "learning_rate": 0.0002883923796761219, "loss": 4.9349, "step": 148950 }, { "epoch": 0.8984455083754417, "grad_norm": 1.1346328258514404, "learning_rate": 0.0002883845293984981, "loss": 5.1773, "step": 149000 }, { "epoch": 0.8987470001567758, "grad_norm": 0.26701146364212036, "learning_rate": 0.0002883766765741045, "loss": 4.6645, "step": 149050 }, { "epoch": 0.8990484919381098, "grad_norm": 1.4048895835876465, "learning_rate": 0.0002883688212030856, "loss": 4.8376, "step": 149100 }, { "epoch": 0.8993499837194439, "grad_norm": 1.8907647132873535, "learning_rate": 0.00028836096328558605, "loss": 5.2933, "step": 149150 }, { "epoch": 0.8996514755007778, "grad_norm": 0.554224967956543, "learning_rate": 0.00028835310282175044, "loss": 5.1533, "step": 149200 }, { "epoch": 0.8999529672821119, "grad_norm": 1.1079721450805664, "learning_rate": 0.0002883452398117234, "loss": 5.2655, "step": 149250 }, { "epoch": 0.9002544590634459, "grad_norm": 1.1429238319396973, "learning_rate": 0.0002883373742556496, "loss": 5.1802, "step": 149300 }, { "epoch": 0.90055595084478, "grad_norm": 1.5779178142547607, "learning_rate": 0.0002883295061536739, "loss": 4.9917, "step": 149350 }, { "epoch": 0.900857442626114, "grad_norm": 1.4552134275436401, "learning_rate": 0.000288321635505941, "loss": 4.9991, "step": 149400 }, { "epoch": 0.9011589344074481, "grad_norm": 1.118432879447937, "learning_rate": 0.0002883137623125958, "loss": 5.4337, "step": 149450 }, { "epoch": 0.9014604261887821, "grad_norm": 1.49931800365448, "learning_rate": 0.00028830588657378323, "loss": 5.5366, "step": 149500 }, { "epoch": 0.9017619179701162, "grad_norm": 0.6015401482582092, "learning_rate": 0.00028829800828964814, "loss": 5.256, "step": 149550 }, { "epoch": 0.9020634097514502, "grad_norm": 2.1076507568359375, "learning_rate": 0.0002882901274603356, "loss": 5.0206, "step": 149600 }, { "epoch": 0.9023649015327843, "grad_norm": 1.3246594667434692, "learning_rate": 0.0002882822440859906, "loss": 4.5675, "step": 149650 }, { "epoch": 0.9026663933141182, "grad_norm": 1.2611678838729858, "learning_rate": 0.0002882743581667583, "loss": 4.9573, "step": 149700 }, { "epoch": 0.9029678850954523, "grad_norm": 1.1638009548187256, "learning_rate": 0.0002882664697027837, "loss": 5.4794, "step": 149750 }, { "epoch": 0.9032693768767863, "grad_norm": 1.2002309560775757, "learning_rate": 0.000288258578694212, "loss": 5.186, "step": 149800 }, { "epoch": 0.9035708686581204, "grad_norm": 1.5088084936141968, "learning_rate": 0.00028825068514118856, "loss": 5.2854, "step": 149850 }, { "epoch": 0.9038723604394544, "grad_norm": 2.0225725173950195, "learning_rate": 0.00028824278904385854, "loss": 5.3453, "step": 149900 }, { "epoch": 0.9041738522207885, "grad_norm": 1.9504542350769043, "learning_rate": 0.0002882348904023672, "loss": 4.8574, "step": 149950 }, { "epoch": 0.9044753440021225, "grad_norm": 1.8106045722961426, "learning_rate": 0.00028822698921686005, "loss": 5.275, "step": 150000 }, { "epoch": 0.9044753440021225, "eval_loss": 5.6946330070495605, "eval_runtime": 39.0844, "eval_samples_per_second": 13.1, "eval_steps_per_second": 6.55, "eval_tts_loss": 7.182933123204412, "step": 150000 }, { "epoch": 0.9047768357834566, "grad_norm": 1.1548739671707153, "learning_rate": 0.00028821908548748237, "loss": 4.9383, "step": 150050 }, { "epoch": 0.9050783275647906, "grad_norm": 1.1502496004104614, "learning_rate": 0.00028821117921437974, "loss": 5.6535, "step": 150100 }, { "epoch": 0.9053798193461247, "grad_norm": 1.0410882234573364, "learning_rate": 0.0002882032703976975, "loss": 4.835, "step": 150150 }, { "epoch": 0.9056813111274586, "grad_norm": 1.1461116075515747, "learning_rate": 0.0002881953590375814, "loss": 5.1598, "step": 150200 }, { "epoch": 0.9059828029087927, "grad_norm": 1.6753275394439697, "learning_rate": 0.00028818744513417683, "loss": 5.2577, "step": 150250 }, { "epoch": 0.9062842946901267, "grad_norm": 0.9384399056434631, "learning_rate": 0.0002881795286876296, "loss": 5.2718, "step": 150300 }, { "epoch": 0.9065857864714608, "grad_norm": 1.0105115175247192, "learning_rate": 0.0002881716096980853, "loss": 4.7276, "step": 150350 }, { "epoch": 0.9068872782527948, "grad_norm": 1.2735682725906372, "learning_rate": 0.0002881636881656897, "loss": 5.2538, "step": 150400 }, { "epoch": 0.9071887700341289, "grad_norm": 1.2025561332702637, "learning_rate": 0.00028815576409058866, "loss": 5.4834, "step": 150450 }, { "epoch": 0.9074902618154629, "grad_norm": 0.9722779393196106, "learning_rate": 0.0002881478374729279, "loss": 4.9685, "step": 150500 }, { "epoch": 0.907791753596797, "grad_norm": 1.0342727899551392, "learning_rate": 0.00028813990831285334, "loss": 5.2386, "step": 150550 }, { "epoch": 0.908093245378131, "grad_norm": 0.6828578114509583, "learning_rate": 0.0002881319766105109, "loss": 4.9494, "step": 150600 }, { "epoch": 0.9083947371594651, "grad_norm": 1.042580008506775, "learning_rate": 0.00028812404236604655, "loss": 5.2074, "step": 150650 }, { "epoch": 0.908696228940799, "grad_norm": 0.5266819000244141, "learning_rate": 0.0002881161055796063, "loss": 4.7513, "step": 150700 }, { "epoch": 0.9089977207221331, "grad_norm": 1.1887564659118652, "learning_rate": 0.00028810816625133624, "loss": 5.595, "step": 150750 }, { "epoch": 0.9092992125034671, "grad_norm": 1.04921293258667, "learning_rate": 0.0002881002243813825, "loss": 5.3902, "step": 150800 }, { "epoch": 0.9096007042848012, "grad_norm": 1.0684316158294678, "learning_rate": 0.0002880922799698912, "loss": 5.2278, "step": 150850 }, { "epoch": 0.9099021960661352, "grad_norm": 1.5458465814590454, "learning_rate": 0.00028808433301700853, "loss": 5.3949, "step": 150900 }, { "epoch": 0.9102036878474693, "grad_norm": 1.9140100479125977, "learning_rate": 0.0002880763835228808, "loss": 5.5582, "step": 150950 }, { "epoch": 0.9105051796288033, "grad_norm": 0.5953182578086853, "learning_rate": 0.0002880684314876542, "loss": 5.1457, "step": 151000 }, { "epoch": 0.9108066714101374, "grad_norm": 0.889725923538208, "learning_rate": 0.00028806047691147524, "loss": 4.8528, "step": 151050 }, { "epoch": 0.9111081631914714, "grad_norm": 0.7214066982269287, "learning_rate": 0.0002880525197944902, "loss": 5.7267, "step": 151100 }, { "epoch": 0.9114096549728055, "grad_norm": 0.9872736930847168, "learning_rate": 0.00028804456013684554, "loss": 5.1828, "step": 151150 }, { "epoch": 0.9117111467541394, "grad_norm": 1.0250866413116455, "learning_rate": 0.00028803659793868774, "loss": 5.0975, "step": 151200 }, { "epoch": 0.9120126385354735, "grad_norm": 0.7108745574951172, "learning_rate": 0.0002880286332001633, "loss": 5.1171, "step": 151250 }, { "epoch": 0.9123141303168075, "grad_norm": 1.1595265865325928, "learning_rate": 0.00028802066592141897, "loss": 5.3087, "step": 151300 }, { "epoch": 0.9126156220981416, "grad_norm": 1.4652717113494873, "learning_rate": 0.00028801269610260116, "loss": 5.1814, "step": 151350 }, { "epoch": 0.9129171138794756, "grad_norm": 1.3413293361663818, "learning_rate": 0.00028800472374385664, "loss": 4.9008, "step": 151400 }, { "epoch": 0.9132186056608097, "grad_norm": 1.0203913450241089, "learning_rate": 0.00028799674884533215, "loss": 5.0687, "step": 151450 }, { "epoch": 0.9135200974421437, "grad_norm": 1.2094836235046387, "learning_rate": 0.00028798877140717445, "loss": 5.4612, "step": 151500 }, { "epoch": 0.9138215892234778, "grad_norm": 1.061000108718872, "learning_rate": 0.0002879807914295303, "loss": 5.2406, "step": 151550 }, { "epoch": 0.9141230810048118, "grad_norm": 1.6715731620788574, "learning_rate": 0.00028797280891254654, "loss": 5.1401, "step": 151600 }, { "epoch": 0.9144245727861459, "grad_norm": 1.7661603689193726, "learning_rate": 0.0002879648238563702, "loss": 5.0399, "step": 151650 }, { "epoch": 0.9147260645674798, "grad_norm": 0.501611053943634, "learning_rate": 0.00028795683626114813, "loss": 5.097, "step": 151700 }, { "epoch": 0.9150275563488139, "grad_norm": 1.8965076208114624, "learning_rate": 0.0002879488461270274, "loss": 5.6866, "step": 151750 }, { "epoch": 0.9153290481301479, "grad_norm": 1.0662275552749634, "learning_rate": 0.000287940853454155, "loss": 5.4392, "step": 151800 }, { "epoch": 0.915630539911482, "grad_norm": 1.0546038150787354, "learning_rate": 0.000287932858242678, "loss": 5.0555, "step": 151850 }, { "epoch": 0.915932031692816, "grad_norm": 1.4942253828048706, "learning_rate": 0.00028792486049274363, "loss": 5.3017, "step": 151900 }, { "epoch": 0.9162335234741501, "grad_norm": 2.2794580459594727, "learning_rate": 0.00028791686020449903, "loss": 4.8027, "step": 151950 }, { "epoch": 0.9165350152554841, "grad_norm": 1.1135092973709106, "learning_rate": 0.0002879088573780914, "loss": 5.0725, "step": 152000 }, { "epoch": 0.9168365070368182, "grad_norm": 1.1933757066726685, "learning_rate": 0.0002879008520136681, "loss": 5.1936, "step": 152050 }, { "epoch": 0.9171379988181522, "grad_norm": 0.7464392185211182, "learning_rate": 0.00028789284411137643, "loss": 5.395, "step": 152100 }, { "epoch": 0.9174394905994863, "grad_norm": 0.5796415209770203, "learning_rate": 0.00028788483367136374, "loss": 4.9964, "step": 152150 }, { "epoch": 0.9177409823808202, "grad_norm": 1.616685152053833, "learning_rate": 0.00028787682069377743, "loss": 5.5295, "step": 152200 }, { "epoch": 0.9180424741621543, "grad_norm": 1.878179907798767, "learning_rate": 0.000287868805178765, "loss": 4.8437, "step": 152250 }, { "epoch": 0.9183439659434883, "grad_norm": 0.9927974939346313, "learning_rate": 0.00028786078712647394, "loss": 5.0216, "step": 152300 }, { "epoch": 0.9186454577248224, "grad_norm": 1.1809433698654175, "learning_rate": 0.0002878527665370519, "loss": 5.2621, "step": 152350 }, { "epoch": 0.9189469495061564, "grad_norm": 1.063910722732544, "learning_rate": 0.00028784474341064635, "loss": 5.1936, "step": 152400 }, { "epoch": 0.9192484412874905, "grad_norm": 1.0842292308807373, "learning_rate": 0.00028783671774740505, "loss": 5.0836, "step": 152450 }, { "epoch": 0.9195499330688246, "grad_norm": 1.1189279556274414, "learning_rate": 0.0002878286895474756, "loss": 5.1801, "step": 152500 }, { "epoch": 0.9198514248501586, "grad_norm": 1.359502911567688, "learning_rate": 0.0002878206588110059, "loss": 5.4581, "step": 152550 }, { "epoch": 0.9201529166314927, "grad_norm": 0.6721907258033752, "learning_rate": 0.0002878126255381436, "loss": 4.9204, "step": 152600 }, { "epoch": 0.9204544084128267, "grad_norm": 1.1024022102355957, "learning_rate": 0.0002878045897290366, "loss": 5.1031, "step": 152650 }, { "epoch": 0.9207559001941608, "grad_norm": 1.6390401124954224, "learning_rate": 0.00028779655138383285, "loss": 5.4975, "step": 152700 }, { "epoch": 0.9210573919754947, "grad_norm": 1.1761778593063354, "learning_rate": 0.00028778851050268014, "loss": 5.3345, "step": 152750 }, { "epoch": 0.9213588837568288, "grad_norm": 0.669963002204895, "learning_rate": 0.00028778046708572655, "loss": 5.0512, "step": 152800 }, { "epoch": 0.9216603755381628, "grad_norm": 1.03511381149292, "learning_rate": 0.00028777242113312007, "loss": 5.2869, "step": 152850 }, { "epoch": 0.9219618673194969, "grad_norm": 1.080496072769165, "learning_rate": 0.00028776437264500885, "loss": 5.1807, "step": 152900 }, { "epoch": 0.9222633591008309, "grad_norm": 2.061359167098999, "learning_rate": 0.0002877563216215409, "loss": 5.2121, "step": 152950 }, { "epoch": 0.922564850882165, "grad_norm": 1.4358723163604736, "learning_rate": 0.00028774826806286444, "loss": 5.6091, "step": 153000 }, { "epoch": 0.922564850882165, "eval_loss": 5.6906609535217285, "eval_runtime": 39.203, "eval_samples_per_second": 13.06, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.1362446905934345, "step": 153000 }, { "epoch": 0.922866342663499, "grad_norm": 0.3655581772327423, "learning_rate": 0.0002877402119691277, "loss": 4.9846, "step": 153050 }, { "epoch": 0.9231678344448331, "grad_norm": 1.6186728477478027, "learning_rate": 0.00028773215334047895, "loss": 5.2209, "step": 153100 }, { "epoch": 0.9234693262261671, "grad_norm": 2.066009283065796, "learning_rate": 0.0002877240921770664, "loss": 5.1035, "step": 153150 }, { "epoch": 0.9237708180075012, "grad_norm": 1.0419787168502808, "learning_rate": 0.00028771602847903855, "loss": 5.488, "step": 153200 }, { "epoch": 0.9240723097888351, "grad_norm": 1.859923243522644, "learning_rate": 0.0002877079622465437, "loss": 5.4249, "step": 153250 }, { "epoch": 0.9243738015701692, "grad_norm": 1.163357138633728, "learning_rate": 0.0002876998934797303, "loss": 5.0891, "step": 153300 }, { "epoch": 0.9246752933515032, "grad_norm": 1.505170464515686, "learning_rate": 0.00028769182217874687, "loss": 4.88, "step": 153350 }, { "epoch": 0.9249767851328373, "grad_norm": 1.0022644996643066, "learning_rate": 0.000287683748343742, "loss": 5.1738, "step": 153400 }, { "epoch": 0.9252782769141713, "grad_norm": 1.137071132659912, "learning_rate": 0.0002876756719748642, "loss": 5.0722, "step": 153450 }, { "epoch": 0.9255797686955054, "grad_norm": 1.1166737079620361, "learning_rate": 0.0002876675930722621, "loss": 5.1826, "step": 153500 }, { "epoch": 0.9258812604768394, "grad_norm": 1.17910897731781, "learning_rate": 0.0002876595116360844, "loss": 5.4838, "step": 153550 }, { "epoch": 0.9261827522581735, "grad_norm": 1.3166781663894653, "learning_rate": 0.0002876514276664799, "loss": 4.82, "step": 153600 }, { "epoch": 0.9264842440395075, "grad_norm": 1.0418568849563599, "learning_rate": 0.00028764334116359734, "loss": 5.3143, "step": 153650 }, { "epoch": 0.9267857358208416, "grad_norm": 1.6072839498519897, "learning_rate": 0.00028763525212758547, "loss": 5.2927, "step": 153700 }, { "epoch": 0.9270872276021755, "grad_norm": 0.700353741645813, "learning_rate": 0.00028762716055859316, "loss": 5.0088, "step": 153750 }, { "epoch": 0.9273887193835096, "grad_norm": 1.277675747871399, "learning_rate": 0.0002876190664567694, "loss": 5.2272, "step": 153800 }, { "epoch": 0.9276902111648436, "grad_norm": 2.418794631958008, "learning_rate": 0.0002876109698222631, "loss": 5.405, "step": 153850 }, { "epoch": 0.9279917029461777, "grad_norm": 0.8814477324485779, "learning_rate": 0.0002876028706552233, "loss": 5.5669, "step": 153900 }, { "epoch": 0.9282931947275117, "grad_norm": 0.617274820804596, "learning_rate": 0.00028759476895579906, "loss": 5.2878, "step": 153950 }, { "epoch": 0.9285946865088458, "grad_norm": 1.4597575664520264, "learning_rate": 0.00028758666472413944, "loss": 5.6214, "step": 154000 }, { "epoch": 0.9288961782901798, "grad_norm": 1.8305727243423462, "learning_rate": 0.0002875785579603936, "loss": 5.0896, "step": 154050 }, { "epoch": 0.9291976700715139, "grad_norm": 1.856857180595398, "learning_rate": 0.00028757044866471065, "loss": 5.5281, "step": 154100 }, { "epoch": 0.9294991618528479, "grad_norm": 0.4971928894519806, "learning_rate": 0.00028756233683724, "loss": 5.2698, "step": 154150 }, { "epoch": 0.929800653634182, "grad_norm": 1.3152552843093872, "learning_rate": 0.00028755422247813084, "loss": 5.0668, "step": 154200 }, { "epoch": 0.930102145415516, "grad_norm": 1.8337149620056152, "learning_rate": 0.00028754610558753255, "loss": 5.6653, "step": 154250 }, { "epoch": 0.93040363719685, "grad_norm": 1.1786352396011353, "learning_rate": 0.0002875379861655944, "loss": 5.2455, "step": 154300 }, { "epoch": 0.930705128978184, "grad_norm": 1.0302740335464478, "learning_rate": 0.000287529864212466, "loss": 4.9765, "step": 154350 }, { "epoch": 0.9310066207595181, "grad_norm": 1.0717605352401733, "learning_rate": 0.00028752173972829663, "loss": 5.4006, "step": 154400 }, { "epoch": 0.9313081125408521, "grad_norm": 1.4475189447402954, "learning_rate": 0.0002875136127132359, "loss": 5.2206, "step": 154450 }, { "epoch": 0.9316096043221862, "grad_norm": 0.7747493386268616, "learning_rate": 0.0002875054831674334, "loss": 5.3533, "step": 154500 }, { "epoch": 0.9319110961035202, "grad_norm": 1.165152907371521, "learning_rate": 0.00028749735109103873, "loss": 4.7424, "step": 154550 }, { "epoch": 0.9322125878848543, "grad_norm": 0.5932610630989075, "learning_rate": 0.0002874892164842014, "loss": 5.3337, "step": 154600 }, { "epoch": 0.9325140796661883, "grad_norm": 2.546983003616333, "learning_rate": 0.0002874810793470714, "loss": 5.8651, "step": 154650 }, { "epoch": 0.9328155714475224, "grad_norm": 1.5157440900802612, "learning_rate": 0.0002874729396797983, "loss": 5.3739, "step": 154700 }, { "epoch": 0.9331170632288563, "grad_norm": 1.8677711486816406, "learning_rate": 0.0002874647974825319, "loss": 5.2363, "step": 154750 }, { "epoch": 0.9334185550101904, "grad_norm": 1.1460150480270386, "learning_rate": 0.0002874566527554221, "loss": 4.7052, "step": 154800 }, { "epoch": 0.9337200467915244, "grad_norm": 1.6958034038543701, "learning_rate": 0.00028744850549861876, "loss": 5.5336, "step": 154850 }, { "epoch": 0.9340215385728585, "grad_norm": 1.7212144136428833, "learning_rate": 0.0002874403557122718, "loss": 5.0002, "step": 154900 }, { "epoch": 0.9343230303541925, "grad_norm": 1.2491487264633179, "learning_rate": 0.00028743220339653123, "loss": 4.9551, "step": 154950 }, { "epoch": 0.9346245221355266, "grad_norm": 1.1535229682922363, "learning_rate": 0.00028742404855154716, "loss": 5.9734, "step": 155000 }, { "epoch": 0.9349260139168606, "grad_norm": 1.476196527481079, "learning_rate": 0.0002874158911774695, "loss": 5.0294, "step": 155050 }, { "epoch": 0.9352275056981947, "grad_norm": 1.683231234550476, "learning_rate": 0.0002874077312744485, "loss": 4.9787, "step": 155100 }, { "epoch": 0.9355289974795287, "grad_norm": 1.4417330026626587, "learning_rate": 0.00028739956884263427, "loss": 4.9103, "step": 155150 }, { "epoch": 0.9358304892608628, "grad_norm": 1.4914827346801758, "learning_rate": 0.0002873914038821771, "loss": 5.4068, "step": 155200 }, { "epoch": 0.9361319810421967, "grad_norm": 1.3618189096450806, "learning_rate": 0.0002873832363932272, "loss": 5.4847, "step": 155250 }, { "epoch": 0.9364334728235308, "grad_norm": 1.014369010925293, "learning_rate": 0.0002873750663759349, "loss": 5.1383, "step": 155300 }, { "epoch": 0.9367349646048648, "grad_norm": 1.7547460794448853, "learning_rate": 0.0002873668938304505, "loss": 5.3154, "step": 155350 }, { "epoch": 0.9370364563861989, "grad_norm": 2.9673752784729004, "learning_rate": 0.0002873587187569245, "loss": 5.2074, "step": 155400 }, { "epoch": 0.9373379481675329, "grad_norm": 0.9821957945823669, "learning_rate": 0.00028735054115550725, "loss": 5.7843, "step": 155450 }, { "epoch": 0.937639439948867, "grad_norm": 1.254030704498291, "learning_rate": 0.0002873423610263494, "loss": 5.9619, "step": 155500 }, { "epoch": 0.937940931730201, "grad_norm": 1.316757082939148, "learning_rate": 0.0002873341783696013, "loss": 4.849, "step": 155550 }, { "epoch": 0.9382424235115351, "grad_norm": 0.7972604632377625, "learning_rate": 0.00028732599318541364, "loss": 4.7908, "step": 155600 }, { "epoch": 0.9385439152928691, "grad_norm": 0.48111000657081604, "learning_rate": 0.0002873178054739371, "loss": 4.95, "step": 155650 }, { "epoch": 0.9388454070742032, "grad_norm": 1.0949296951293945, "learning_rate": 0.00028730961523532224, "loss": 5.1037, "step": 155700 }, { "epoch": 0.9391468988555371, "grad_norm": 1.787550449371338, "learning_rate": 0.0002873014224697199, "loss": 4.8882, "step": 155750 }, { "epoch": 0.9394483906368712, "grad_norm": 1.464937448501587, "learning_rate": 0.00028729322717728083, "loss": 5.3629, "step": 155800 }, { "epoch": 0.9397498824182052, "grad_norm": 1.141615867614746, "learning_rate": 0.00028728502935815585, "loss": 5.4957, "step": 155850 }, { "epoch": 0.9400513741995393, "grad_norm": 0.6396550536155701, "learning_rate": 0.00028727682901249585, "loss": 5.0265, "step": 155900 }, { "epoch": 0.9403528659808733, "grad_norm": 0.38299790024757385, "learning_rate": 0.00028726862614045164, "loss": 4.7687, "step": 155950 }, { "epoch": 0.9406543577622074, "grad_norm": 1.1170941591262817, "learning_rate": 0.0002872604207421743, "loss": 4.8051, "step": 156000 }, { "epoch": 0.9406543577622074, "eval_loss": 5.681239128112793, "eval_runtime": 39.0415, "eval_samples_per_second": 13.114, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.226691822999492, "step": 156000 }, { "epoch": 0.9409558495435414, "grad_norm": 1.1835739612579346, "learning_rate": 0.00028725221281781485, "loss": 5.231, "step": 156050 }, { "epoch": 0.9412573413248755, "grad_norm": 1.994027853012085, "learning_rate": 0.0002872440023675242, "loss": 5.2578, "step": 156100 }, { "epoch": 0.9415588331062095, "grad_norm": 2.405381441116333, "learning_rate": 0.00028723578939145357, "loss": 5.3678, "step": 156150 }, { "epoch": 0.9418603248875436, "grad_norm": 1.3751730918884277, "learning_rate": 0.00028722757388975406, "loss": 4.9882, "step": 156200 }, { "epoch": 0.9421618166688777, "grad_norm": 1.8018766641616821, "learning_rate": 0.00028721935586257694, "loss": 5.1177, "step": 156250 }, { "epoch": 0.9424633084502116, "grad_norm": 1.3992455005645752, "learning_rate": 0.0002872111353100734, "loss": 4.9875, "step": 156300 }, { "epoch": 0.9427648002315457, "grad_norm": 1.5679166316986084, "learning_rate": 0.0002872029122323947, "loss": 5.2126, "step": 156350 }, { "epoch": 0.9430662920128797, "grad_norm": 1.0654597282409668, "learning_rate": 0.0002871946866296922, "loss": 5.4371, "step": 156400 }, { "epoch": 0.9433677837942138, "grad_norm": 1.3742411136627197, "learning_rate": 0.0002871864585021173, "loss": 5.415, "step": 156450 }, { "epoch": 0.9436692755755478, "grad_norm": 1.3297288417816162, "learning_rate": 0.00028717822784982133, "loss": 5.5263, "step": 156500 }, { "epoch": 0.9439707673568819, "grad_norm": 1.7704334259033203, "learning_rate": 0.0002871699946729559, "loss": 5.495, "step": 156550 }, { "epoch": 0.9442722591382159, "grad_norm": 0.570009171962738, "learning_rate": 0.0002871617589716725, "loss": 4.9708, "step": 156600 }, { "epoch": 0.94457375091955, "grad_norm": 1.4769442081451416, "learning_rate": 0.0002871535207461226, "loss": 5.3203, "step": 156650 }, { "epoch": 0.944875242700884, "grad_norm": 1.6025233268737793, "learning_rate": 0.0002871452799964579, "loss": 5.4954, "step": 156700 }, { "epoch": 0.9451767344822181, "grad_norm": 1.0084046125411987, "learning_rate": 0.0002871370367228301, "loss": 5.1355, "step": 156750 }, { "epoch": 0.945478226263552, "grad_norm": 1.0468761920928955, "learning_rate": 0.00028712879092539077, "loss": 5.4674, "step": 156800 }, { "epoch": 0.9457797180448861, "grad_norm": 1.6450432538986206, "learning_rate": 0.00028712054260429177, "loss": 5.1877, "step": 156850 }, { "epoch": 0.9460812098262201, "grad_norm": 3.543611526489258, "learning_rate": 0.0002871122917596849, "loss": 5.8227, "step": 156900 }, { "epoch": 0.9463827016075542, "grad_norm": 2.016862630844116, "learning_rate": 0.00028710403839172187, "loss": 5.7628, "step": 156950 }, { "epoch": 0.9466841933888882, "grad_norm": 1.1392916440963745, "learning_rate": 0.00028709578250055476, "loss": 5.0629, "step": 157000 }, { "epoch": 0.9469856851702223, "grad_norm": 1.462442398071289, "learning_rate": 0.00028708752408633536, "loss": 5.6797, "step": 157050 }, { "epoch": 0.9472871769515563, "grad_norm": 1.08036470413208, "learning_rate": 0.0002870792631492157, "loss": 5.1391, "step": 157100 }, { "epoch": 0.9475886687328904, "grad_norm": 0.9671501517295837, "learning_rate": 0.0002870709996893479, "loss": 5.4253, "step": 157150 }, { "epoch": 0.9478901605142244, "grad_norm": 1.2123503684997559, "learning_rate": 0.00028706273370688397, "loss": 5.4374, "step": 157200 }, { "epoch": 0.9481916522955585, "grad_norm": 1.5081409215927124, "learning_rate": 0.000287054465201976, "loss": 5.199, "step": 157250 }, { "epoch": 0.9484931440768924, "grad_norm": 1.6117068529129028, "learning_rate": 0.00028704619417477614, "loss": 4.9359, "step": 157300 }, { "epoch": 0.9487946358582265, "grad_norm": 0.45961794257164, "learning_rate": 0.0002870379206254367, "loss": 5.438, "step": 157350 }, { "epoch": 0.9490961276395605, "grad_norm": 0.7751233577728271, "learning_rate": 0.00028702964455410985, "loss": 4.9679, "step": 157400 }, { "epoch": 0.9493976194208946, "grad_norm": 0.9068114161491394, "learning_rate": 0.00028702136596094795, "loss": 5.119, "step": 157450 }, { "epoch": 0.9496991112022286, "grad_norm": 1.352318525314331, "learning_rate": 0.00028701308484610337, "loss": 5.5071, "step": 157500 }, { "epoch": 0.9500006029835627, "grad_norm": 1.2668931484222412, "learning_rate": 0.0002870048012097285, "loss": 5.2963, "step": 157550 }, { "epoch": 0.9503020947648967, "grad_norm": 1.4168813228607178, "learning_rate": 0.00028699651505197576, "loss": 5.2214, "step": 157600 }, { "epoch": 0.9506035865462308, "grad_norm": 1.1288690567016602, "learning_rate": 0.00028698822637299767, "loss": 5.3165, "step": 157650 }, { "epoch": 0.9509050783275648, "grad_norm": 1.121492624282837, "learning_rate": 0.00028697993517294676, "loss": 5.3004, "step": 157700 }, { "epoch": 0.9512065701088989, "grad_norm": 1.1793073415756226, "learning_rate": 0.00028697164145197564, "loss": 5.3731, "step": 157750 }, { "epoch": 0.9515080618902328, "grad_norm": 1.059881329536438, "learning_rate": 0.00028696334521023697, "loss": 5.3584, "step": 157800 }, { "epoch": 0.9518095536715669, "grad_norm": 1.123091697692871, "learning_rate": 0.00028695504644788333, "loss": 5.2546, "step": 157850 }, { "epoch": 0.9521110454529009, "grad_norm": 1.6565293073654175, "learning_rate": 0.0002869467451650675, "loss": 5.0883, "step": 157900 }, { "epoch": 0.952412537234235, "grad_norm": 0.8356490731239319, "learning_rate": 0.00028693844136194226, "loss": 4.9696, "step": 157950 }, { "epoch": 0.952714029015569, "grad_norm": 1.4571679830551147, "learning_rate": 0.00028693013503866045, "loss": 5.2811, "step": 158000 }, { "epoch": 0.9530155207969031, "grad_norm": 2.6171536445617676, "learning_rate": 0.00028692182619537496, "loss": 5.5091, "step": 158050 }, { "epoch": 0.9533170125782371, "grad_norm": 1.0920251607894897, "learning_rate": 0.00028691351483223855, "loss": 5.1491, "step": 158100 }, { "epoch": 0.9536185043595712, "grad_norm": 1.46067476272583, "learning_rate": 0.00028690520094940437, "loss": 4.9562, "step": 158150 }, { "epoch": 0.9539199961409052, "grad_norm": 1.5288447141647339, "learning_rate": 0.0002868968845470253, "loss": 4.5564, "step": 158200 }, { "epoch": 0.9542214879222393, "grad_norm": 1.9776289463043213, "learning_rate": 0.00028688856562525443, "loss": 5.4138, "step": 158250 }, { "epoch": 0.9545229797035732, "grad_norm": 1.3882122039794922, "learning_rate": 0.00028688024418424486, "loss": 5.3749, "step": 158300 }, { "epoch": 0.9548244714849073, "grad_norm": 1.1053030490875244, "learning_rate": 0.00028687192022414977, "loss": 5.0248, "step": 158350 }, { "epoch": 0.9551259632662413, "grad_norm": 1.5714857578277588, "learning_rate": 0.0002868635937451223, "loss": 5.2405, "step": 158400 }, { "epoch": 0.9554274550475754, "grad_norm": 1.7433533668518066, "learning_rate": 0.0002868552647473157, "loss": 5.3064, "step": 158450 }, { "epoch": 0.9557289468289094, "grad_norm": 1.7518503665924072, "learning_rate": 0.00028684693323088325, "loss": 4.688, "step": 158500 }, { "epoch": 0.9560304386102435, "grad_norm": 1.314279317855835, "learning_rate": 0.00028683859919597824, "loss": 5.0342, "step": 158550 }, { "epoch": 0.9563319303915775, "grad_norm": 1.0173089504241943, "learning_rate": 0.0002868302626427542, "loss": 5.0756, "step": 158600 }, { "epoch": 0.9566334221729116, "grad_norm": 1.5423470735549927, "learning_rate": 0.00028682192357136434, "loss": 5.1675, "step": 158650 }, { "epoch": 0.9569349139542456, "grad_norm": 2.1208667755126953, "learning_rate": 0.00028681358198196226, "loss": 5.1209, "step": 158700 }, { "epoch": 0.9572364057355797, "grad_norm": 1.402405858039856, "learning_rate": 0.00028680523787470144, "loss": 5.254, "step": 158750 }, { "epoch": 0.9575378975169136, "grad_norm": 0.9559354782104492, "learning_rate": 0.0002867968912497355, "loss": 5.172, "step": 158800 }, { "epoch": 0.9578393892982477, "grad_norm": 1.6275203227996826, "learning_rate": 0.00028678854210721795, "loss": 5.2154, "step": 158850 }, { "epoch": 0.9581408810795817, "grad_norm": 1.0680928230285645, "learning_rate": 0.0002867801904473025, "loss": 5.3702, "step": 158900 }, { "epoch": 0.9584423728609158, "grad_norm": 1.095552682876587, "learning_rate": 0.0002867718362701428, "loss": 4.8672, "step": 158950 }, { "epoch": 0.9587438646422498, "grad_norm": 0.6702092885971069, "learning_rate": 0.0002867634795758927, "loss": 5.1401, "step": 159000 }, { "epoch": 0.9587438646422498, "eval_loss": 5.679960250854492, "eval_runtime": 39.0431, "eval_samples_per_second": 13.114, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.294058940889374, "step": 159000 }, { "epoch": 0.9590453564235839, "grad_norm": 1.2164103984832764, "learning_rate": 0.0002867551203647059, "loss": 5.499, "step": 159050 }, { "epoch": 0.9593468482049179, "grad_norm": 1.8155300617218018, "learning_rate": 0.0002867467586367363, "loss": 5.3904, "step": 159100 }, { "epoch": 0.959648339986252, "grad_norm": 1.6162075996398926, "learning_rate": 0.0002867383943921378, "loss": 4.8073, "step": 159150 }, { "epoch": 0.959949831767586, "grad_norm": 0.47649192810058594, "learning_rate": 0.0002867300276310642, "loss": 4.8816, "step": 159200 }, { "epoch": 0.9602513235489201, "grad_norm": 0.7019538283348083, "learning_rate": 0.0002867216583536696, "loss": 5.1414, "step": 159250 }, { "epoch": 0.960552815330254, "grad_norm": 1.2072488069534302, "learning_rate": 0.00028671328656010796, "loss": 5.3454, "step": 159300 }, { "epoch": 0.9608543071115881, "grad_norm": 1.2450177669525146, "learning_rate": 0.0002867049122505334, "loss": 4.7721, "step": 159350 }, { "epoch": 0.9611557988929221, "grad_norm": 1.6000854969024658, "learning_rate": 0.0002866965354251, "loss": 5.2248, "step": 159400 }, { "epoch": 0.9614572906742562, "grad_norm": 1.005121111869812, "learning_rate": 0.000286688156083962, "loss": 5.1478, "step": 159450 }, { "epoch": 0.9617587824555902, "grad_norm": 1.1473897695541382, "learning_rate": 0.0002866797742272735, "loss": 5.2361, "step": 159500 }, { "epoch": 0.9620602742369243, "grad_norm": 1.8352391719818115, "learning_rate": 0.00028667138985518883, "loss": 4.9448, "step": 159550 }, { "epoch": 0.9623617660182583, "grad_norm": 1.0379916429519653, "learning_rate": 0.0002866630029678623, "loss": 5.0667, "step": 159600 }, { "epoch": 0.9626632577995924, "grad_norm": 0.9576144218444824, "learning_rate": 0.00028665461356544813, "loss": 5.3567, "step": 159650 }, { "epoch": 0.9629647495809264, "grad_norm": 1.0531004667282104, "learning_rate": 0.0002866462216481009, "loss": 5.2187, "step": 159700 }, { "epoch": 0.9632662413622605, "grad_norm": 1.5523455142974854, "learning_rate": 0.00028663782721597493, "loss": 5.0122, "step": 159750 }, { "epoch": 0.9635677331435945, "grad_norm": 0.5603106617927551, "learning_rate": 0.00028662943026922476, "loss": 4.6042, "step": 159800 }, { "epoch": 0.9638692249249285, "grad_norm": 2.8877851963043213, "learning_rate": 0.0002866210308080049, "loss": 4.9082, "step": 159850 }, { "epoch": 0.9641707167062626, "grad_norm": 1.017879605293274, "learning_rate": 0.00028661262883246993, "loss": 5.5397, "step": 159900 }, { "epoch": 0.9644722084875966, "grad_norm": 1.854726791381836, "learning_rate": 0.0002866042243427745, "loss": 5.4314, "step": 159950 }, { "epoch": 0.9647737002689307, "grad_norm": 1.6658324003219604, "learning_rate": 0.0002865958173390732, "loss": 4.666, "step": 160000 }, { "epoch": 0.9650751920502647, "grad_norm": 1.548567295074463, "learning_rate": 0.0002865874078215209, "loss": 5.3663, "step": 160050 }, { "epoch": 0.9653766838315988, "grad_norm": 2.081678867340088, "learning_rate": 0.00028657899579027224, "loss": 5.4674, "step": 160100 }, { "epoch": 0.9656781756129328, "grad_norm": 1.1595813035964966, "learning_rate": 0.00028657058124548207, "loss": 5.4405, "step": 160150 }, { "epoch": 0.9659796673942669, "grad_norm": 1.64620041847229, "learning_rate": 0.00028656216418730527, "loss": 5.4475, "step": 160200 }, { "epoch": 0.9662811591756009, "grad_norm": 2.540269613265991, "learning_rate": 0.0002865537446158967, "loss": 4.9371, "step": 160250 }, { "epoch": 0.966582650956935, "grad_norm": 0.6362331509590149, "learning_rate": 0.00028654532253141133, "loss": 4.9845, "step": 160300 }, { "epoch": 0.966884142738269, "grad_norm": 1.3657262325286865, "learning_rate": 0.00028653689793400417, "loss": 5.119, "step": 160350 }, { "epoch": 0.967185634519603, "grad_norm": 1.169128179550171, "learning_rate": 0.0002865284708238303, "loss": 5.5219, "step": 160400 }, { "epoch": 0.967487126300937, "grad_norm": 1.4556084871292114, "learning_rate": 0.0002865200412010447, "loss": 5.1598, "step": 160450 }, { "epoch": 0.9677886180822711, "grad_norm": 1.0472179651260376, "learning_rate": 0.0002865116090658026, "loss": 5.1488, "step": 160500 }, { "epoch": 0.9680901098636051, "grad_norm": 0.9404919743537903, "learning_rate": 0.0002865031744182591, "loss": 4.9824, "step": 160550 }, { "epoch": 0.9683916016449392, "grad_norm": 1.015669584274292, "learning_rate": 0.0002864947372585695, "loss": 5.2354, "step": 160600 }, { "epoch": 0.9686930934262732, "grad_norm": 0.7669971585273743, "learning_rate": 0.0002864862975868891, "loss": 4.958, "step": 160650 }, { "epoch": 0.9689945852076073, "grad_norm": 1.477534532546997, "learning_rate": 0.0002864778554033731, "loss": 5.1885, "step": 160700 }, { "epoch": 0.9692960769889413, "grad_norm": 1.350481390953064, "learning_rate": 0.00028646941070817697, "loss": 5.1563, "step": 160750 }, { "epoch": 0.9695975687702754, "grad_norm": 1.3967812061309814, "learning_rate": 0.0002864609635014561, "loss": 4.8746, "step": 160800 }, { "epoch": 0.9698990605516093, "grad_norm": 1.5228937864303589, "learning_rate": 0.0002864525137833659, "loss": 5.5983, "step": 160850 }, { "epoch": 0.9702005523329434, "grad_norm": 1.1704891920089722, "learning_rate": 0.00028644406155406193, "loss": 5.2442, "step": 160900 }, { "epoch": 0.9705020441142774, "grad_norm": 1.3065762519836426, "learning_rate": 0.00028643560681369974, "loss": 5.2957, "step": 160950 }, { "epoch": 0.9708035358956115, "grad_norm": 1.6861201524734497, "learning_rate": 0.0002864271495624349, "loss": 4.8506, "step": 161000 }, { "epoch": 0.9711050276769455, "grad_norm": 1.1208776235580444, "learning_rate": 0.00028641868980042303, "loss": 5.0608, "step": 161050 }, { "epoch": 0.9714065194582796, "grad_norm": 1.1450281143188477, "learning_rate": 0.00028641022752781984, "loss": 4.9869, "step": 161100 }, { "epoch": 0.9717080112396136, "grad_norm": 1.1875343322753906, "learning_rate": 0.00028640176274478116, "loss": 5.0509, "step": 161150 }, { "epoch": 0.9720095030209477, "grad_norm": 1.0246950387954712, "learning_rate": 0.00028639329545146263, "loss": 5.1882, "step": 161200 }, { "epoch": 0.9723109948022817, "grad_norm": 1.0598448514938354, "learning_rate": 0.0002863848256480202, "loss": 5.1591, "step": 161250 }, { "epoch": 0.9726124865836158, "grad_norm": 1.6605243682861328, "learning_rate": 0.00028637635333460964, "loss": 5.4218, "step": 161300 }, { "epoch": 0.9729139783649497, "grad_norm": 2.2580056190490723, "learning_rate": 0.0002863678785113869, "loss": 4.8246, "step": 161350 }, { "epoch": 0.9732154701462838, "grad_norm": 1.3645694255828857, "learning_rate": 0.000286359401178508, "loss": 4.9687, "step": 161400 }, { "epoch": 0.9735169619276178, "grad_norm": 1.366003394126892, "learning_rate": 0.0002863509213361289, "loss": 5.4448, "step": 161450 }, { "epoch": 0.9738184537089519, "grad_norm": 1.0223708152770996, "learning_rate": 0.00028634243898440566, "loss": 5.2802, "step": 161500 }, { "epoch": 0.9741199454902859, "grad_norm": 1.0738648176193237, "learning_rate": 0.00028633395412349446, "loss": 5.6154, "step": 161550 }, { "epoch": 0.97442143727162, "grad_norm": 1.1313939094543457, "learning_rate": 0.0002863254667535513, "loss": 5.4401, "step": 161600 }, { "epoch": 0.974722929052954, "grad_norm": 2.240259885787964, "learning_rate": 0.0002863169768747326, "loss": 5.5265, "step": 161650 }, { "epoch": 0.9750244208342881, "grad_norm": 1.126526951789856, "learning_rate": 0.0002863084844871944, "loss": 5.2646, "step": 161700 }, { "epoch": 0.9753259126156221, "grad_norm": 1.4702166318893433, "learning_rate": 0.000286299989591093, "loss": 5.769, "step": 161750 }, { "epoch": 0.9756274043969562, "grad_norm": 1.3935444355010986, "learning_rate": 0.00028629149218658493, "loss": 5.414, "step": 161800 }, { "epoch": 0.9759288961782902, "grad_norm": 1.319547176361084, "learning_rate": 0.0002862829922738264, "loss": 5.2083, "step": 161850 }, { "epoch": 0.9762303879596242, "grad_norm": 0.9943134784698486, "learning_rate": 0.0002862744898529738, "loss": 5.2454, "step": 161900 }, { "epoch": 0.9765318797409582, "grad_norm": 1.6352397203445435, "learning_rate": 0.0002862659849241838, "loss": 5.4468, "step": 161950 }, { "epoch": 0.9768333715222923, "grad_norm": 1.342061161994934, "learning_rate": 0.00028625747748761276, "loss": 5.215, "step": 162000 }, { "epoch": 0.9768333715222923, "eval_loss": 5.658607482910156, "eval_runtime": 39.2399, "eval_samples_per_second": 13.048, "eval_steps_per_second": 6.524, "eval_tts_loss": 7.217766798108572, "step": 162000 }, { "epoch": 0.9771348633036263, "grad_norm": 1.1942977905273438, "learning_rate": 0.0002862489675434173, "loss": 5.474, "step": 162050 }, { "epoch": 0.9774363550849604, "grad_norm": 1.0020674467086792, "learning_rate": 0.00028624045509175405, "loss": 5.2914, "step": 162100 }, { "epoch": 0.9777378468662944, "grad_norm": 1.9230053424835205, "learning_rate": 0.00028623194013277964, "loss": 5.0359, "step": 162150 }, { "epoch": 0.9780393386476285, "grad_norm": 1.5087312459945679, "learning_rate": 0.00028622342266665076, "loss": 5.5223, "step": 162200 }, { "epoch": 0.9783408304289625, "grad_norm": 1.1398727893829346, "learning_rate": 0.00028621490269352417, "loss": 5.2105, "step": 162250 }, { "epoch": 0.9786423222102966, "grad_norm": 0.5290941596031189, "learning_rate": 0.00028620638021355677, "loss": 5.116, "step": 162300 }, { "epoch": 0.9789438139916306, "grad_norm": 1.6581902503967285, "learning_rate": 0.0002861978552269052, "loss": 5.4044, "step": 162350 }, { "epoch": 0.9792453057729646, "grad_norm": 1.1701014041900635, "learning_rate": 0.00028618932773372656, "loss": 5.1144, "step": 162400 }, { "epoch": 0.9795467975542986, "grad_norm": 1.670994520187378, "learning_rate": 0.00028618079773417773, "loss": 5.0301, "step": 162450 }, { "epoch": 0.9798482893356327, "grad_norm": 1.4540083408355713, "learning_rate": 0.0002861722652284156, "loss": 5.2099, "step": 162500 }, { "epoch": 0.9801497811169667, "grad_norm": 1.7801440954208374, "learning_rate": 0.00028616373021659724, "loss": 4.99, "step": 162550 }, { "epoch": 0.9804512728983008, "grad_norm": 1.4987680912017822, "learning_rate": 0.00028615519269887976, "loss": 5.2208, "step": 162600 }, { "epoch": 0.9807527646796348, "grad_norm": 1.4012480974197388, "learning_rate": 0.00028614665267542025, "loss": 4.7825, "step": 162650 }, { "epoch": 0.9810542564609689, "grad_norm": 1.1021407842636108, "learning_rate": 0.0002861381101463759, "loss": 5.1745, "step": 162700 }, { "epoch": 0.9813557482423029, "grad_norm": 1.7445770502090454, "learning_rate": 0.0002861295651119039, "loss": 5.3625, "step": 162750 }, { "epoch": 0.981657240023637, "grad_norm": 1.1012113094329834, "learning_rate": 0.0002861210175721616, "loss": 5.549, "step": 162800 }, { "epoch": 0.981958731804971, "grad_norm": 1.2392910718917847, "learning_rate": 0.00028611246752730614, "loss": 4.9727, "step": 162850 }, { "epoch": 0.982260223586305, "grad_norm": 1.3255871534347534, "learning_rate": 0.000286103914977495, "loss": 4.9333, "step": 162900 }, { "epoch": 0.982561715367639, "grad_norm": 1.3872698545455933, "learning_rate": 0.0002860953599228855, "loss": 5.5728, "step": 162950 }, { "epoch": 0.9828632071489731, "grad_norm": 0.9957082271575928, "learning_rate": 0.0002860868023636351, "loss": 4.793, "step": 163000 }, { "epoch": 0.9831646989303071, "grad_norm": 1.5318853855133057, "learning_rate": 0.0002860782422999014, "loss": 4.8254, "step": 163050 }, { "epoch": 0.9834661907116412, "grad_norm": 1.603813648223877, "learning_rate": 0.00028606967973184174, "loss": 5.019, "step": 163100 }, { "epoch": 0.9837676824929752, "grad_norm": 1.4190490245819092, "learning_rate": 0.00028606111465961383, "loss": 4.7282, "step": 163150 }, { "epoch": 0.9840691742743093, "grad_norm": 0.9312903881072998, "learning_rate": 0.0002860525470833753, "loss": 5.4656, "step": 163200 }, { "epoch": 0.9843706660556433, "grad_norm": 1.529245138168335, "learning_rate": 0.0002860439770032838, "loss": 4.898, "step": 163250 }, { "epoch": 0.9846721578369774, "grad_norm": 1.3076131343841553, "learning_rate": 0.0002860354044194971, "loss": 5.0206, "step": 163300 }, { "epoch": 0.9849736496183114, "grad_norm": 1.1889207363128662, "learning_rate": 0.0002860268293321728, "loss": 5.0576, "step": 163350 }, { "epoch": 0.9852751413996454, "grad_norm": 1.3577243089675903, "learning_rate": 0.00028601825174146895, "loss": 5.4775, "step": 163400 }, { "epoch": 0.9855766331809794, "grad_norm": 1.015804648399353, "learning_rate": 0.00028600967164754324, "loss": 5.1459, "step": 163450 }, { "epoch": 0.9858781249623135, "grad_norm": 1.5423462390899658, "learning_rate": 0.0002860010890505536, "loss": 5.0837, "step": 163500 }, { "epoch": 0.9861796167436475, "grad_norm": 1.1370787620544434, "learning_rate": 0.00028599250395065795, "loss": 5.4026, "step": 163550 }, { "epoch": 0.9864811085249816, "grad_norm": 1.544995903968811, "learning_rate": 0.0002859839163480144, "loss": 5.6125, "step": 163600 }, { "epoch": 0.9867826003063157, "grad_norm": 1.4509356021881104, "learning_rate": 0.00028597532624278094, "loss": 5.1388, "step": 163650 }, { "epoch": 0.9870840920876497, "grad_norm": 1.7052663564682007, "learning_rate": 0.0002859667336351156, "loss": 5.4517, "step": 163700 }, { "epoch": 0.9873855838689838, "grad_norm": 1.0534886121749878, "learning_rate": 0.00028595813852517663, "loss": 5.3704, "step": 163750 }, { "epoch": 0.9876870756503178, "grad_norm": 1.0704901218414307, "learning_rate": 0.0002859495409131221, "loss": 5.0065, "step": 163800 }, { "epoch": 0.9879885674316519, "grad_norm": 1.7805166244506836, "learning_rate": 0.00028594094079911026, "loss": 5.147, "step": 163850 }, { "epoch": 0.9882900592129858, "grad_norm": 1.605783224105835, "learning_rate": 0.0002859323381832994, "loss": 5.417, "step": 163900 }, { "epoch": 0.98859155099432, "grad_norm": 1.4657938480377197, "learning_rate": 0.0002859237330658479, "loss": 4.8965, "step": 163950 }, { "epoch": 0.9888930427756539, "grad_norm": 0.677395224571228, "learning_rate": 0.000285915125446914, "loss": 5.1664, "step": 164000 }, { "epoch": 0.989194534556988, "grad_norm": 1.3033111095428467, "learning_rate": 0.00028590651532665624, "loss": 5.2257, "step": 164050 }, { "epoch": 0.989496026338322, "grad_norm": 1.5703026056289673, "learning_rate": 0.00028589790270523296, "loss": 5.45, "step": 164100 }, { "epoch": 0.9897975181196561, "grad_norm": 1.0956568717956543, "learning_rate": 0.0002858892875828028, "loss": 5.1368, "step": 164150 }, { "epoch": 0.9900990099009901, "grad_norm": 1.6455731391906738, "learning_rate": 0.00028588066995952413, "loss": 5.2239, "step": 164200 }, { "epoch": 0.9904005016823242, "grad_norm": 1.1507912874221802, "learning_rate": 0.00028587204983555566, "loss": 4.6151, "step": 164250 }, { "epoch": 0.9907019934636582, "grad_norm": 1.6962883472442627, "learning_rate": 0.000285863427211056, "loss": 5.6378, "step": 164300 }, { "epoch": 0.9910034852449923, "grad_norm": 1.085607647895813, "learning_rate": 0.0002858548020861839, "loss": 4.8432, "step": 164350 }, { "epoch": 0.9913049770263263, "grad_norm": 1.123114824295044, "learning_rate": 0.00028584617446109805, "loss": 5.0198, "step": 164400 }, { "epoch": 0.9916064688076603, "grad_norm": 1.2561765909194946, "learning_rate": 0.0002858375443359572, "loss": 4.8618, "step": 164450 }, { "epoch": 0.9919079605889943, "grad_norm": 1.085960030555725, "learning_rate": 0.0002858289117109202, "loss": 5.2854, "step": 164500 }, { "epoch": 0.9922094523703284, "grad_norm": 0.9056730270385742, "learning_rate": 0.0002858202765861459, "loss": 4.8921, "step": 164550 }, { "epoch": 0.9925109441516624, "grad_norm": 1.4401495456695557, "learning_rate": 0.0002858116389617932, "loss": 5.0949, "step": 164600 }, { "epoch": 0.9928124359329965, "grad_norm": 0.9236772656440735, "learning_rate": 0.0002858029988380212, "loss": 4.8765, "step": 164650 }, { "epoch": 0.9931139277143305, "grad_norm": 1.1768697500228882, "learning_rate": 0.00028579435621498873, "loss": 5.2365, "step": 164700 }, { "epoch": 0.9934154194956646, "grad_norm": 2.086521863937378, "learning_rate": 0.000285785711092855, "loss": 5.3573, "step": 164750 }, { "epoch": 0.9937169112769986, "grad_norm": 1.7992589473724365, "learning_rate": 0.00028577706347177896, "loss": 5.4427, "step": 164800 }, { "epoch": 0.9940184030583327, "grad_norm": 1.6407090425491333, "learning_rate": 0.0002857684133519198, "loss": 4.9703, "step": 164850 }, { "epoch": 0.9943198948396667, "grad_norm": 1.9836337566375732, "learning_rate": 0.00028575976073343687, "loss": 4.7979, "step": 164900 }, { "epoch": 0.9946213866210007, "grad_norm": 0.9127053618431091, "learning_rate": 0.0002857511056164892, "loss": 5.5253, "step": 164950 }, { "epoch": 0.9949228784023347, "grad_norm": 1.4649827480316162, "learning_rate": 0.0002857424480012361, "loss": 5.2839, "step": 165000 }, { "epoch": 0.9949228784023347, "eval_loss": 5.671165466308594, "eval_runtime": 39.0465, "eval_samples_per_second": 13.113, "eval_steps_per_second": 6.556, "eval_tts_loss": 7.2064489884766525, "step": 165000 }, { "epoch": 0.9952243701836688, "grad_norm": 1.688293695449829, "learning_rate": 0.0002857337878878371, "loss": 5.0538, "step": 165050 }, { "epoch": 0.9955258619650028, "grad_norm": 1.7680010795593262, "learning_rate": 0.00028572512527645134, "loss": 5.5823, "step": 165100 }, { "epoch": 0.9958273537463369, "grad_norm": 0.39983540773391724, "learning_rate": 0.0002857164601672384, "loss": 5.3078, "step": 165150 }, { "epoch": 0.9961288455276709, "grad_norm": 1.8838379383087158, "learning_rate": 0.00028570779256035763, "loss": 5.1314, "step": 165200 }, { "epoch": 0.996430337309005, "grad_norm": 1.166282057762146, "learning_rate": 0.0002856991224559686, "loss": 5.0194, "step": 165250 }, { "epoch": 0.996731829090339, "grad_norm": 0.6866438388824463, "learning_rate": 0.00028569044985423086, "loss": 5.2885, "step": 165300 }, { "epoch": 0.9970333208716731, "grad_norm": 0.9227458238601685, "learning_rate": 0.00028568177475530405, "loss": 5.1106, "step": 165350 }, { "epoch": 0.997334812653007, "grad_norm": 1.1427098512649536, "learning_rate": 0.00028567309715934786, "loss": 5.3669, "step": 165400 }, { "epoch": 0.9976363044343411, "grad_norm": 1.6873548030853271, "learning_rate": 0.0002856644170665219, "loss": 5.3463, "step": 165450 }, { "epoch": 0.9979377962156751, "grad_norm": 1.6654683351516724, "learning_rate": 0.00028565573447698596, "loss": 5.3777, "step": 165500 }, { "epoch": 0.9982392879970092, "grad_norm": 1.7132155895233154, "learning_rate": 0.0002856470493908998, "loss": 5.4729, "step": 165550 }, { "epoch": 0.9985407797783432, "grad_norm": 1.163101077079773, "learning_rate": 0.0002856383618084233, "loss": 4.9853, "step": 165600 }, { "epoch": 0.9988422715596773, "grad_norm": 1.4985449314117432, "learning_rate": 0.00028562967172971626, "loss": 5.659, "step": 165650 }, { "epoch": 0.9991437633410113, "grad_norm": 0.8539741635322571, "learning_rate": 0.0002856209791549387, "loss": 4.6881, "step": 165700 }, { "epoch": 0.9994452551223454, "grad_norm": 1.1476261615753174, "learning_rate": 0.00028561228408425056, "loss": 5.0823, "step": 165750 }, { "epoch": 0.9997467469036794, "grad_norm": 2.479600667953491, "learning_rate": 0.00028560358651781185, "loss": 5.5193, "step": 165800 }, { "epoch": 1.0000482386850134, "grad_norm": 1.1407890319824219, "learning_rate": 0.00028559488645578266, "loss": 5.4442, "step": 165850 }, { "epoch": 1.0003497304663476, "grad_norm": 1.9224570989608765, "learning_rate": 0.0002855861838983231, "loss": 5.0968, "step": 165900 }, { "epoch": 1.0006512222476815, "grad_norm": 1.129863977432251, "learning_rate": 0.0002855774788455933, "loss": 5.2747, "step": 165950 }, { "epoch": 1.0009527140290155, "grad_norm": 1.3493777513504028, "learning_rate": 0.00028556877129775346, "loss": 5.1052, "step": 166000 }, { "epoch": 1.0012542058103495, "grad_norm": 1.1590770483016968, "learning_rate": 0.00028556006125496385, "loss": 5.1446, "step": 166050 }, { "epoch": 1.0015556975916837, "grad_norm": 1.0508099794387817, "learning_rate": 0.0002855513487173848, "loss": 5.4416, "step": 166100 }, { "epoch": 1.0018571893730177, "grad_norm": 0.9718881845474243, "learning_rate": 0.00028554263368517666, "loss": 5.2133, "step": 166150 }, { "epoch": 1.0021586811543517, "grad_norm": 1.1663620471954346, "learning_rate": 0.00028553391615849974, "loss": 5.2235, "step": 166200 }, { "epoch": 1.0024601729356857, "grad_norm": 1.304509162902832, "learning_rate": 0.0002855251961375145, "loss": 5.5994, "step": 166250 }, { "epoch": 1.0027616647170199, "grad_norm": 1.1574549674987793, "learning_rate": 0.00028551647362238137, "loss": 4.8332, "step": 166300 }, { "epoch": 1.0030631564983539, "grad_norm": 1.8096781969070435, "learning_rate": 0.000285507748613261, "loss": 5.0953, "step": 166350 }, { "epoch": 1.0033646482796879, "grad_norm": 0.9647696614265442, "learning_rate": 0.00028549902111031395, "loss": 5.4401, "step": 166400 }, { "epoch": 1.0036661400610218, "grad_norm": 1.235223650932312, "learning_rate": 0.0002854902911137007, "loss": 5.0371, "step": 166450 }, { "epoch": 1.003967631842356, "grad_norm": 1.2554482221603394, "learning_rate": 0.000285481558623582, "loss": 5.8563, "step": 166500 }, { "epoch": 1.00426912362369, "grad_norm": 1.3481732606887817, "learning_rate": 0.0002854728236401186, "loss": 5.0433, "step": 166550 }, { "epoch": 1.004570615405024, "grad_norm": 1.2127629518508911, "learning_rate": 0.00028546408616347114, "loss": 5.1976, "step": 166600 }, { "epoch": 1.004872107186358, "grad_norm": 1.9563778638839722, "learning_rate": 0.00028545534619380057, "loss": 5.5042, "step": 166650 }, { "epoch": 1.0051735989676922, "grad_norm": 1.2532026767730713, "learning_rate": 0.00028544660373126763, "loss": 5.3454, "step": 166700 }, { "epoch": 1.0054750907490262, "grad_norm": 1.4976719617843628, "learning_rate": 0.00028543785877603323, "loss": 5.014, "step": 166750 }, { "epoch": 1.0057765825303602, "grad_norm": 1.5328762531280518, "learning_rate": 0.0002854291113282584, "loss": 5.3842, "step": 166800 }, { "epoch": 1.0060780743116942, "grad_norm": 2.116807460784912, "learning_rate": 0.00028542036138810395, "loss": 4.9812, "step": 166850 }, { "epoch": 1.0063795660930284, "grad_norm": 1.1272799968719482, "learning_rate": 0.000285411608955731, "loss": 5.4035, "step": 166900 }, { "epoch": 1.0066810578743624, "grad_norm": 2.192249059677124, "learning_rate": 0.0002854028540313007, "loss": 5.089, "step": 166950 }, { "epoch": 1.0069825496556963, "grad_norm": 0.97234708070755, "learning_rate": 0.000285394096614974, "loss": 5.4477, "step": 167000 }, { "epoch": 1.0072840414370303, "grad_norm": 1.4995300769805908, "learning_rate": 0.0002853853367069122, "loss": 5.1508, "step": 167050 }, { "epoch": 1.0075855332183645, "grad_norm": 1.114480972290039, "learning_rate": 0.00028537657430727654, "loss": 5.3229, "step": 167100 }, { "epoch": 1.0078870249996985, "grad_norm": 1.9128762483596802, "learning_rate": 0.00028536780941622824, "loss": 5.052, "step": 167150 }, { "epoch": 1.0081885167810325, "grad_norm": 1.1081236600875854, "learning_rate": 0.00028535904203392853, "loss": 4.9429, "step": 167200 }, { "epoch": 1.0084900085623665, "grad_norm": 1.1521334648132324, "learning_rate": 0.0002853502721605388, "loss": 5.5081, "step": 167250 }, { "epoch": 1.0087915003437007, "grad_norm": 1.0340059995651245, "learning_rate": 0.0002853414997962205, "loss": 5.3554, "step": 167300 }, { "epoch": 1.0090929921250347, "grad_norm": 1.7663995027542114, "learning_rate": 0.0002853327249411351, "loss": 5.2511, "step": 167350 }, { "epoch": 1.0093944839063687, "grad_norm": 1.2023100852966309, "learning_rate": 0.00028532394759544393, "loss": 5.1848, "step": 167400 }, { "epoch": 1.0096959756877029, "grad_norm": 0.348593145608902, "learning_rate": 0.00028531516775930863, "loss": 4.9962, "step": 167450 }, { "epoch": 1.0099974674690368, "grad_norm": 1.8057384490966797, "learning_rate": 0.0002853063854328908, "loss": 5.3594, "step": 167500 }, { "epoch": 1.0102989592503708, "grad_norm": 1.063029408454895, "learning_rate": 0.000285297600616352, "loss": 5.1259, "step": 167550 }, { "epoch": 1.0106004510317048, "grad_norm": 2.2014265060424805, "learning_rate": 0.00028528881330985406, "loss": 4.9878, "step": 167600 }, { "epoch": 1.010901942813039, "grad_norm": 1.8758608102798462, "learning_rate": 0.00028528002351355846, "loss": 5.2219, "step": 167650 }, { "epoch": 1.011203434594373, "grad_norm": 1.5074063539505005, "learning_rate": 0.00028527123122762713, "loss": 5.1668, "step": 167700 }, { "epoch": 1.011504926375707, "grad_norm": 1.8625437021255493, "learning_rate": 0.0002852624364522218, "loss": 5.3894, "step": 167750 }, { "epoch": 1.011806418157041, "grad_norm": 0.4323553442955017, "learning_rate": 0.00028525363918750437, "loss": 5.2676, "step": 167800 }, { "epoch": 1.0121079099383752, "grad_norm": 2.958794116973877, "learning_rate": 0.0002852448394336368, "loss": 5.3112, "step": 167850 }, { "epoch": 1.0124094017197092, "grad_norm": 1.1304806470870972, "learning_rate": 0.00028523603719078086, "loss": 5.2509, "step": 167900 }, { "epoch": 1.0127108935010432, "grad_norm": 1.1215916872024536, "learning_rate": 0.00028522723245909867, "loss": 5.265, "step": 167950 }, { "epoch": 1.0130123852823771, "grad_norm": 1.0617082118988037, "learning_rate": 0.00028521842523875226, "loss": 4.4985, "step": 168000 }, { "epoch": 1.0130123852823771, "eval_loss": 5.66669225692749, "eval_runtime": 39.1447, "eval_samples_per_second": 13.08, "eval_steps_per_second": 6.54, "eval_tts_loss": 7.231743600724044, "step": 168000 }, { "epoch": 1.0133138770637113, "grad_norm": 1.2680619955062866, "learning_rate": 0.0002852096155299037, "loss": 4.9536, "step": 168050 }, { "epoch": 1.0136153688450453, "grad_norm": 1.073631763458252, "learning_rate": 0.0002852008033327152, "loss": 4.8497, "step": 168100 }, { "epoch": 1.0139168606263793, "grad_norm": 1.4889531135559082, "learning_rate": 0.00028519198864734873, "loss": 5.068, "step": 168150 }, { "epoch": 1.0142183524077133, "grad_norm": 2.1167943477630615, "learning_rate": 0.0002851831714739667, "loss": 5.1668, "step": 168200 }, { "epoch": 1.0145198441890475, "grad_norm": 1.1226085424423218, "learning_rate": 0.0002851743518127313, "loss": 5.8003, "step": 168250 }, { "epoch": 1.0148213359703815, "grad_norm": 1.019120693206787, "learning_rate": 0.0002851655296638049, "loss": 5.1119, "step": 168300 }, { "epoch": 1.0151228277517155, "grad_norm": 1.9140961170196533, "learning_rate": 0.00028515670502734977, "loss": 5.6614, "step": 168350 }, { "epoch": 1.0154243195330495, "grad_norm": 1.1274645328521729, "learning_rate": 0.00028514787790352833, "loss": 5.3369, "step": 168400 }, { "epoch": 1.0157258113143837, "grad_norm": 0.7055413722991943, "learning_rate": 0.00028513904829250314, "loss": 5.4827, "step": 168450 }, { "epoch": 1.0160273030957176, "grad_norm": 1.1682708263397217, "learning_rate": 0.00028513021619443657, "loss": 5.0963, "step": 168500 }, { "epoch": 1.0163287948770516, "grad_norm": 1.9351691007614136, "learning_rate": 0.00028512138160949124, "loss": 5.0266, "step": 168550 }, { "epoch": 1.0166302866583856, "grad_norm": 1.3371727466583252, "learning_rate": 0.0002851125445378297, "loss": 5.6477, "step": 168600 }, { "epoch": 1.0169317784397198, "grad_norm": 1.2568777799606323, "learning_rate": 0.0002851037049796146, "loss": 5.319, "step": 168650 }, { "epoch": 1.0172332702210538, "grad_norm": 1.437023639678955, "learning_rate": 0.0002850948629350086, "loss": 5.5627, "step": 168700 }, { "epoch": 1.0175347620023878, "grad_norm": 1.4856027364730835, "learning_rate": 0.00028508601840417444, "loss": 5.4797, "step": 168750 }, { "epoch": 1.0178362537837218, "grad_norm": 1.2064710855484009, "learning_rate": 0.00028507717138727493, "loss": 4.9761, "step": 168800 }, { "epoch": 1.018137745565056, "grad_norm": 1.7981373071670532, "learning_rate": 0.0002850683218844728, "loss": 5.0864, "step": 168850 }, { "epoch": 1.01843923734639, "grad_norm": 1.6170579195022583, "learning_rate": 0.00028505946989593097, "loss": 5.6254, "step": 168900 }, { "epoch": 1.018740729127724, "grad_norm": 1.508662223815918, "learning_rate": 0.0002850506154218123, "loss": 5.2899, "step": 168950 }, { "epoch": 1.019042220909058, "grad_norm": 1.1708358526229858, "learning_rate": 0.0002850417584622799, "loss": 5.3649, "step": 169000 }, { "epoch": 1.0193437126903921, "grad_norm": 2.2046408653259277, "learning_rate": 0.0002850328990174965, "loss": 4.8949, "step": 169050 }, { "epoch": 1.0196452044717261, "grad_norm": 1.2460949420928955, "learning_rate": 0.0002850240370876254, "loss": 5.4355, "step": 169100 }, { "epoch": 1.0199466962530601, "grad_norm": 1.7187395095825195, "learning_rate": 0.00028501517267282955, "loss": 5.1331, "step": 169150 }, { "epoch": 1.020248188034394, "grad_norm": 2.0045461654663086, "learning_rate": 0.00028500630577327213, "loss": 4.8713, "step": 169200 }, { "epoch": 1.0205496798157283, "grad_norm": 1.0169366598129272, "learning_rate": 0.0002849974363891163, "loss": 5.2787, "step": 169250 }, { "epoch": 1.0208511715970623, "grad_norm": 1.109613060951233, "learning_rate": 0.00028498856452052534, "loss": 4.8, "step": 169300 }, { "epoch": 1.0211526633783963, "grad_norm": 1.9689172506332397, "learning_rate": 0.0002849796901676624, "loss": 5.0529, "step": 169350 }, { "epoch": 1.0214541551597303, "grad_norm": 1.8127249479293823, "learning_rate": 0.000284970813330691, "loss": 5.2346, "step": 169400 }, { "epoch": 1.0217556469410645, "grad_norm": 1.3677676916122437, "learning_rate": 0.0002849619340097744, "loss": 5.574, "step": 169450 }, { "epoch": 1.0220571387223985, "grad_norm": 1.3878545761108398, "learning_rate": 0.00028495305220507594, "loss": 5.2701, "step": 169500 }, { "epoch": 1.0223586305037324, "grad_norm": 1.3384488821029663, "learning_rate": 0.00028494416791675914, "loss": 5.2492, "step": 169550 }, { "epoch": 1.0226601222850664, "grad_norm": 1.1820489168167114, "learning_rate": 0.0002849352811449875, "loss": 4.9099, "step": 169600 }, { "epoch": 1.0229616140664006, "grad_norm": 1.277307152748108, "learning_rate": 0.0002849263918899246, "loss": 5.5839, "step": 169650 }, { "epoch": 1.0232631058477346, "grad_norm": 1.4964858293533325, "learning_rate": 0.000284917500151734, "loss": 5.3314, "step": 169700 }, { "epoch": 1.0235645976290686, "grad_norm": 1.8582234382629395, "learning_rate": 0.0002849086059305793, "loss": 4.9557, "step": 169750 }, { "epoch": 1.0238660894104026, "grad_norm": 1.0983136892318726, "learning_rate": 0.00028489970922662434, "loss": 5.4508, "step": 169800 }, { "epoch": 1.0241675811917368, "grad_norm": 1.2851970195770264, "learning_rate": 0.0002848908100400327, "loss": 5.2389, "step": 169850 }, { "epoch": 1.0244690729730708, "grad_norm": 1.4716085195541382, "learning_rate": 0.00028488190837096814, "loss": 4.9937, "step": 169900 }, { "epoch": 1.0247705647544048, "grad_norm": 1.6202969551086426, "learning_rate": 0.0002848730042195946, "loss": 5.0521, "step": 169950 }, { "epoch": 1.0250720565357387, "grad_norm": 1.2281121015548706, "learning_rate": 0.0002848640975860759, "loss": 4.9714, "step": 170000 }, { "epoch": 1.025373548317073, "grad_norm": 1.2633475065231323, "learning_rate": 0.00028485518847057593, "loss": 5.2178, "step": 170050 }, { "epoch": 1.025675040098407, "grad_norm": 1.4828799962997437, "learning_rate": 0.0002848462768732587, "loss": 5.0088, "step": 170100 }, { "epoch": 1.025976531879741, "grad_norm": 1.439505696296692, "learning_rate": 0.0002848373627942881, "loss": 5.2516, "step": 170150 }, { "epoch": 1.026278023661075, "grad_norm": 1.7184579372406006, "learning_rate": 0.0002848284462338283, "loss": 5.2581, "step": 170200 }, { "epoch": 1.026579515442409, "grad_norm": 1.2383882999420166, "learning_rate": 0.0002848195271920434, "loss": 4.8889, "step": 170250 }, { "epoch": 1.026881007223743, "grad_norm": 0.8643097281455994, "learning_rate": 0.0002848106056690975, "loss": 5.1763, "step": 170300 }, { "epoch": 1.027182499005077, "grad_norm": 0.46179673075675964, "learning_rate": 0.0002848016816651547, "loss": 5.1123, "step": 170350 }, { "epoch": 1.027483990786411, "grad_norm": 1.8005422353744507, "learning_rate": 0.00028479275518037935, "loss": 4.4383, "step": 170400 }, { "epoch": 1.0277854825677453, "grad_norm": 1.6312962770462036, "learning_rate": 0.00028478382621493576, "loss": 5.0169, "step": 170450 }, { "epoch": 1.0280869743490793, "grad_norm": 1.0414592027664185, "learning_rate": 0.00028477489476898813, "loss": 5.2225, "step": 170500 }, { "epoch": 1.0283884661304132, "grad_norm": 2.1724188327789307, "learning_rate": 0.00028476596084270095, "loss": 5.2785, "step": 170550 }, { "epoch": 1.0286899579117472, "grad_norm": 1.8512825965881348, "learning_rate": 0.0002847570244362385, "loss": 4.8937, "step": 170600 }, { "epoch": 1.0289914496930814, "grad_norm": 1.14168119430542, "learning_rate": 0.00028474808554976544, "loss": 4.8161, "step": 170650 }, { "epoch": 1.0292929414744154, "grad_norm": 0.7393837571144104, "learning_rate": 0.0002847391441834461, "loss": 4.9672, "step": 170700 }, { "epoch": 1.0295944332557494, "grad_norm": 1.2832591533660889, "learning_rate": 0.0002847302003374451, "loss": 5.0975, "step": 170750 }, { "epoch": 1.0298959250370834, "grad_norm": 1.2016764879226685, "learning_rate": 0.00028472125401192704, "loss": 5.2644, "step": 170800 }, { "epoch": 1.0301974168184176, "grad_norm": 1.5435986518859863, "learning_rate": 0.0002847123052070565, "loss": 5.4327, "step": 170850 }, { "epoch": 1.0304989085997516, "grad_norm": 1.2784219980239868, "learning_rate": 0.00028470335392299825, "loss": 5.3873, "step": 170900 }, { "epoch": 1.0308004003810856, "grad_norm": 1.3019580841064453, "learning_rate": 0.0002846944001599171, "loss": 5.0086, "step": 170950 }, { "epoch": 1.0311018921624195, "grad_norm": 1.0518265962600708, "learning_rate": 0.0002846854439179776, "loss": 4.7404, "step": 171000 }, { "epoch": 1.0311018921624195, "eval_loss": 5.669318199157715, "eval_runtime": 39.065, "eval_samples_per_second": 13.106, "eval_steps_per_second": 6.553, "eval_tts_loss": 7.1539516989062415, "step": 171000 }, { "epoch": 1.0314033839437537, "grad_norm": 1.1470900774002075, "learning_rate": 0.0002846764851973448, "loss": 5.3181, "step": 171050 }, { "epoch": 1.0317048757250877, "grad_norm": 1.8200784921646118, "learning_rate": 0.00028466752399818344, "loss": 5.223, "step": 171100 }, { "epoch": 1.0320063675064217, "grad_norm": 1.1360580921173096, "learning_rate": 0.0002846585603206585, "loss": 5.4406, "step": 171150 }, { "epoch": 1.032307859287756, "grad_norm": 2.0265262126922607, "learning_rate": 0.0002846495941649349, "loss": 4.9981, "step": 171200 }, { "epoch": 1.03260935106909, "grad_norm": 1.127638816833496, "learning_rate": 0.0002846406255311777, "loss": 5.2586, "step": 171250 }, { "epoch": 1.032910842850424, "grad_norm": 1.1191216707229614, "learning_rate": 0.00028463165441955195, "loss": 5.2816, "step": 171300 }, { "epoch": 1.0332123346317579, "grad_norm": 1.3139292001724243, "learning_rate": 0.00028462268083022266, "loss": 5.0433, "step": 171350 }, { "epoch": 1.033513826413092, "grad_norm": 1.3775134086608887, "learning_rate": 0.0002846137047633551, "loss": 5.2094, "step": 171400 }, { "epoch": 1.033815318194426, "grad_norm": 1.531082034111023, "learning_rate": 0.0002846047262191144, "loss": 5.2554, "step": 171450 }, { "epoch": 1.03411680997576, "grad_norm": 1.2007426023483276, "learning_rate": 0.0002845957451976658, "loss": 5.0641, "step": 171500 }, { "epoch": 1.034418301757094, "grad_norm": 1.5250176191329956, "learning_rate": 0.00028458676169917463, "loss": 5.4093, "step": 171550 }, { "epoch": 1.0347197935384282, "grad_norm": 1.2824820280075073, "learning_rate": 0.0002845777757238061, "loss": 5.4213, "step": 171600 }, { "epoch": 1.0350212853197622, "grad_norm": 1.6099777221679688, "learning_rate": 0.00028456878727172575, "loss": 5.3637, "step": 171650 }, { "epoch": 1.0353227771010962, "grad_norm": 1.2707066535949707, "learning_rate": 0.00028455979634309887, "loss": 5.2782, "step": 171700 }, { "epoch": 1.0356242688824302, "grad_norm": 1.4569491147994995, "learning_rate": 0.0002845508029380909, "loss": 4.8876, "step": 171750 }, { "epoch": 1.0359257606637644, "grad_norm": 1.7485415935516357, "learning_rate": 0.0002845418070568675, "loss": 5.1919, "step": 171800 }, { "epoch": 1.0362272524450984, "grad_norm": 0.7062082290649414, "learning_rate": 0.0002845328086995941, "loss": 5.2671, "step": 171850 }, { "epoch": 1.0365287442264324, "grad_norm": 1.730145812034607, "learning_rate": 0.00028452380786643636, "loss": 5.0594, "step": 171900 }, { "epoch": 1.0368302360077664, "grad_norm": 1.1224908828735352, "learning_rate": 0.00028451480455755993, "loss": 5.1803, "step": 171950 }, { "epoch": 1.0371317277891006, "grad_norm": 0.9822213649749756, "learning_rate": 0.0002845057987731305, "loss": 5.38, "step": 172000 }, { "epoch": 1.0374332195704346, "grad_norm": 1.0550786256790161, "learning_rate": 0.00028449679051331374, "loss": 5.2261, "step": 172050 }, { "epoch": 1.0377347113517685, "grad_norm": 1.0441229343414307, "learning_rate": 0.00028448777977827547, "loss": 4.9866, "step": 172100 }, { "epoch": 1.0380362031331025, "grad_norm": 1.5509852170944214, "learning_rate": 0.0002844787665681816, "loss": 4.8983, "step": 172150 }, { "epoch": 1.0383376949144367, "grad_norm": 1.3055269718170166, "learning_rate": 0.0002844697508831979, "loss": 4.9291, "step": 172200 }, { "epoch": 1.0386391866957707, "grad_norm": 1.3617615699768066, "learning_rate": 0.0002844607327234904, "loss": 5.3106, "step": 172250 }, { "epoch": 1.0389406784771047, "grad_norm": 1.235087275505066, "learning_rate": 0.00028445171208922495, "loss": 5.2773, "step": 172300 }, { "epoch": 1.0392421702584387, "grad_norm": 1.6191834211349487, "learning_rate": 0.00028444268898056757, "loss": 5.2194, "step": 172350 }, { "epoch": 1.0395436620397729, "grad_norm": 0.9100556373596191, "learning_rate": 0.00028443366339768445, "loss": 5.0502, "step": 172400 }, { "epoch": 1.0398451538211069, "grad_norm": 1.0487231016159058, "learning_rate": 0.0002844246353407415, "loss": 4.4444, "step": 172450 }, { "epoch": 1.0401466456024409, "grad_norm": 1.0399647951126099, "learning_rate": 0.00028441560480990506, "loss": 4.921, "step": 172500 }, { "epoch": 1.0404481373837748, "grad_norm": 1.8907113075256348, "learning_rate": 0.0002844065718053412, "loss": 5.2549, "step": 172550 }, { "epoch": 1.040749629165109, "grad_norm": 2.0090150833129883, "learning_rate": 0.0002843975363272162, "loss": 5.3856, "step": 172600 }, { "epoch": 1.041051120946443, "grad_norm": 0.9131385087966919, "learning_rate": 0.00028438849837569635, "loss": 5.5656, "step": 172650 }, { "epoch": 1.041352612727777, "grad_norm": 1.4326099157333374, "learning_rate": 0.0002843794579509479, "loss": 5.2881, "step": 172700 }, { "epoch": 1.041654104509111, "grad_norm": 1.1448014974594116, "learning_rate": 0.0002843704150531374, "loss": 4.894, "step": 172750 }, { "epoch": 1.0419555962904452, "grad_norm": 1.5738749504089355, "learning_rate": 0.0002843613696824311, "loss": 5.581, "step": 172800 }, { "epoch": 1.0422570880717792, "grad_norm": 1.4732722043991089, "learning_rate": 0.0002843523218389956, "loss": 5.3081, "step": 172850 }, { "epoch": 1.0425585798531132, "grad_norm": 1.6300110816955566, "learning_rate": 0.0002843432715229973, "loss": 5.094, "step": 172900 }, { "epoch": 1.0428600716344472, "grad_norm": 0.5239542126655579, "learning_rate": 0.00028433421873460277, "loss": 4.83, "step": 172950 }, { "epoch": 1.0431615634157814, "grad_norm": 1.5809475183486938, "learning_rate": 0.0002843251634739787, "loss": 5.0509, "step": 173000 }, { "epoch": 1.0434630551971154, "grad_norm": 1.71309494972229, "learning_rate": 0.00028431610574129165, "loss": 5.0441, "step": 173050 }, { "epoch": 1.0437645469784493, "grad_norm": 1.0716718435287476, "learning_rate": 0.0002843070455367084, "loss": 5.0643, "step": 173100 }, { "epoch": 1.0440660387597833, "grad_norm": 1.3868937492370605, "learning_rate": 0.00028429798286039564, "loss": 5.3071, "step": 173150 }, { "epoch": 1.0443675305411175, "grad_norm": 1.0517919063568115, "learning_rate": 0.0002842889177125201, "loss": 5.7069, "step": 173200 }, { "epoch": 1.0446690223224515, "grad_norm": 1.0380215644836426, "learning_rate": 0.0002842798500932487, "loss": 5.175, "step": 173250 }, { "epoch": 1.0449705141037855, "grad_norm": 1.2550522089004517, "learning_rate": 0.0002842707800027483, "loss": 5.166, "step": 173300 }, { "epoch": 1.0452720058851195, "grad_norm": 1.471093773841858, "learning_rate": 0.0002842617074411858, "loss": 5.3012, "step": 173350 }, { "epoch": 1.0455734976664537, "grad_norm": 1.4930171966552734, "learning_rate": 0.0002842526324087282, "loss": 5.2199, "step": 173400 }, { "epoch": 1.0458749894477877, "grad_norm": 1.2808502912521362, "learning_rate": 0.00028424355490554246, "loss": 5.0325, "step": 173450 }, { "epoch": 1.0461764812291217, "grad_norm": 1.2832098007202148, "learning_rate": 0.0002842344749317957, "loss": 5.3802, "step": 173500 }, { "epoch": 1.0464779730104556, "grad_norm": 1.3144298791885376, "learning_rate": 0.000284225392487655, "loss": 5.2624, "step": 173550 }, { "epoch": 1.0467794647917898, "grad_norm": 1.3185632228851318, "learning_rate": 0.0002842163075732875, "loss": 4.9368, "step": 173600 }, { "epoch": 1.0470809565731238, "grad_norm": 1.695967674255371, "learning_rate": 0.00028420722018886033, "loss": 4.8168, "step": 173650 }, { "epoch": 1.0473824483544578, "grad_norm": 0.8716504573822021, "learning_rate": 0.0002841981303345408, "loss": 4.5822, "step": 173700 }, { "epoch": 1.0476839401357918, "grad_norm": 1.7192696332931519, "learning_rate": 0.00028418903801049624, "loss": 5.0206, "step": 173750 }, { "epoch": 1.047985431917126, "grad_norm": 2.1863811016082764, "learning_rate": 0.00028417994321689393, "loss": 5.2791, "step": 173800 }, { "epoch": 1.04828692369846, "grad_norm": 1.1784250736236572, "learning_rate": 0.00028417084595390126, "loss": 5.6223, "step": 173850 }, { "epoch": 1.048588415479794, "grad_norm": 0.4903421700000763, "learning_rate": 0.0002841617462216856, "loss": 5.1723, "step": 173900 }, { "epoch": 1.048889907261128, "grad_norm": 1.0614197254180908, "learning_rate": 0.00028415264402041444, "loss": 4.8088, "step": 173950 }, { "epoch": 1.0491913990424622, "grad_norm": 1.1121766567230225, "learning_rate": 0.0002841435393502554, "loss": 5.1248, "step": 174000 }, { "epoch": 1.0491913990424622, "eval_loss": 5.658250331878662, "eval_runtime": 39.119, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.058625548288072, "step": 174000 }, { "epoch": 1.0494928908237962, "grad_norm": 2.15714693069458, "learning_rate": 0.0002841344322113759, "loss": 5.1375, "step": 174050 }, { "epoch": 1.0497943826051301, "grad_norm": 1.1746362447738647, "learning_rate": 0.0002841253226039436, "loss": 5.1198, "step": 174100 }, { "epoch": 1.0500958743864641, "grad_norm": 0.8003747463226318, "learning_rate": 0.0002841162105281262, "loss": 4.5847, "step": 174150 }, { "epoch": 1.0503973661677983, "grad_norm": 1.178815245628357, "learning_rate": 0.00028410709598409125, "loss": 5.4984, "step": 174200 }, { "epoch": 1.0506988579491323, "grad_norm": 1.0063910484313965, "learning_rate": 0.0002840979789720066, "loss": 5.384, "step": 174250 }, { "epoch": 1.0510003497304663, "grad_norm": 2.0856385231018066, "learning_rate": 0.00028408885949204003, "loss": 5.2472, "step": 174300 }, { "epoch": 1.0513018415118003, "grad_norm": 2.050337314605713, "learning_rate": 0.00028407973754435937, "loss": 5.1075, "step": 174350 }, { "epoch": 1.0516033332931345, "grad_norm": 0.7592726349830627, "learning_rate": 0.00028407061312913246, "loss": 5.4604, "step": 174400 }, { "epoch": 1.0519048250744685, "grad_norm": 1.690390944480896, "learning_rate": 0.00028406148624652726, "loss": 5.3126, "step": 174450 }, { "epoch": 1.0522063168558025, "grad_norm": 1.0091112852096558, "learning_rate": 0.00028405235689671175, "loss": 4.953, "step": 174500 }, { "epoch": 1.0525078086371367, "grad_norm": 0.7227573394775391, "learning_rate": 0.00028404322507985383, "loss": 5.1349, "step": 174550 }, { "epoch": 1.0528093004184707, "grad_norm": 1.4193600416183472, "learning_rate": 0.0002840340907961217, "loss": 5.3347, "step": 174600 }, { "epoch": 1.0531107921998046, "grad_norm": 1.6488064527511597, "learning_rate": 0.0002840249540456834, "loss": 4.9329, "step": 174650 }, { "epoch": 1.0534122839811386, "grad_norm": 1.3886584043502808, "learning_rate": 0.0002840158148287071, "loss": 5.2334, "step": 174700 }, { "epoch": 1.0537137757624726, "grad_norm": 1.4975112676620483, "learning_rate": 0.00028400667314536096, "loss": 4.6831, "step": 174750 }, { "epoch": 1.0540152675438068, "grad_norm": 1.5008914470672607, "learning_rate": 0.0002839975289958132, "loss": 5.2517, "step": 174800 }, { "epoch": 1.0543167593251408, "grad_norm": 1.7215638160705566, "learning_rate": 0.0002839883823802322, "loss": 5.1161, "step": 174850 }, { "epoch": 1.0546182511064748, "grad_norm": 1.2201296091079712, "learning_rate": 0.00028397923329878623, "loss": 5.1159, "step": 174900 }, { "epoch": 1.054919742887809, "grad_norm": 1.6511861085891724, "learning_rate": 0.0002839700817516437, "loss": 5.2862, "step": 174950 }, { "epoch": 1.055221234669143, "grad_norm": 1.2603026628494263, "learning_rate": 0.0002839609277389729, "loss": 5.0867, "step": 175000 }, { "epoch": 1.055522726450477, "grad_norm": 0.3695918321609497, "learning_rate": 0.0002839517712609425, "loss": 5.2222, "step": 175050 }, { "epoch": 1.055824218231811, "grad_norm": 1.5265346765518188, "learning_rate": 0.00028394261231772086, "loss": 5.6294, "step": 175100 }, { "epoch": 1.0561257100131451, "grad_norm": 1.0528879165649414, "learning_rate": 0.0002839334509094766, "loss": 4.8114, "step": 175150 }, { "epoch": 1.0564272017944791, "grad_norm": 1.2261357307434082, "learning_rate": 0.0002839242870363783, "loss": 4.8474, "step": 175200 }, { "epoch": 1.0567286935758131, "grad_norm": 1.3827847242355347, "learning_rate": 0.0002839151206985946, "loss": 5.1687, "step": 175250 }, { "epoch": 1.057030185357147, "grad_norm": 1.8418129682540894, "learning_rate": 0.00028390595189629425, "loss": 5.4063, "step": 175300 }, { "epoch": 1.0573316771384813, "grad_norm": 2.0745973587036133, "learning_rate": 0.0002838967806296459, "loss": 5.5597, "step": 175350 }, { "epoch": 1.0576331689198153, "grad_norm": 1.520316243171692, "learning_rate": 0.00028388760689881845, "loss": 5.199, "step": 175400 }, { "epoch": 1.0579346607011493, "grad_norm": 0.8522021770477295, "learning_rate": 0.0002838784307039806, "loss": 5.0778, "step": 175450 }, { "epoch": 1.0582361524824833, "grad_norm": 1.529756784439087, "learning_rate": 0.0002838692520453013, "loss": 5.2218, "step": 175500 }, { "epoch": 1.0585376442638175, "grad_norm": 1.3343534469604492, "learning_rate": 0.00028386007092294946, "loss": 5.3693, "step": 175550 }, { "epoch": 1.0588391360451515, "grad_norm": 1.9295142889022827, "learning_rate": 0.0002838508873370941, "loss": 5.5174, "step": 175600 }, { "epoch": 1.0591406278264854, "grad_norm": 0.8432401418685913, "learning_rate": 0.0002838417012879041, "loss": 5.0838, "step": 175650 }, { "epoch": 1.0594421196078194, "grad_norm": 1.6889712810516357, "learning_rate": 0.0002838325127755486, "loss": 4.8094, "step": 175700 }, { "epoch": 1.0597436113891536, "grad_norm": 1.0126906633377075, "learning_rate": 0.00028382332180019673, "loss": 5.2335, "step": 175750 }, { "epoch": 1.0600451031704876, "grad_norm": 2.2282538414001465, "learning_rate": 0.0002838141283620176, "loss": 4.7493, "step": 175800 }, { "epoch": 1.0603465949518216, "grad_norm": 1.0662941932678223, "learning_rate": 0.0002838049324611804, "loss": 5.4135, "step": 175850 }, { "epoch": 1.0606480867331556, "grad_norm": 1.1989936828613281, "learning_rate": 0.0002837957340978543, "loss": 5.1375, "step": 175900 }, { "epoch": 1.0609495785144898, "grad_norm": 1.1157798767089844, "learning_rate": 0.00028378653327220874, "loss": 5.2444, "step": 175950 }, { "epoch": 1.0612510702958238, "grad_norm": 0.7183984518051147, "learning_rate": 0.00028377732998441287, "loss": 5.175, "step": 176000 }, { "epoch": 1.0615525620771578, "grad_norm": 0.90440833568573, "learning_rate": 0.0002837681242346362, "loss": 5.6029, "step": 176050 }, { "epoch": 1.0618540538584917, "grad_norm": 1.1856491565704346, "learning_rate": 0.0002837589160230481, "loss": 5.1306, "step": 176100 }, { "epoch": 1.062155545639826, "grad_norm": 1.596678614616394, "learning_rate": 0.000283749705349818, "loss": 4.9799, "step": 176150 }, { "epoch": 1.06245703742116, "grad_norm": 1.5770305395126343, "learning_rate": 0.00028374049221511555, "loss": 5.1393, "step": 176200 }, { "epoch": 1.062758529202494, "grad_norm": 1.7334778308868408, "learning_rate": 0.0002837312766191101, "loss": 4.9379, "step": 176250 }, { "epoch": 1.063060020983828, "grad_norm": 1.293083667755127, "learning_rate": 0.0002837220585619714, "loss": 4.6022, "step": 176300 }, { "epoch": 1.063361512765162, "grad_norm": 1.7198519706726074, "learning_rate": 0.00028371283804386904, "loss": 5.2054, "step": 176350 }, { "epoch": 1.063663004546496, "grad_norm": 1.6985646486282349, "learning_rate": 0.0002837036150649727, "loss": 4.9733, "step": 176400 }, { "epoch": 1.06396449632783, "grad_norm": 2.5599615573883057, "learning_rate": 0.00028369438962545204, "loss": 4.6933, "step": 176450 }, { "epoch": 1.064265988109164, "grad_norm": 1.720046043395996, "learning_rate": 0.000283685161725477, "loss": 5.0011, "step": 176500 }, { "epoch": 1.0645674798904983, "grad_norm": 1.9850646257400513, "learning_rate": 0.0002836759313652174, "loss": 4.909, "step": 176550 }, { "epoch": 1.0648689716718323, "grad_norm": 1.6044411659240723, "learning_rate": 0.000283666698544843, "loss": 5.4292, "step": 176600 }, { "epoch": 1.0651704634531662, "grad_norm": 1.3007478713989258, "learning_rate": 0.0002836574632645238, "loss": 5.3209, "step": 176650 }, { "epoch": 1.0654719552345002, "grad_norm": 1.3704993724822998, "learning_rate": 0.00028364822552442965, "loss": 5.5565, "step": 176700 }, { "epoch": 1.0657734470158344, "grad_norm": 2.2223291397094727, "learning_rate": 0.0002836389853247307, "loss": 5.5837, "step": 176750 }, { "epoch": 1.0660749387971684, "grad_norm": 1.5850380659103394, "learning_rate": 0.0002836297426655969, "loss": 5.2812, "step": 176800 }, { "epoch": 1.0663764305785024, "grad_norm": 1.5209633111953735, "learning_rate": 0.0002836204975471984, "loss": 5.1374, "step": 176850 }, { "epoch": 1.0666779223598364, "grad_norm": 1.1817504167556763, "learning_rate": 0.0002836112499697054, "loss": 5.1032, "step": 176900 }, { "epoch": 1.0669794141411706, "grad_norm": 1.2910133600234985, "learning_rate": 0.0002836019999332879, "loss": 5.0882, "step": 176950 }, { "epoch": 1.0672809059225046, "grad_norm": 1.1849116086959839, "learning_rate": 0.00028359274743811633, "loss": 5.0531, "step": 177000 }, { "epoch": 1.0672809059225046, "eval_loss": 5.65696907043457, "eval_runtime": 39.1968, "eval_samples_per_second": 13.062, "eval_steps_per_second": 6.531, "eval_tts_loss": 7.105802893509736, "step": 177000 }, { "epoch": 1.0675823977038386, "grad_norm": 1.4842357635498047, "learning_rate": 0.0002835834924843608, "loss": 5.3423, "step": 177050 }, { "epoch": 1.0678838894851725, "grad_norm": 1.4392093420028687, "learning_rate": 0.00028357423507219183, "loss": 5.2686, "step": 177100 }, { "epoch": 1.0681853812665068, "grad_norm": 1.072418451309204, "learning_rate": 0.00028356497520177966, "loss": 5.2378, "step": 177150 }, { "epoch": 1.0684868730478407, "grad_norm": 1.2704533338546753, "learning_rate": 0.0002835557128732947, "loss": 4.9587, "step": 177200 }, { "epoch": 1.0687883648291747, "grad_norm": 1.143209457397461, "learning_rate": 0.00028354644808690745, "loss": 5.2113, "step": 177250 }, { "epoch": 1.0690898566105087, "grad_norm": 1.5894304513931274, "learning_rate": 0.00028353718084278844, "loss": 5.1298, "step": 177300 }, { "epoch": 1.069391348391843, "grad_norm": 1.1302824020385742, "learning_rate": 0.00028352791114110814, "loss": 5.0244, "step": 177350 }, { "epoch": 1.069692840173177, "grad_norm": 1.2257181406021118, "learning_rate": 0.00028351863898203717, "loss": 5.3577, "step": 177400 }, { "epoch": 1.0699943319545109, "grad_norm": 1.8124737739562988, "learning_rate": 0.0002835093643657462, "loss": 5.39, "step": 177450 }, { "epoch": 1.0702958237358449, "grad_norm": 1.4130516052246094, "learning_rate": 0.00028350008729240595, "loss": 4.8982, "step": 177500 }, { "epoch": 1.070597315517179, "grad_norm": 1.7054184675216675, "learning_rate": 0.00028349080776218707, "loss": 5.0729, "step": 177550 }, { "epoch": 1.070898807298513, "grad_norm": 1.9482219219207764, "learning_rate": 0.0002834815257752604, "loss": 5.4038, "step": 177600 }, { "epoch": 1.071200299079847, "grad_norm": 1.5640803575515747, "learning_rate": 0.0002834722413317967, "loss": 5.5116, "step": 177650 }, { "epoch": 1.071501790861181, "grad_norm": 1.1688361167907715, "learning_rate": 0.00028346295443196695, "loss": 5.3103, "step": 177700 }, { "epoch": 1.0718032826425152, "grad_norm": 1.6103187799453735, "learning_rate": 0.00028345366507594194, "loss": 5.2152, "step": 177750 }, { "epoch": 1.0721047744238492, "grad_norm": 1.1779364347457886, "learning_rate": 0.0002834443732638926, "loss": 5.5385, "step": 177800 }, { "epoch": 1.0724062662051832, "grad_norm": 1.1251434087753296, "learning_rate": 0.0002834350789959901, "loss": 4.6888, "step": 177850 }, { "epoch": 1.0727077579865172, "grad_norm": 2.4804704189300537, "learning_rate": 0.0002834257822724054, "loss": 5.3494, "step": 177900 }, { "epoch": 1.0730092497678514, "grad_norm": 1.3000129461288452, "learning_rate": 0.00028341648309330954, "loss": 5.1688, "step": 177950 }, { "epoch": 1.0733107415491854, "grad_norm": 1.4406238794326782, "learning_rate": 0.0002834071814588737, "loss": 4.7991, "step": 178000 }, { "epoch": 1.0736122333305194, "grad_norm": 1.551135778427124, "learning_rate": 0.0002833978773692691, "loss": 5.3382, "step": 178050 }, { "epoch": 1.0739137251118533, "grad_norm": 1.1353768110275269, "learning_rate": 0.0002833885708246669, "loss": 4.9856, "step": 178100 }, { "epoch": 1.0742152168931876, "grad_norm": 1.5627453327178955, "learning_rate": 0.0002833792618252384, "loss": 5.4559, "step": 178150 }, { "epoch": 1.0745167086745215, "grad_norm": 1.7764450311660767, "learning_rate": 0.000283369950371155, "loss": 5.2354, "step": 178200 }, { "epoch": 1.0748182004558555, "grad_norm": 1.3625026941299438, "learning_rate": 0.00028336063646258795, "loss": 5.3377, "step": 178250 }, { "epoch": 1.0751196922371897, "grad_norm": 1.2358375787734985, "learning_rate": 0.00028335132009970867, "loss": 5.3352, "step": 178300 }, { "epoch": 1.0754211840185237, "grad_norm": 1.3051469326019287, "learning_rate": 0.0002833420012826887, "loss": 5.4919, "step": 178350 }, { "epoch": 1.0757226757998577, "grad_norm": 1.583042860031128, "learning_rate": 0.00028333268001169953, "loss": 5.4653, "step": 178400 }, { "epoch": 1.0760241675811917, "grad_norm": 1.5050984621047974, "learning_rate": 0.00028332335628691255, "loss": 5.2295, "step": 178450 }, { "epoch": 1.0763256593625257, "grad_norm": 1.016768217086792, "learning_rate": 0.0002833140301084995, "loss": 5.3801, "step": 178500 }, { "epoch": 1.0766271511438599, "grad_norm": 1.1859915256500244, "learning_rate": 0.000283304701476632, "loss": 5.1803, "step": 178550 }, { "epoch": 1.0769286429251939, "grad_norm": 1.191945195198059, "learning_rate": 0.0002832953703914817, "loss": 4.9809, "step": 178600 }, { "epoch": 1.0772301347065278, "grad_norm": 0.8866909742355347, "learning_rate": 0.0002832860368532203, "loss": 5.1459, "step": 178650 }, { "epoch": 1.077531626487862, "grad_norm": 1.618242621421814, "learning_rate": 0.00028327670086201963, "loss": 5.4256, "step": 178700 }, { "epoch": 1.077833118269196, "grad_norm": 0.8321172595024109, "learning_rate": 0.00028326736241805143, "loss": 5.0521, "step": 178750 }, { "epoch": 1.07813461005053, "grad_norm": 2.037337303161621, "learning_rate": 0.00028325802152148765, "loss": 5.1535, "step": 178800 }, { "epoch": 1.078436101831864, "grad_norm": 1.9060475826263428, "learning_rate": 0.00028324867817250014, "loss": 5.2828, "step": 178850 }, { "epoch": 1.078737593613198, "grad_norm": 1.5095758438110352, "learning_rate": 0.00028323933237126085, "loss": 5.1893, "step": 178900 }, { "epoch": 1.0790390853945322, "grad_norm": 1.4009321928024292, "learning_rate": 0.00028322998411794184, "loss": 4.8386, "step": 178950 }, { "epoch": 1.0793405771758662, "grad_norm": 1.0421829223632812, "learning_rate": 0.00028322063341271506, "loss": 5.2972, "step": 179000 }, { "epoch": 1.0796420689572002, "grad_norm": 1.2256029844284058, "learning_rate": 0.00028321128025575264, "loss": 5.069, "step": 179050 }, { "epoch": 1.0799435607385344, "grad_norm": 1.8486034870147705, "learning_rate": 0.00028320192464722675, "loss": 5.452, "step": 179100 }, { "epoch": 1.0802450525198684, "grad_norm": 1.569428563117981, "learning_rate": 0.0002831925665873095, "loss": 5.376, "step": 179150 }, { "epoch": 1.0805465443012023, "grad_norm": 1.4246095418930054, "learning_rate": 0.0002831832060761731, "loss": 5.4712, "step": 179200 }, { "epoch": 1.0808480360825363, "grad_norm": 1.0885868072509766, "learning_rate": 0.0002831738431139899, "loss": 5.6897, "step": 179250 }, { "epoch": 1.0811495278638705, "grad_norm": 1.535190463066101, "learning_rate": 0.0002831644777009321, "loss": 4.9069, "step": 179300 }, { "epoch": 1.0814510196452045, "grad_norm": 2.253230333328247, "learning_rate": 0.00028315510983717223, "loss": 5.075, "step": 179350 }, { "epoch": 1.0817525114265385, "grad_norm": 0.9288419485092163, "learning_rate": 0.00028314573952288255, "loss": 5.2914, "step": 179400 }, { "epoch": 1.0820540032078725, "grad_norm": 1.681607723236084, "learning_rate": 0.0002831363667582355, "loss": 5.3278, "step": 179450 }, { "epoch": 1.0823554949892067, "grad_norm": 1.3258296251296997, "learning_rate": 0.00028312699154340367, "loss": 5.0662, "step": 179500 }, { "epoch": 1.0826569867705407, "grad_norm": 0.8821828961372375, "learning_rate": 0.0002831176138785595, "loss": 5.3308, "step": 179550 }, { "epoch": 1.0829584785518747, "grad_norm": 1.716820478439331, "learning_rate": 0.0002831082337638756, "loss": 5.1768, "step": 179600 }, { "epoch": 1.0832599703332086, "grad_norm": 1.6377936601638794, "learning_rate": 0.0002830988511995247, "loss": 5.4938, "step": 179650 }, { "epoch": 1.0835614621145429, "grad_norm": 1.3539170026779175, "learning_rate": 0.0002830894661856793, "loss": 5.1156, "step": 179700 }, { "epoch": 1.0838629538958768, "grad_norm": 1.6222363710403442, "learning_rate": 0.00028308007872251224, "loss": 5.0811, "step": 179750 }, { "epoch": 1.0841644456772108, "grad_norm": 1.708933711051941, "learning_rate": 0.00028307068881019623, "loss": 5.6557, "step": 179800 }, { "epoch": 1.0844659374585448, "grad_norm": 1.6381220817565918, "learning_rate": 0.00028306129644890414, "loss": 5.0412, "step": 179850 }, { "epoch": 1.084767429239879, "grad_norm": 1.1277289390563965, "learning_rate": 0.00028305190163880874, "loss": 5.1514, "step": 179900 }, { "epoch": 1.085068921021213, "grad_norm": 1.4564307928085327, "learning_rate": 0.00028304250438008295, "loss": 5.033, "step": 179950 }, { "epoch": 1.085370412802547, "grad_norm": 0.4444867670536041, "learning_rate": 0.00028303310467289977, "loss": 5.1325, "step": 180000 }, { "epoch": 1.085370412802547, "eval_loss": 5.639659881591797, "eval_runtime": 38.9214, "eval_samples_per_second": 13.155, "eval_steps_per_second": 6.577, "eval_tts_loss": 7.19022297841909, "step": 180000 }, { "epoch": 1.085671904583881, "grad_norm": 1.3935341835021973, "learning_rate": 0.0002830237025174321, "loss": 4.5863, "step": 180050 }, { "epoch": 1.0859733963652152, "grad_norm": 1.1704981327056885, "learning_rate": 0.00028301429791385304, "loss": 4.9183, "step": 180100 }, { "epoch": 1.0862748881465492, "grad_norm": 1.5442876815795898, "learning_rate": 0.00028300489086233567, "loss": 5.1735, "step": 180150 }, { "epoch": 1.0865763799278831, "grad_norm": 1.0316473245620728, "learning_rate": 0.0002829954813630531, "loss": 4.7904, "step": 180200 }, { "epoch": 1.0868778717092171, "grad_norm": 1.3085618019104004, "learning_rate": 0.0002829860694161784, "loss": 4.5738, "step": 180250 }, { "epoch": 1.0871793634905513, "grad_norm": 1.4784886837005615, "learning_rate": 0.000282976655021885, "loss": 5.0275, "step": 180300 }, { "epoch": 1.0874808552718853, "grad_norm": 1.1743570566177368, "learning_rate": 0.00028296723818034596, "loss": 4.6868, "step": 180350 }, { "epoch": 1.0877823470532193, "grad_norm": 0.8532216548919678, "learning_rate": 0.00028295781889173466, "loss": 5.1874, "step": 180400 }, { "epoch": 1.0880838388345533, "grad_norm": 1.2920033931732178, "learning_rate": 0.00028294839715622445, "loss": 5.0027, "step": 180450 }, { "epoch": 1.0883853306158875, "grad_norm": 1.404274582862854, "learning_rate": 0.0002829389729739887, "loss": 5.1077, "step": 180500 }, { "epoch": 1.0886868223972215, "grad_norm": 1.271293044090271, "learning_rate": 0.0002829295463452009, "loss": 5.0661, "step": 180550 }, { "epoch": 1.0889883141785555, "grad_norm": 2.0502521991729736, "learning_rate": 0.0002829201172700345, "loss": 5.2893, "step": 180600 }, { "epoch": 1.0892898059598894, "grad_norm": 1.0069313049316406, "learning_rate": 0.00028291068574866303, "loss": 5.2707, "step": 180650 }, { "epoch": 1.0895912977412237, "grad_norm": 2.376244068145752, "learning_rate": 0.00028290125178126, "loss": 4.497, "step": 180700 }, { "epoch": 1.0898927895225576, "grad_norm": 1.6725925207138062, "learning_rate": 0.0002828918153679992, "loss": 5.2615, "step": 180750 }, { "epoch": 1.0901942813038916, "grad_norm": 1.686850905418396, "learning_rate": 0.00028288237650905406, "loss": 5.0921, "step": 180800 }, { "epoch": 1.0904957730852256, "grad_norm": 0.7315508127212524, "learning_rate": 0.0002828729352045985, "loss": 4.8895, "step": 180850 }, { "epoch": 1.0907972648665598, "grad_norm": 1.179991364479065, "learning_rate": 0.00028286349145480616, "loss": 5.3122, "step": 180900 }, { "epoch": 1.0910987566478938, "grad_norm": 1.833229660987854, "learning_rate": 0.0002828540452598509, "loss": 5.1379, "step": 180950 }, { "epoch": 1.0914002484292278, "grad_norm": 1.3566107749938965, "learning_rate": 0.0002828445966199065, "loss": 5.3704, "step": 181000 }, { "epoch": 1.0917017402105618, "grad_norm": 0.9752807021141052, "learning_rate": 0.00028283514553514687, "loss": 5.0474, "step": 181050 }, { "epoch": 1.092003231991896, "grad_norm": 1.4805009365081787, "learning_rate": 0.00028282569200574595, "loss": 5.3592, "step": 181100 }, { "epoch": 1.09230472377323, "grad_norm": 0.7126824855804443, "learning_rate": 0.00028281623603187777, "loss": 5.2154, "step": 181150 }, { "epoch": 1.092606215554564, "grad_norm": 1.2091187238693237, "learning_rate": 0.00028280677761371623, "loss": 4.9032, "step": 181200 }, { "epoch": 1.092907707335898, "grad_norm": 1.896178126335144, "learning_rate": 0.0002827973167514355, "loss": 5.0806, "step": 181250 }, { "epoch": 1.0932091991172321, "grad_norm": 0.3042243719100952, "learning_rate": 0.0002827878534452097, "loss": 5.0008, "step": 181300 }, { "epoch": 1.0935106908985661, "grad_norm": 1.62445068359375, "learning_rate": 0.0002827783876952129, "loss": 5.2947, "step": 181350 }, { "epoch": 1.0938121826799, "grad_norm": 1.5339677333831787, "learning_rate": 0.00028276891950161946, "loss": 5.4291, "step": 181400 }, { "epoch": 1.094113674461234, "grad_norm": 0.5309788584709167, "learning_rate": 0.0002827594488646034, "loss": 5.4967, "step": 181450 }, { "epoch": 1.0944151662425683, "grad_norm": 1.7395565509796143, "learning_rate": 0.0002827499757843392, "loss": 4.9544, "step": 181500 }, { "epoch": 1.0947166580239023, "grad_norm": 0.701613187789917, "learning_rate": 0.00028274050026100116, "loss": 4.7898, "step": 181550 }, { "epoch": 1.0950181498052363, "grad_norm": 0.5083491802215576, "learning_rate": 0.0002827310222947637, "loss": 4.8579, "step": 181600 }, { "epoch": 1.0953196415865702, "grad_norm": 1.2273558378219604, "learning_rate": 0.00028272154188580103, "loss": 5.4872, "step": 181650 }, { "epoch": 1.0956211333679045, "grad_norm": 1.168215274810791, "learning_rate": 0.0002827120590342879, "loss": 5.1327, "step": 181700 }, { "epoch": 1.0959226251492384, "grad_norm": 1.9095325469970703, "learning_rate": 0.0002827025737403987, "loss": 5.2272, "step": 181750 }, { "epoch": 1.0962241169305724, "grad_norm": 1.0929069519042969, "learning_rate": 0.000282693086004308, "loss": 5.1498, "step": 181800 }, { "epoch": 1.0965256087119064, "grad_norm": 1.1653860807418823, "learning_rate": 0.0002826835958261904, "loss": 5.2684, "step": 181850 }, { "epoch": 1.0968271004932406, "grad_norm": 1.2424743175506592, "learning_rate": 0.00028267410320622054, "loss": 4.599, "step": 181900 }, { "epoch": 1.0971285922745746, "grad_norm": 1.0296772718429565, "learning_rate": 0.00028266460814457324, "loss": 5.5328, "step": 181950 }, { "epoch": 1.0974300840559086, "grad_norm": 1.137259840965271, "learning_rate": 0.0002826551106414231, "loss": 4.8692, "step": 182000 }, { "epoch": 1.0977315758372428, "grad_norm": 1.6612836122512817, "learning_rate": 0.00028264561069694493, "loss": 5.4967, "step": 182050 }, { "epoch": 1.0980330676185768, "grad_norm": 1.411877155303955, "learning_rate": 0.0002826361083113136, "loss": 5.403, "step": 182100 }, { "epoch": 1.0983345593999108, "grad_norm": 1.4006059169769287, "learning_rate": 0.000282626603484704, "loss": 5.2646, "step": 182150 }, { "epoch": 1.0986360511812447, "grad_norm": 1.2626278400421143, "learning_rate": 0.000282617096217291, "loss": 5.4787, "step": 182200 }, { "epoch": 1.0989375429625787, "grad_norm": 1.2244073152542114, "learning_rate": 0.00028260758650924963, "loss": 5.0916, "step": 182250 }, { "epoch": 1.099239034743913, "grad_norm": 1.3744454383850098, "learning_rate": 0.00028259807436075487, "loss": 4.8254, "step": 182300 }, { "epoch": 1.099540526525247, "grad_norm": 1.3178207874298096, "learning_rate": 0.00028258855977198176, "loss": 5.7149, "step": 182350 }, { "epoch": 1.099842018306581, "grad_norm": 0.7610295414924622, "learning_rate": 0.0002825790427431054, "loss": 5.1926, "step": 182400 }, { "epoch": 1.100143510087915, "grad_norm": 0.7692704200744629, "learning_rate": 0.00028256952327430104, "loss": 5.1926, "step": 182450 }, { "epoch": 1.100445001869249, "grad_norm": 1.0019376277923584, "learning_rate": 0.00028256000136574373, "loss": 4.6579, "step": 182500 }, { "epoch": 1.100746493650583, "grad_norm": 1.1856638193130493, "learning_rate": 0.0002825504770176088, "loss": 5.1704, "step": 182550 }, { "epoch": 1.101047985431917, "grad_norm": 1.0865720510482788, "learning_rate": 0.00028254095023007146, "loss": 5.1341, "step": 182600 }, { "epoch": 1.101349477213251, "grad_norm": 1.1739696264266968, "learning_rate": 0.0002825314210033071, "loss": 5.0482, "step": 182650 }, { "epoch": 1.1016509689945853, "grad_norm": 1.096392273902893, "learning_rate": 0.0002825218893374911, "loss": 4.7481, "step": 182700 }, { "epoch": 1.1019524607759192, "grad_norm": 0.5198074579238892, "learning_rate": 0.0002825123552327988, "loss": 4.974, "step": 182750 }, { "epoch": 1.1022539525572532, "grad_norm": 1.836201786994934, "learning_rate": 0.0002825028186894057, "loss": 4.779, "step": 182800 }, { "epoch": 1.1025554443385874, "grad_norm": 0.8041704297065735, "learning_rate": 0.00028249327970748735, "loss": 4.836, "step": 182850 }, { "epoch": 1.1028569361199214, "grad_norm": 1.2546354532241821, "learning_rate": 0.00028248373828721927, "loss": 5.3051, "step": 182900 }, { "epoch": 1.1031584279012554, "grad_norm": 1.5982600450515747, "learning_rate": 0.000282474194428777, "loss": 5.2604, "step": 182950 }, { "epoch": 1.1034599196825894, "grad_norm": 1.2125132083892822, "learning_rate": 0.0002824646481323363, "loss": 4.8979, "step": 183000 }, { "epoch": 1.1034599196825894, "eval_loss": 5.674833297729492, "eval_runtime": 39.2042, "eval_samples_per_second": 13.06, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.147015316881288, "step": 183000 }, { "epoch": 1.1037614114639236, "grad_norm": 1.513508677482605, "learning_rate": 0.0002824550993980727, "loss": 4.9257, "step": 183050 }, { "epoch": 1.1040629032452576, "grad_norm": 1.157760739326477, "learning_rate": 0.0002824455482261621, "loss": 5.4911, "step": 183100 }, { "epoch": 1.1043643950265916, "grad_norm": 1.1123489141464233, "learning_rate": 0.00028243599461678015, "loss": 5.3499, "step": 183150 }, { "epoch": 1.1046658868079255, "grad_norm": 1.2326006889343262, "learning_rate": 0.0002824264385701027, "loss": 4.8033, "step": 183200 }, { "epoch": 1.1049673785892598, "grad_norm": 1.1977006196975708, "learning_rate": 0.0002824168800863057, "loss": 5.266, "step": 183250 }, { "epoch": 1.1052688703705937, "grad_norm": 1.620728611946106, "learning_rate": 0.00028240731916556494, "loss": 5.5051, "step": 183300 }, { "epoch": 1.1055703621519277, "grad_norm": 1.070458173751831, "learning_rate": 0.0002823977558080564, "loss": 4.9118, "step": 183350 }, { "epoch": 1.1058718539332617, "grad_norm": 0.257414847612381, "learning_rate": 0.0002823881900139562, "loss": 4.9058, "step": 183400 }, { "epoch": 1.106173345714596, "grad_norm": 1.6216927766799927, "learning_rate": 0.00028237862178344023, "loss": 4.9949, "step": 183450 }, { "epoch": 1.10647483749593, "grad_norm": 1.648545503616333, "learning_rate": 0.00028236905111668465, "loss": 5.3269, "step": 183500 }, { "epoch": 1.1067763292772639, "grad_norm": 0.8659826517105103, "learning_rate": 0.0002823594780138656, "loss": 5.4454, "step": 183550 }, { "epoch": 1.1070778210585979, "grad_norm": 1.614243507385254, "learning_rate": 0.0002823499024751592, "loss": 5.1683, "step": 183600 }, { "epoch": 1.107379312839932, "grad_norm": 1.5513204336166382, "learning_rate": 0.0002823403245007417, "loss": 5.3109, "step": 183650 }, { "epoch": 1.107680804621266, "grad_norm": 1.0670057535171509, "learning_rate": 0.0002823307440907895, "loss": 5.2271, "step": 183700 }, { "epoch": 1.1079822964026, "grad_norm": 1.30763840675354, "learning_rate": 0.0002823211612454787, "loss": 5.1672, "step": 183750 }, { "epoch": 1.108283788183934, "grad_norm": 2.5280392169952393, "learning_rate": 0.0002823115759649858, "loss": 5.0741, "step": 183800 }, { "epoch": 1.1085852799652682, "grad_norm": 1.056873083114624, "learning_rate": 0.0002823019882494872, "loss": 5.0649, "step": 183850 }, { "epoch": 1.1088867717466022, "grad_norm": 0.7075055241584778, "learning_rate": 0.00028229239809915923, "loss": 5.2355, "step": 183900 }, { "epoch": 1.1091882635279362, "grad_norm": 1.3778613805770874, "learning_rate": 0.0002822828055141785, "loss": 5.177, "step": 183950 }, { "epoch": 1.1094897553092702, "grad_norm": 1.0943998098373413, "learning_rate": 0.0002822732104947215, "loss": 5.4113, "step": 184000 }, { "epoch": 1.1097912470906044, "grad_norm": 1.6085234880447388, "learning_rate": 0.0002822636130409648, "loss": 5.3019, "step": 184050 }, { "epoch": 1.1100927388719384, "grad_norm": 1.0888705253601074, "learning_rate": 0.00028225401315308513, "loss": 5.0588, "step": 184100 }, { "epoch": 1.1103942306532724, "grad_norm": 1.140872836112976, "learning_rate": 0.00028224441083125906, "loss": 5.1417, "step": 184150 }, { "epoch": 1.1106957224346063, "grad_norm": 1.6326576471328735, "learning_rate": 0.0002822348060756633, "loss": 5.1718, "step": 184200 }, { "epoch": 1.1109972142159406, "grad_norm": 1.3469504117965698, "learning_rate": 0.00028222519888647466, "loss": 5.2924, "step": 184250 }, { "epoch": 1.1112987059972745, "grad_norm": 1.232838749885559, "learning_rate": 0.00028221558926386995, "loss": 5.2404, "step": 184300 }, { "epoch": 1.1116001977786085, "grad_norm": 1.6275427341461182, "learning_rate": 0.000282205977208026, "loss": 4.7466, "step": 184350 }, { "epoch": 1.1119016895599425, "grad_norm": 1.5211961269378662, "learning_rate": 0.0002821963627191196, "loss": 5.2292, "step": 184400 }, { "epoch": 1.1122031813412767, "grad_norm": 1.2359402179718018, "learning_rate": 0.00028218674579732795, "loss": 5.227, "step": 184450 }, { "epoch": 1.1125046731226107, "grad_norm": 2.368140459060669, "learning_rate": 0.00028217712644282777, "loss": 5.0427, "step": 184500 }, { "epoch": 1.1128061649039447, "grad_norm": 1.535837173461914, "learning_rate": 0.00028216750465579626, "loss": 5.3118, "step": 184550 }, { "epoch": 1.1131076566852787, "grad_norm": 1.5698790550231934, "learning_rate": 0.0002821578804364105, "loss": 5.5192, "step": 184600 }, { "epoch": 1.1134091484666129, "grad_norm": 1.5717387199401855, "learning_rate": 0.0002821482537848475, "loss": 5.1247, "step": 184650 }, { "epoch": 1.1137106402479469, "grad_norm": 1.471139907836914, "learning_rate": 0.00028213862470128444, "loss": 5.2914, "step": 184700 }, { "epoch": 1.1140121320292808, "grad_norm": 1.617720365524292, "learning_rate": 0.0002821289931858986, "loss": 5.1044, "step": 184750 }, { "epoch": 1.1143136238106148, "grad_norm": 1.4011073112487793, "learning_rate": 0.0002821193592388671, "loss": 4.9496, "step": 184800 }, { "epoch": 1.114615115591949, "grad_norm": 1.9814960956573486, "learning_rate": 0.0002821097228603675, "loss": 4.7257, "step": 184850 }, { "epoch": 1.114916607373283, "grad_norm": 1.7565337419509888, "learning_rate": 0.00028210008405057685, "loss": 5.2262, "step": 184900 }, { "epoch": 1.115218099154617, "grad_norm": 1.720570683479309, "learning_rate": 0.00028209044280967275, "loss": 5.3866, "step": 184950 }, { "epoch": 1.115519590935951, "grad_norm": 0.9250810742378235, "learning_rate": 0.00028208079913783255, "loss": 5.2508, "step": 185000 }, { "epoch": 1.1158210827172852, "grad_norm": 1.163602590560913, "learning_rate": 0.0002820711530352337, "loss": 4.7788, "step": 185050 }, { "epoch": 1.1161225744986192, "grad_norm": 0.6637985110282898, "learning_rate": 0.00028206150450205375, "loss": 5.2007, "step": 185100 }, { "epoch": 1.1164240662799532, "grad_norm": 1.8817274570465088, "learning_rate": 0.0002820518535384703, "loss": 4.8412, "step": 185150 }, { "epoch": 1.1167255580612871, "grad_norm": 1.2591520547866821, "learning_rate": 0.00028204220014466086, "loss": 4.9274, "step": 185200 }, { "epoch": 1.1170270498426214, "grad_norm": 1.0828992128372192, "learning_rate": 0.00028203254432080326, "loss": 4.856, "step": 185250 }, { "epoch": 1.1173285416239553, "grad_norm": 1.033410906791687, "learning_rate": 0.00028202288606707506, "loss": 5.448, "step": 185300 }, { "epoch": 1.1176300334052893, "grad_norm": 1.3814611434936523, "learning_rate": 0.000282013225383654, "loss": 5.5423, "step": 185350 }, { "epoch": 1.1179315251866233, "grad_norm": 1.6649916172027588, "learning_rate": 0.000282003562270718, "loss": 5.1804, "step": 185400 }, { "epoch": 1.1182330169679575, "grad_norm": 0.989023745059967, "learning_rate": 0.0002819938967284448, "loss": 5.1112, "step": 185450 }, { "epoch": 1.1185345087492915, "grad_norm": 1.009531855583191, "learning_rate": 0.00028198422875701224, "loss": 4.7107, "step": 185500 }, { "epoch": 1.1188360005306255, "grad_norm": 1.321474313735962, "learning_rate": 0.00028197455835659836, "loss": 4.8071, "step": 185550 }, { "epoch": 1.1191374923119595, "grad_norm": 1.7131448984146118, "learning_rate": 0.0002819648855273811, "loss": 4.6822, "step": 185600 }, { "epoch": 1.1194389840932937, "grad_norm": 0.9823644757270813, "learning_rate": 0.00028195521026953835, "loss": 4.7335, "step": 185650 }, { "epoch": 1.1197404758746277, "grad_norm": 2.1484534740448, "learning_rate": 0.0002819455325832483, "loss": 5.2688, "step": 185700 }, { "epoch": 1.1200419676559616, "grad_norm": 1.1639310121536255, "learning_rate": 0.00028193585246868904, "loss": 4.6837, "step": 185750 }, { "epoch": 1.1203434594372959, "grad_norm": 1.379204273223877, "learning_rate": 0.00028192616992603873, "loss": 5.1475, "step": 185800 }, { "epoch": 1.1206449512186298, "grad_norm": 1.1789718866348267, "learning_rate": 0.00028191648495547546, "loss": 5.5038, "step": 185850 }, { "epoch": 1.1209464429999638, "grad_norm": 0.7003433704376221, "learning_rate": 0.0002819067975571776, "loss": 5.1124, "step": 185900 }, { "epoch": 1.1212479347812978, "grad_norm": 1.226332664489746, "learning_rate": 0.0002818971077313233, "loss": 5.09, "step": 185950 }, { "epoch": 1.1215494265626318, "grad_norm": 1.4952410459518433, "learning_rate": 0.00028188741547809104, "loss": 4.9701, "step": 186000 }, { "epoch": 1.1215494265626318, "eval_loss": 5.638899803161621, "eval_runtime": 39.0104, "eval_samples_per_second": 13.125, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.196434325060411, "step": 186000 }, { "epoch": 1.121850918343966, "grad_norm": 1.1668733358383179, "learning_rate": 0.00028187772079765907, "loss": 4.9606, "step": 186050 }, { "epoch": 1.1221524101253, "grad_norm": 1.7647494077682495, "learning_rate": 0.00028186802369020584, "loss": 5.3608, "step": 186100 }, { "epoch": 1.122453901906634, "grad_norm": 0.8735373020172119, "learning_rate": 0.0002818583241559098, "loss": 5.3697, "step": 186150 }, { "epoch": 1.1227553936879682, "grad_norm": 1.2058995962142944, "learning_rate": 0.00028184862219494955, "loss": 5.0266, "step": 186200 }, { "epoch": 1.1230568854693022, "grad_norm": 1.1616865396499634, "learning_rate": 0.0002818389178075035, "loss": 5.0199, "step": 186250 }, { "epoch": 1.1233583772506361, "grad_norm": 2.529729127883911, "learning_rate": 0.00028182921099375035, "loss": 5.0698, "step": 186300 }, { "epoch": 1.1236598690319701, "grad_norm": 1.9741357564926147, "learning_rate": 0.0002818195017538687, "loss": 5.2134, "step": 186350 }, { "epoch": 1.1239613608133043, "grad_norm": 2.0783605575561523, "learning_rate": 0.0002818097900880372, "loss": 5.1378, "step": 186400 }, { "epoch": 1.1242628525946383, "grad_norm": 3.145364761352539, "learning_rate": 0.00028180007599643464, "loss": 5.232, "step": 186450 }, { "epoch": 1.1245643443759723, "grad_norm": 1.590342402458191, "learning_rate": 0.0002817903594792398, "loss": 5.1764, "step": 186500 }, { "epoch": 1.1248658361573063, "grad_norm": 1.2509845495224, "learning_rate": 0.0002817806405366314, "loss": 4.7463, "step": 186550 }, { "epoch": 1.1251673279386405, "grad_norm": 2.055493116378784, "learning_rate": 0.0002817709191687884, "loss": 5.1817, "step": 186600 }, { "epoch": 1.1254688197199745, "grad_norm": 1.6149464845657349, "learning_rate": 0.00028176119537588973, "loss": 4.9546, "step": 186650 }, { "epoch": 1.1257703115013085, "grad_norm": 1.3227636814117432, "learning_rate": 0.0002817514691581143, "loss": 5.3014, "step": 186700 }, { "epoch": 1.1260718032826424, "grad_norm": 0.8608360886573792, "learning_rate": 0.000281741740515641, "loss": 5.2005, "step": 186750 }, { "epoch": 1.1263732950639764, "grad_norm": 2.0555150508880615, "learning_rate": 0.00028173200944864904, "loss": 5.506, "step": 186800 }, { "epoch": 1.1266747868453106, "grad_norm": 1.496124267578125, "learning_rate": 0.0002817222759573174, "loss": 4.8304, "step": 186850 }, { "epoch": 1.1269762786266446, "grad_norm": 0.9972660541534424, "learning_rate": 0.00028171254004182527, "loss": 5.1607, "step": 186900 }, { "epoch": 1.1272777704079786, "grad_norm": 2.080538272857666, "learning_rate": 0.00028170280170235185, "loss": 5.2115, "step": 186950 }, { "epoch": 1.1275792621893128, "grad_norm": 1.3131717443466187, "learning_rate": 0.0002816930609390763, "loss": 5.5431, "step": 187000 }, { "epoch": 1.1278807539706468, "grad_norm": 1.8158154487609863, "learning_rate": 0.00028168331775217783, "loss": 5.1852, "step": 187050 }, { "epoch": 1.1281822457519808, "grad_norm": 1.0371254682540894, "learning_rate": 0.00028167357214183583, "loss": 5.4746, "step": 187100 }, { "epoch": 1.1284837375333148, "grad_norm": 1.1879075765609741, "learning_rate": 0.00028166382410822966, "loss": 4.6675, "step": 187150 }, { "epoch": 1.128785229314649, "grad_norm": 1.0669647455215454, "learning_rate": 0.00028165407365153875, "loss": 5.3289, "step": 187200 }, { "epoch": 1.129086721095983, "grad_norm": 1.3276599645614624, "learning_rate": 0.00028164432077194243, "loss": 5.2643, "step": 187250 }, { "epoch": 1.129388212877317, "grad_norm": 2.383532762527466, "learning_rate": 0.0002816345654696203, "loss": 5.0034, "step": 187300 }, { "epoch": 1.129689704658651, "grad_norm": 1.0827738046646118, "learning_rate": 0.00028162480774475185, "loss": 5.365, "step": 187350 }, { "epoch": 1.1299911964399851, "grad_norm": 1.4025092124938965, "learning_rate": 0.0002816150475975166, "loss": 4.9655, "step": 187400 }, { "epoch": 1.1302926882213191, "grad_norm": 1.5408847332000732, "learning_rate": 0.00028160528502809426, "loss": 5.5101, "step": 187450 }, { "epoch": 1.130594180002653, "grad_norm": 1.3620647192001343, "learning_rate": 0.00028159552003666444, "loss": 5.1103, "step": 187500 }, { "epoch": 1.130895671783987, "grad_norm": 1.1760339736938477, "learning_rate": 0.0002815857526234069, "loss": 5.1209, "step": 187550 }, { "epoch": 1.1311971635653213, "grad_norm": 1.1580239534378052, "learning_rate": 0.00028157598278850135, "loss": 4.9047, "step": 187600 }, { "epoch": 1.1314986553466553, "grad_norm": 1.2345781326293945, "learning_rate": 0.0002815662105321276, "loss": 5.1453, "step": 187650 }, { "epoch": 1.1318001471279893, "grad_norm": 1.1386715173721313, "learning_rate": 0.0002815564358544655, "loss": 5.4345, "step": 187700 }, { "epoch": 1.1321016389093232, "grad_norm": 1.196548581123352, "learning_rate": 0.00028154665875569494, "loss": 5.6465, "step": 187750 }, { "epoch": 1.1324031306906575, "grad_norm": 1.3208179473876953, "learning_rate": 0.00028153687923599584, "loss": 5.1246, "step": 187800 }, { "epoch": 1.1327046224719914, "grad_norm": 1.304329514503479, "learning_rate": 0.00028152709729554825, "loss": 5.1862, "step": 187850 }, { "epoch": 1.1330061142533254, "grad_norm": 1.3608157634735107, "learning_rate": 0.0002815173129345321, "loss": 5.2409, "step": 187900 }, { "epoch": 1.1333076060346594, "grad_norm": 1.5249302387237549, "learning_rate": 0.0002815075261531275, "loss": 5.035, "step": 187950 }, { "epoch": 1.1336090978159936, "grad_norm": 1.0991209745407104, "learning_rate": 0.0002814977369515145, "loss": 5.2057, "step": 188000 }, { "epoch": 1.1339105895973276, "grad_norm": 1.6286346912384033, "learning_rate": 0.0002814879453298733, "loss": 5.2307, "step": 188050 }, { "epoch": 1.1342120813786616, "grad_norm": 2.454179286956787, "learning_rate": 0.0002814781512883842, "loss": 5.0911, "step": 188100 }, { "epoch": 1.1345135731599956, "grad_norm": 1.099913477897644, "learning_rate": 0.00028146835482722733, "loss": 5.1332, "step": 188150 }, { "epoch": 1.1348150649413298, "grad_norm": 1.340327501296997, "learning_rate": 0.00028145855594658293, "loss": 5.0746, "step": 188200 }, { "epoch": 1.1351165567226638, "grad_norm": 0.41060903668403625, "learning_rate": 0.0002814487546466315, "loss": 5.0766, "step": 188250 }, { "epoch": 1.1354180485039977, "grad_norm": 1.1046122312545776, "learning_rate": 0.0002814389509275533, "loss": 5.0444, "step": 188300 }, { "epoch": 1.1357195402853317, "grad_norm": 1.2183218002319336, "learning_rate": 0.00028142914478952876, "loss": 4.6885, "step": 188350 }, { "epoch": 1.136021032066666, "grad_norm": 1.9328845739364624, "learning_rate": 0.00028141933623273836, "loss": 5.2866, "step": 188400 }, { "epoch": 1.136322523848, "grad_norm": 1.3735898733139038, "learning_rate": 0.00028140952525736266, "loss": 4.8668, "step": 188450 }, { "epoch": 1.136624015629334, "grad_norm": 0.9694177508354187, "learning_rate": 0.0002813997118635822, "loss": 5.1682, "step": 188500 }, { "epoch": 1.136925507410668, "grad_norm": 0.21731162071228027, "learning_rate": 0.0002813898960515775, "loss": 5.3472, "step": 188550 }, { "epoch": 1.137226999192002, "grad_norm": 1.379750370979309, "learning_rate": 0.00028138007782152924, "loss": 5.3659, "step": 188600 }, { "epoch": 1.137528490973336, "grad_norm": 1.1582145690917969, "learning_rate": 0.00028137025717361826, "loss": 5.2053, "step": 188650 }, { "epoch": 1.13782998275467, "grad_norm": 2.0517632961273193, "learning_rate": 0.00028136043410802506, "loss": 5.4103, "step": 188700 }, { "epoch": 1.1381314745360043, "grad_norm": 0.40456894040107727, "learning_rate": 0.00028135060862493055, "loss": 5.1272, "step": 188750 }, { "epoch": 1.1384329663173383, "grad_norm": 1.8980306386947632, "learning_rate": 0.00028134078072451557, "loss": 4.839, "step": 188800 }, { "epoch": 1.1387344580986722, "grad_norm": 1.1142340898513794, "learning_rate": 0.00028133095040696097, "loss": 4.7691, "step": 188850 }, { "epoch": 1.1390359498800062, "grad_norm": 2.031322717666626, "learning_rate": 0.0002813211176724476, "loss": 4.7452, "step": 188900 }, { "epoch": 1.1393374416613402, "grad_norm": 1.389369249343872, "learning_rate": 0.0002813112825211565, "loss": 5.0179, "step": 188950 }, { "epoch": 1.1396389334426744, "grad_norm": 0.27312445640563965, "learning_rate": 0.0002813014449532686, "loss": 5.1895, "step": 189000 }, { "epoch": 1.1396389334426744, "eval_loss": 5.630044937133789, "eval_runtime": 38.9897, "eval_samples_per_second": 13.132, "eval_steps_per_second": 6.566, "eval_tts_loss": 7.228348841062322, "step": 189000 }, { "epoch": 1.1399404252240084, "grad_norm": 1.6915806531906128, "learning_rate": 0.000281291604968965, "loss": 5.0964, "step": 189050 }, { "epoch": 1.1402419170053424, "grad_norm": 1.5885722637176514, "learning_rate": 0.0002812817625684268, "loss": 5.6222, "step": 189100 }, { "epoch": 1.1405434087866766, "grad_norm": 1.6738959550857544, "learning_rate": 0.000281271917751835, "loss": 5.2246, "step": 189150 }, { "epoch": 1.1408449005680106, "grad_norm": 2.651317834854126, "learning_rate": 0.000281262070519371, "loss": 5.2388, "step": 189200 }, { "epoch": 1.1411463923493446, "grad_norm": 1.2048159837722778, "learning_rate": 0.0002812522208712159, "loss": 4.929, "step": 189250 }, { "epoch": 1.1414478841306785, "grad_norm": 1.2662147283554077, "learning_rate": 0.000281242368807551, "loss": 5.1652, "step": 189300 }, { "epoch": 1.1417493759120125, "grad_norm": 0.9117621779441833, "learning_rate": 0.0002812325143285575, "loss": 5.1383, "step": 189350 }, { "epoch": 1.1420508676933467, "grad_norm": 1.5050725936889648, "learning_rate": 0.000281222657434417, "loss": 5.231, "step": 189400 }, { "epoch": 1.1423523594746807, "grad_norm": 2.3532824516296387, "learning_rate": 0.0002812127981253106, "loss": 5.1672, "step": 189450 }, { "epoch": 1.1426538512560147, "grad_norm": 1.2174291610717773, "learning_rate": 0.00028120293640142, "loss": 5.4529, "step": 189500 }, { "epoch": 1.142955343037349, "grad_norm": 1.5448813438415527, "learning_rate": 0.0002811930722629266, "loss": 5.169, "step": 189550 }, { "epoch": 1.143256834818683, "grad_norm": 1.1378343105316162, "learning_rate": 0.00028118320571001186, "loss": 4.539, "step": 189600 }, { "epoch": 1.1435583266000169, "grad_norm": 1.0695677995681763, "learning_rate": 0.00028117333674285746, "loss": 4.4516, "step": 189650 }, { "epoch": 1.1438598183813509, "grad_norm": 1.6760063171386719, "learning_rate": 0.000281163465361645, "loss": 5.1483, "step": 189700 }, { "epoch": 1.1441613101626849, "grad_norm": 1.0583415031433105, "learning_rate": 0.0002811535915665561, "loss": 5.2697, "step": 189750 }, { "epoch": 1.144462801944019, "grad_norm": 1.4690290689468384, "learning_rate": 0.00028114371535777256, "loss": 5.4844, "step": 189800 }, { "epoch": 1.144764293725353, "grad_norm": 1.0141549110412598, "learning_rate": 0.0002811338367354761, "loss": 5.1015, "step": 189850 }, { "epoch": 1.145065785506687, "grad_norm": 1.112923264503479, "learning_rate": 0.00028112395569984845, "loss": 5.0426, "step": 189900 }, { "epoch": 1.1453672772880212, "grad_norm": 1.6862199306488037, "learning_rate": 0.00028111407225107156, "loss": 4.94, "step": 189950 }, { "epoch": 1.1456687690693552, "grad_norm": 2.907735586166382, "learning_rate": 0.00028110418638932727, "loss": 5.3655, "step": 190000 }, { "epoch": 1.1459702608506892, "grad_norm": 1.6840403079986572, "learning_rate": 0.0002810942981147975, "loss": 5.2958, "step": 190050 }, { "epoch": 1.1462717526320232, "grad_norm": 1.612610936164856, "learning_rate": 0.0002810844074276644, "loss": 5.3314, "step": 190100 }, { "epoch": 1.1465732444133572, "grad_norm": 1.1632415056228638, "learning_rate": 0.00028107451432810964, "loss": 4.6354, "step": 190150 }, { "epoch": 1.1468747361946914, "grad_norm": 1.3001283407211304, "learning_rate": 0.0002810646188163156, "loss": 5.1801, "step": 190200 }, { "epoch": 1.1471762279760254, "grad_norm": 1.4298402070999146, "learning_rate": 0.0002810547208924643, "loss": 4.7062, "step": 190250 }, { "epoch": 1.1474777197573593, "grad_norm": 1.017475962638855, "learning_rate": 0.0002810448205567379, "loss": 5.3913, "step": 190300 }, { "epoch": 1.1477792115386936, "grad_norm": 1.072568655014038, "learning_rate": 0.00028103491780931856, "loss": 5.0483, "step": 190350 }, { "epoch": 1.1480807033200275, "grad_norm": 2.0675764083862305, "learning_rate": 0.00028102501265038856, "loss": 5.22, "step": 190400 }, { "epoch": 1.1483821951013615, "grad_norm": 1.3992373943328857, "learning_rate": 0.00028101510508013016, "loss": 5.4676, "step": 190450 }, { "epoch": 1.1486836868826955, "grad_norm": 1.6903566122055054, "learning_rate": 0.0002810051950987257, "loss": 4.9046, "step": 190500 }, { "epoch": 1.1489851786640295, "grad_norm": 1.866632342338562, "learning_rate": 0.00028099528270635765, "loss": 5.2505, "step": 190550 }, { "epoch": 1.1492866704453637, "grad_norm": 1.1998556852340698, "learning_rate": 0.0002809853679032083, "loss": 5.3379, "step": 190600 }, { "epoch": 1.1495881622266977, "grad_norm": 0.3771515190601349, "learning_rate": 0.00028097545068946024, "loss": 4.7879, "step": 190650 }, { "epoch": 1.1498896540080317, "grad_norm": 0.8855519890785217, "learning_rate": 0.0002809655310652959, "loss": 5.3565, "step": 190700 }, { "epoch": 1.1501911457893659, "grad_norm": 1.940394639968872, "learning_rate": 0.0002809556090308978, "loss": 5.1362, "step": 190750 }, { "epoch": 1.1504926375706999, "grad_norm": 0.643152117729187, "learning_rate": 0.00028094568458644873, "loss": 4.7379, "step": 190800 }, { "epoch": 1.1507941293520338, "grad_norm": 1.4545855522155762, "learning_rate": 0.00028093575773213106, "loss": 4.9737, "step": 190850 }, { "epoch": 1.1510956211333678, "grad_norm": 1.2519017457962036, "learning_rate": 0.0002809258284681277, "loss": 5.2075, "step": 190900 }, { "epoch": 1.151397112914702, "grad_norm": 0.6261531710624695, "learning_rate": 0.00028091589679462133, "loss": 5.0277, "step": 190950 }, { "epoch": 1.151698604696036, "grad_norm": 2.752915143966675, "learning_rate": 0.00028090596271179464, "loss": 5.3089, "step": 191000 }, { "epoch": 1.15200009647737, "grad_norm": 1.1344833374023438, "learning_rate": 0.00028089602621983055, "loss": 4.5506, "step": 191050 }, { "epoch": 1.152301588258704, "grad_norm": 1.728891372680664, "learning_rate": 0.0002808860873189119, "loss": 5.0759, "step": 191100 }, { "epoch": 1.1526030800400382, "grad_norm": 1.0889031887054443, "learning_rate": 0.0002808761460092216, "loss": 5.1565, "step": 191150 }, { "epoch": 1.1529045718213722, "grad_norm": 1.1134023666381836, "learning_rate": 0.0002808662022909426, "loss": 5.34, "step": 191200 }, { "epoch": 1.1532060636027062, "grad_norm": 1.167879581451416, "learning_rate": 0.0002808562561642579, "loss": 5.2444, "step": 191250 }, { "epoch": 1.1535075553840402, "grad_norm": 1.0854800939559937, "learning_rate": 0.00028084630762935056, "loss": 5.3165, "step": 191300 }, { "epoch": 1.1538090471653744, "grad_norm": 0.4949403703212738, "learning_rate": 0.00028083635668640363, "loss": 5.1932, "step": 191350 }, { "epoch": 1.1541105389467083, "grad_norm": 1.3907326459884644, "learning_rate": 0.0002808264033356003, "loss": 4.9304, "step": 191400 }, { "epoch": 1.1544120307280423, "grad_norm": 1.0247265100479126, "learning_rate": 0.00028081644757712367, "loss": 4.9445, "step": 191450 }, { "epoch": 1.1547135225093763, "grad_norm": 1.8694039583206177, "learning_rate": 0.0002808064894111571, "loss": 5.4443, "step": 191500 }, { "epoch": 1.1550150142907105, "grad_norm": 1.3744457960128784, "learning_rate": 0.0002807965288378836, "loss": 4.9991, "step": 191550 }, { "epoch": 1.1553165060720445, "grad_norm": 1.137048363685608, "learning_rate": 0.0002807865658574868, "loss": 5.1516, "step": 191600 }, { "epoch": 1.1556179978533785, "grad_norm": 1.506661295890808, "learning_rate": 0.00028077660047014984, "loss": 4.9916, "step": 191650 }, { "epoch": 1.1559194896347125, "grad_norm": 0.6899194121360779, "learning_rate": 0.0002807666326760562, "loss": 5.0296, "step": 191700 }, { "epoch": 1.1562209814160467, "grad_norm": 1.3620749711990356, "learning_rate": 0.0002807566624753892, "loss": 5.3125, "step": 191750 }, { "epoch": 1.1565224731973807, "grad_norm": 1.2256529331207275, "learning_rate": 0.00028074668986833256, "loss": 4.8226, "step": 191800 }, { "epoch": 1.1568239649787146, "grad_norm": 0.7661783695220947, "learning_rate": 0.00028073671485506963, "loss": 5.2351, "step": 191850 }, { "epoch": 1.1571254567600486, "grad_norm": 1.196624755859375, "learning_rate": 0.000280726737435784, "loss": 4.9678, "step": 191900 }, { "epoch": 1.1574269485413828, "grad_norm": 1.4586546421051025, "learning_rate": 0.00028071675761065935, "loss": 5.4388, "step": 191950 }, { "epoch": 1.1577284403227168, "grad_norm": 1.865029215812683, "learning_rate": 0.0002807067753798793, "loss": 5.6342, "step": 192000 }, { "epoch": 1.1577284403227168, "eval_loss": 5.628931999206543, "eval_runtime": 39.1613, "eval_samples_per_second": 13.074, "eval_steps_per_second": 6.537, "eval_tts_loss": 7.153658659977838, "step": 192000 }, { "epoch": 1.1580299321040508, "grad_norm": 1.6830133199691772, "learning_rate": 0.0002806967907436277, "loss": 5.3728, "step": 192050 }, { "epoch": 1.1583314238853848, "grad_norm": 1.6778476238250732, "learning_rate": 0.00028068680370208805, "loss": 5.0853, "step": 192100 }, { "epoch": 1.158632915666719, "grad_norm": 1.3882999420166016, "learning_rate": 0.00028067681425544426, "loss": 5.21, "step": 192150 }, { "epoch": 1.158934407448053, "grad_norm": 1.549195408821106, "learning_rate": 0.00028066682240388024, "loss": 5.1604, "step": 192200 }, { "epoch": 1.159235899229387, "grad_norm": 1.9642118215560913, "learning_rate": 0.0002806568281475798, "loss": 4.9836, "step": 192250 }, { "epoch": 1.159537391010721, "grad_norm": 1.1975209712982178, "learning_rate": 0.000280646831486727, "loss": 5.1545, "step": 192300 }, { "epoch": 1.1598388827920552, "grad_norm": 0.9973523616790771, "learning_rate": 0.0002806368324215056, "loss": 5.4118, "step": 192350 }, { "epoch": 1.1601403745733891, "grad_norm": 0.22077567875385284, "learning_rate": 0.00028062683095209974, "loss": 4.9046, "step": 192400 }, { "epoch": 1.1604418663547231, "grad_norm": 1.3318142890930176, "learning_rate": 0.0002806168270786935, "loss": 5.282, "step": 192450 }, { "epoch": 1.1607433581360573, "grad_norm": 1.6284310817718506, "learning_rate": 0.00028060682080147093, "loss": 5.6526, "step": 192500 }, { "epoch": 1.1610448499173913, "grad_norm": 1.3278313875198364, "learning_rate": 0.00028059681212061623, "loss": 5.4043, "step": 192550 }, { "epoch": 1.1613463416987253, "grad_norm": 1.2754052877426147, "learning_rate": 0.00028058680103631355, "loss": 5.6054, "step": 192600 }, { "epoch": 1.1616478334800593, "grad_norm": 1.1446794271469116, "learning_rate": 0.0002805767875487472, "loss": 5.1374, "step": 192650 }, { "epoch": 1.1619493252613933, "grad_norm": 1.0576056241989136, "learning_rate": 0.00028056677165810135, "loss": 5.4844, "step": 192700 }, { "epoch": 1.1622508170427275, "grad_norm": 1.1292272806167603, "learning_rate": 0.0002805567533645604, "loss": 5.4344, "step": 192750 }, { "epoch": 1.1625523088240615, "grad_norm": 1.4494476318359375, "learning_rate": 0.0002805467326683088, "loss": 5.0319, "step": 192800 }, { "epoch": 1.1628538006053954, "grad_norm": 1.1541330814361572, "learning_rate": 0.0002805367095695308, "loss": 5.5098, "step": 192850 }, { "epoch": 1.1631552923867297, "grad_norm": 0.6972658634185791, "learning_rate": 0.00028052668406841104, "loss": 4.8302, "step": 192900 }, { "epoch": 1.1634567841680636, "grad_norm": 1.552588939666748, "learning_rate": 0.00028051665616513386, "loss": 5.5425, "step": 192950 }, { "epoch": 1.1637582759493976, "grad_norm": 1.6909620761871338, "learning_rate": 0.00028050662585988386, "loss": 5.764, "step": 193000 }, { "epoch": 1.1640597677307316, "grad_norm": 0.8894562125205994, "learning_rate": 0.0002804965931528457, "loss": 4.8719, "step": 193050 }, { "epoch": 1.1643612595120656, "grad_norm": 1.8061796426773071, "learning_rate": 0.000280486558044204, "loss": 5.0512, "step": 193100 }, { "epoch": 1.1646627512933998, "grad_norm": 1.4103350639343262, "learning_rate": 0.00028047652053414334, "loss": 5.3875, "step": 193150 }, { "epoch": 1.1649642430747338, "grad_norm": 1.3658123016357422, "learning_rate": 0.0002804664806228486, "loss": 5.0669, "step": 193200 }, { "epoch": 1.1652657348560678, "grad_norm": 1.2820665836334229, "learning_rate": 0.0002804564383105044, "loss": 5.0683, "step": 193250 }, { "epoch": 1.165567226637402, "grad_norm": 1.410030484199524, "learning_rate": 0.00028044639359729566, "loss": 5.0837, "step": 193300 }, { "epoch": 1.165868718418736, "grad_norm": 1.5527492761611938, "learning_rate": 0.0002804363464834072, "loss": 4.9808, "step": 193350 }, { "epoch": 1.16617021020007, "grad_norm": 1.4526965618133545, "learning_rate": 0.00028042629696902393, "loss": 4.9981, "step": 193400 }, { "epoch": 1.166471701981404, "grad_norm": 1.5553522109985352, "learning_rate": 0.00028041624505433074, "loss": 5.3266, "step": 193450 }, { "epoch": 1.166773193762738, "grad_norm": 1.0575780868530273, "learning_rate": 0.00028040619073951275, "loss": 5.4096, "step": 193500 }, { "epoch": 1.1670746855440721, "grad_norm": 1.1282066106796265, "learning_rate": 0.00028039613402475487, "loss": 5.0643, "step": 193550 }, { "epoch": 1.167376177325406, "grad_norm": 1.1852718591690063, "learning_rate": 0.00028038607491024227, "loss": 5.0099, "step": 193600 }, { "epoch": 1.16767766910674, "grad_norm": 1.4252620935440063, "learning_rate": 0.00028037601339616, "loss": 5.3294, "step": 193650 }, { "epoch": 1.1679791608880743, "grad_norm": 1.1362558603286743, "learning_rate": 0.00028036594948269324, "loss": 4.9782, "step": 193700 }, { "epoch": 1.1682806526694083, "grad_norm": 1.3032013177871704, "learning_rate": 0.00028035588317002727, "loss": 4.8962, "step": 193750 }, { "epoch": 1.1685821444507423, "grad_norm": 1.5048174858093262, "learning_rate": 0.00028034581445834727, "loss": 4.9176, "step": 193800 }, { "epoch": 1.1688836362320763, "grad_norm": 1.8084118366241455, "learning_rate": 0.00028033574334783857, "loss": 5.4022, "step": 193850 }, { "epoch": 1.1691851280134102, "grad_norm": 2.2829172611236572, "learning_rate": 0.0002803256698386865, "loss": 5.2897, "step": 193900 }, { "epoch": 1.1694866197947444, "grad_norm": 1.2585618495941162, "learning_rate": 0.0002803155939310765, "loss": 5.4615, "step": 193950 }, { "epoch": 1.1697881115760784, "grad_norm": 1.946536660194397, "learning_rate": 0.0002803055156251939, "loss": 5.1588, "step": 194000 }, { "epoch": 1.1700896033574124, "grad_norm": 1.3961400985717773, "learning_rate": 0.0002802954349212243, "loss": 5.4053, "step": 194050 }, { "epoch": 1.1703910951387466, "grad_norm": 2.2885489463806152, "learning_rate": 0.0002802853518193532, "loss": 4.6022, "step": 194100 }, { "epoch": 1.1706925869200806, "grad_norm": 1.0848360061645508, "learning_rate": 0.000280275266319766, "loss": 5.1418, "step": 194150 }, { "epoch": 1.1709940787014146, "grad_norm": 1.1315810680389404, "learning_rate": 0.00028026517842264854, "loss": 5.1325, "step": 194200 }, { "epoch": 1.1712955704827486, "grad_norm": 1.7309868335723877, "learning_rate": 0.0002802550881281863, "loss": 5.3259, "step": 194250 }, { "epoch": 1.1715970622640828, "grad_norm": 1.1483211517333984, "learning_rate": 0.0002802449954365651, "loss": 5.0862, "step": 194300 }, { "epoch": 1.1718985540454168, "grad_norm": 1.1199297904968262, "learning_rate": 0.00028023490034797063, "loss": 5.3148, "step": 194350 }, { "epoch": 1.1722000458267507, "grad_norm": 2.0570428371429443, "learning_rate": 0.0002802248028625887, "loss": 5.3709, "step": 194400 }, { "epoch": 1.1725015376080847, "grad_norm": 1.1893017292022705, "learning_rate": 0.000280214702980605, "loss": 5.2272, "step": 194450 }, { "epoch": 1.172803029389419, "grad_norm": 1.2014883756637573, "learning_rate": 0.00028020460070220563, "loss": 5.298, "step": 194500 }, { "epoch": 1.173104521170753, "grad_norm": 0.6895910501480103, "learning_rate": 0.0002801944960275763, "loss": 5.3449, "step": 194550 }, { "epoch": 1.173406012952087, "grad_norm": 1.1374634504318237, "learning_rate": 0.0002801843889569032, "loss": 5.0398, "step": 194600 }, { "epoch": 1.173707504733421, "grad_norm": 1.986419916152954, "learning_rate": 0.0002801742794903721, "loss": 5.297, "step": 194650 }, { "epoch": 1.174008996514755, "grad_norm": 1.3066449165344238, "learning_rate": 0.00028016416762816917, "loss": 4.9398, "step": 194700 }, { "epoch": 1.174310488296089, "grad_norm": 1.6208709478378296, "learning_rate": 0.00028015405337048047, "loss": 4.8698, "step": 194750 }, { "epoch": 1.174611980077423, "grad_norm": 1.2638306617736816, "learning_rate": 0.0002801439367174922, "loss": 5.2609, "step": 194800 }, { "epoch": 1.174913471858757, "grad_norm": 2.1589434146881104, "learning_rate": 0.00028013381766939054, "loss": 4.8307, "step": 194850 }, { "epoch": 1.1752149636400913, "grad_norm": 0.3616599142551422, "learning_rate": 0.0002801236962263616, "loss": 5.1948, "step": 194900 }, { "epoch": 1.1755164554214252, "grad_norm": 0.9476787447929382, "learning_rate": 0.0002801135723885918, "loss": 5.1843, "step": 194950 }, { "epoch": 1.1758179472027592, "grad_norm": 1.7585182189941406, "learning_rate": 0.0002801034461562673, "loss": 5.4612, "step": 195000 }, { "epoch": 1.1758179472027592, "eval_loss": 5.636074066162109, "eval_runtime": 39.1151, "eval_samples_per_second": 13.09, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.140136356835193, "step": 195000 }, { "epoch": 1.1761194389840932, "grad_norm": 1.2101829051971436, "learning_rate": 0.00028009331752957455, "loss": 4.6804, "step": 195050 }, { "epoch": 1.1764209307654274, "grad_norm": 1.097802996635437, "learning_rate": 0.0002800831865087, "loss": 5.3439, "step": 195100 }, { "epoch": 1.1767224225467614, "grad_norm": 1.544421672821045, "learning_rate": 0.00028007305309383, "loss": 5.4601, "step": 195150 }, { "epoch": 1.1770239143280954, "grad_norm": 1.1253108978271484, "learning_rate": 0.0002800629172851511, "loss": 5.4742, "step": 195200 }, { "epoch": 1.1773254061094294, "grad_norm": 1.2356271743774414, "learning_rate": 0.0002800527790828498, "loss": 5.2271, "step": 195250 }, { "epoch": 1.1776268978907636, "grad_norm": 1.7756986618041992, "learning_rate": 0.0002800426384871127, "loss": 5.2364, "step": 195300 }, { "epoch": 1.1779283896720976, "grad_norm": 1.1348810195922852, "learning_rate": 0.00028003249549812646, "loss": 4.6747, "step": 195350 }, { "epoch": 1.1782298814534315, "grad_norm": 1.4483606815338135, "learning_rate": 0.0002800223501160777, "loss": 5.1857, "step": 195400 }, { "epoch": 1.1785313732347655, "grad_norm": 0.8468074798583984, "learning_rate": 0.00028001220234115306, "loss": 5.155, "step": 195450 }, { "epoch": 1.1788328650160997, "grad_norm": 1.24000084400177, "learning_rate": 0.0002800020521735394, "loss": 5.163, "step": 195500 }, { "epoch": 1.1791343567974337, "grad_norm": 1.2273448705673218, "learning_rate": 0.0002799918996134235, "loss": 5.373, "step": 195550 }, { "epoch": 1.1794358485787677, "grad_norm": 2.1052660942077637, "learning_rate": 0.00027998174466099227, "loss": 4.7542, "step": 195600 }, { "epoch": 1.1797373403601017, "grad_norm": 0.7283917665481567, "learning_rate": 0.0002799715873164324, "loss": 5.1744, "step": 195650 }, { "epoch": 1.180038832141436, "grad_norm": 0.784494161605835, "learning_rate": 0.00027996142757993105, "loss": 5.2804, "step": 195700 }, { "epoch": 1.1803403239227699, "grad_norm": 2.7362513542175293, "learning_rate": 0.0002799512654516751, "loss": 5.1573, "step": 195750 }, { "epoch": 1.1806418157041039, "grad_norm": 1.8545078039169312, "learning_rate": 0.0002799411009318515, "loss": 5.1702, "step": 195800 }, { "epoch": 1.1809433074854379, "grad_norm": 1.5015488862991333, "learning_rate": 0.00027993093402064735, "loss": 5.001, "step": 195850 }, { "epoch": 1.181244799266772, "grad_norm": 0.6321857571601868, "learning_rate": 0.0002799207647182498, "loss": 5.3378, "step": 195900 }, { "epoch": 1.181546291048106, "grad_norm": 1.5920709371566772, "learning_rate": 0.000279910593024846, "loss": 4.9448, "step": 195950 }, { "epoch": 1.18184778282944, "grad_norm": 1.4199192523956299, "learning_rate": 0.00027990041894062316, "loss": 5.0288, "step": 196000 }, { "epoch": 1.182149274610774, "grad_norm": 1.2777657508850098, "learning_rate": 0.00027989024246576847, "loss": 5.4695, "step": 196050 }, { "epoch": 1.1824507663921082, "grad_norm": 1.186979055404663, "learning_rate": 0.00027988006360046917, "loss": 5.5439, "step": 196100 }, { "epoch": 1.1827522581734422, "grad_norm": 1.7433035373687744, "learning_rate": 0.0002798698823449127, "loss": 4.9537, "step": 196150 }, { "epoch": 1.1830537499547762, "grad_norm": 1.118714451789856, "learning_rate": 0.00027985969869928634, "loss": 5.009, "step": 196200 }, { "epoch": 1.1833552417361104, "grad_norm": 1.46890127658844, "learning_rate": 0.0002798495126637776, "loss": 5.6131, "step": 196250 }, { "epoch": 1.1836567335174444, "grad_norm": 1.032270908355713, "learning_rate": 0.00027983932423857386, "loss": 5.1733, "step": 196300 }, { "epoch": 1.1839582252987784, "grad_norm": 1.0815178155899048, "learning_rate": 0.00027982913342386266, "loss": 4.8241, "step": 196350 }, { "epoch": 1.1842597170801124, "grad_norm": 1.3503048419952393, "learning_rate": 0.0002798189402198315, "loss": 4.9849, "step": 196400 }, { "epoch": 1.1845612088614463, "grad_norm": 1.100692629814148, "learning_rate": 0.00027980874462666804, "loss": 5.0298, "step": 196450 }, { "epoch": 1.1848627006427805, "grad_norm": 1.5114941596984863, "learning_rate": 0.00027979854664455985, "loss": 5.0113, "step": 196500 }, { "epoch": 1.1851641924241145, "grad_norm": 1.1497571468353271, "learning_rate": 0.0002797883462736947, "loss": 4.8069, "step": 196550 }, { "epoch": 1.1854656842054485, "grad_norm": 1.3961387872695923, "learning_rate": 0.0002797781435142602, "loss": 4.8956, "step": 196600 }, { "epoch": 1.1857671759867827, "grad_norm": 1.2420998811721802, "learning_rate": 0.0002797679383664442, "loss": 5.1948, "step": 196650 }, { "epoch": 1.1860686677681167, "grad_norm": 1.0916370153427124, "learning_rate": 0.00027975773083043447, "loss": 5.2083, "step": 196700 }, { "epoch": 1.1863701595494507, "grad_norm": 1.1776378154754639, "learning_rate": 0.0002797475209064189, "loss": 5.6113, "step": 196750 }, { "epoch": 1.1866716513307847, "grad_norm": 1.2928545475006104, "learning_rate": 0.00027973730859458533, "loss": 4.9307, "step": 196800 }, { "epoch": 1.1869731431121187, "grad_norm": 1.120017647743225, "learning_rate": 0.0002797270938951218, "loss": 4.7793, "step": 196850 }, { "epoch": 1.1872746348934529, "grad_norm": 1.465498924255371, "learning_rate": 0.0002797168768082162, "loss": 5.1887, "step": 196900 }, { "epoch": 1.1875761266747868, "grad_norm": 1.4094030857086182, "learning_rate": 0.00027970665733405655, "loss": 5.1921, "step": 196950 }, { "epoch": 1.1878776184561208, "grad_norm": 0.6548944711685181, "learning_rate": 0.00027969643547283104, "loss": 5.5917, "step": 197000 }, { "epoch": 1.188179110237455, "grad_norm": 0.42463505268096924, "learning_rate": 0.0002796862112247277, "loss": 5.2146, "step": 197050 }, { "epoch": 1.188480602018789, "grad_norm": 1.5307974815368652, "learning_rate": 0.0002796759845899347, "loss": 5.1587, "step": 197100 }, { "epoch": 1.188782093800123, "grad_norm": 0.8674759864807129, "learning_rate": 0.00027966575556864033, "loss": 5.0674, "step": 197150 }, { "epoch": 1.189083585581457, "grad_norm": 1.433497667312622, "learning_rate": 0.0002796555241610327, "loss": 5.0793, "step": 197200 }, { "epoch": 1.189385077362791, "grad_norm": 1.4702426195144653, "learning_rate": 0.0002796452903673002, "loss": 5.3908, "step": 197250 }, { "epoch": 1.1896865691441252, "grad_norm": 1.3756134510040283, "learning_rate": 0.0002796350541876311, "loss": 5.2017, "step": 197300 }, { "epoch": 1.1899880609254592, "grad_norm": 1.1879849433898926, "learning_rate": 0.0002796248156222139, "loss": 5.1969, "step": 197350 }, { "epoch": 1.1902895527067932, "grad_norm": 1.143218994140625, "learning_rate": 0.0002796145746712369, "loss": 5.5052, "step": 197400 }, { "epoch": 1.1905910444881274, "grad_norm": 0.8408507704734802, "learning_rate": 0.00027960433133488865, "loss": 4.4924, "step": 197450 }, { "epoch": 1.1908925362694613, "grad_norm": 1.73350191116333, "learning_rate": 0.00027959408561335764, "loss": 5.3182, "step": 197500 }, { "epoch": 1.1911940280507953, "grad_norm": 2.6623122692108154, "learning_rate": 0.0002795838375068324, "loss": 4.9216, "step": 197550 }, { "epoch": 1.1914955198321293, "grad_norm": 1.6584265232086182, "learning_rate": 0.0002795735870155016, "loss": 4.7865, "step": 197600 }, { "epoch": 1.1917970116134633, "grad_norm": 1.4752849340438843, "learning_rate": 0.0002795633341395538, "loss": 5.3082, "step": 197650 }, { "epoch": 1.1920985033947975, "grad_norm": 1.6600366830825806, "learning_rate": 0.00027955307887917775, "loss": 5.2221, "step": 197700 }, { "epoch": 1.1923999951761315, "grad_norm": 1.4173938035964966, "learning_rate": 0.00027954282123456215, "loss": 5.0243, "step": 197750 }, { "epoch": 1.1927014869574655, "grad_norm": 0.9663242697715759, "learning_rate": 0.0002795325612058958, "loss": 5.2261, "step": 197800 }, { "epoch": 1.1930029787387997, "grad_norm": 1.49470055103302, "learning_rate": 0.00027952229879336753, "loss": 5.173, "step": 197850 }, { "epoch": 1.1933044705201337, "grad_norm": 1.2400941848754883, "learning_rate": 0.0002795120339971662, "loss": 5.1764, "step": 197900 }, { "epoch": 1.1936059623014676, "grad_norm": 1.0468271970748901, "learning_rate": 0.00027950176681748064, "loss": 4.9273, "step": 197950 }, { "epoch": 1.1939074540828016, "grad_norm": 1.067435383796692, "learning_rate": 0.00027949149725449993, "loss": 5.2727, "step": 198000 }, { "epoch": 1.1939074540828016, "eval_loss": 5.613858222961426, "eval_runtime": 38.9162, "eval_samples_per_second": 13.156, "eval_steps_per_second": 6.578, "eval_tts_loss": 7.273023629903954, "step": 198000 }, { "epoch": 1.1942089458641358, "grad_norm": 1.945086121559143, "learning_rate": 0.000279481225308413, "loss": 5.1362, "step": 198050 }, { "epoch": 1.1945104376454698, "grad_norm": 0.8980103135108948, "learning_rate": 0.00027947095097940886, "loss": 5.2568, "step": 198100 }, { "epoch": 1.1948119294268038, "grad_norm": 1.2401330471038818, "learning_rate": 0.00027946067426767667, "loss": 5.3011, "step": 198150 }, { "epoch": 1.1951134212081378, "grad_norm": 1.3299596309661865, "learning_rate": 0.0002794503951734055, "loss": 5.1844, "step": 198200 }, { "epoch": 1.195414912989472, "grad_norm": 1.1612519025802612, "learning_rate": 0.00027944011369678457, "loss": 5.112, "step": 198250 }, { "epoch": 1.195716404770806, "grad_norm": 1.1140722036361694, "learning_rate": 0.0002794298298380031, "loss": 5.2156, "step": 198300 }, { "epoch": 1.19601789655214, "grad_norm": 1.1555149555206299, "learning_rate": 0.00027941954359725026, "loss": 5.2328, "step": 198350 }, { "epoch": 1.196319388333474, "grad_norm": 0.33667948842048645, "learning_rate": 0.0002794092549747154, "loss": 5.3192, "step": 198400 }, { "epoch": 1.1966208801148082, "grad_norm": 1.4352713823318481, "learning_rate": 0.0002793989639705879, "loss": 4.6894, "step": 198450 }, { "epoch": 1.1969223718961421, "grad_norm": 1.2449672222137451, "learning_rate": 0.00027938867058505714, "loss": 4.9183, "step": 198500 }, { "epoch": 1.1972238636774761, "grad_norm": 1.2970243692398071, "learning_rate": 0.00027937837481831256, "loss": 5.4672, "step": 198550 }, { "epoch": 1.1975253554588101, "grad_norm": 1.256252646446228, "learning_rate": 0.00027936807667054356, "loss": 5.2011, "step": 198600 }, { "epoch": 1.1978268472401443, "grad_norm": 0.9879198670387268, "learning_rate": 0.00027935777614193977, "loss": 5.4163, "step": 198650 }, { "epoch": 1.1981283390214783, "grad_norm": 1.6730989217758179, "learning_rate": 0.0002793474732326907, "loss": 4.8044, "step": 198700 }, { "epoch": 1.1984298308028123, "grad_norm": 0.8514532446861267, "learning_rate": 0.000279337167942986, "loss": 4.6771, "step": 198750 }, { "epoch": 1.1987313225841463, "grad_norm": 0.5652899742126465, "learning_rate": 0.0002793268602730153, "loss": 5.4142, "step": 198800 }, { "epoch": 1.1990328143654805, "grad_norm": 0.5966469049453735, "learning_rate": 0.0002793165502229683, "loss": 5.3761, "step": 198850 }, { "epoch": 1.1993343061468145, "grad_norm": 1.1196991205215454, "learning_rate": 0.0002793062377930347, "loss": 5.2802, "step": 198900 }, { "epoch": 1.1996357979281485, "grad_norm": 1.2251888513565063, "learning_rate": 0.0002792959229834043, "loss": 5.3988, "step": 198950 }, { "epoch": 1.1999372897094824, "grad_norm": 1.4194010496139526, "learning_rate": 0.00027928560579426707, "loss": 5.5377, "step": 199000 }, { "epoch": 1.2002387814908166, "grad_norm": 0.29135531187057495, "learning_rate": 0.0002792752862258127, "loss": 4.8349, "step": 199050 }, { "epoch": 1.2005402732721506, "grad_norm": 1.2602856159210205, "learning_rate": 0.00027926496427823114, "loss": 4.985, "step": 199100 }, { "epoch": 1.2008417650534846, "grad_norm": 1.1077449321746826, "learning_rate": 0.0002792546399517124, "loss": 4.6643, "step": 199150 }, { "epoch": 1.2011432568348186, "grad_norm": 1.5963178873062134, "learning_rate": 0.00027924431324644647, "loss": 5.0432, "step": 199200 }, { "epoch": 1.2014447486161528, "grad_norm": 1.5043659210205078, "learning_rate": 0.0002792339841626234, "loss": 5.3686, "step": 199250 }, { "epoch": 1.2017462403974868, "grad_norm": 1.620086431503296, "learning_rate": 0.0002792236527004333, "loss": 4.9291, "step": 199300 }, { "epoch": 1.2020477321788208, "grad_norm": 1.0112261772155762, "learning_rate": 0.0002792133188600662, "loss": 4.9301, "step": 199350 }, { "epoch": 1.2023492239601548, "grad_norm": 1.0584102869033813, "learning_rate": 0.0002792029826417125, "loss": 4.7153, "step": 199400 }, { "epoch": 1.202650715741489, "grad_norm": 1.5122658014297485, "learning_rate": 0.00027919264404556215, "loss": 5.5316, "step": 199450 }, { "epoch": 1.202952207522823, "grad_norm": 1.3798387050628662, "learning_rate": 0.00027918230307180566, "loss": 5.007, "step": 199500 }, { "epoch": 1.203253699304157, "grad_norm": 1.000451922416687, "learning_rate": 0.0002791719597206332, "loss": 5.1836, "step": 199550 }, { "epoch": 1.203555191085491, "grad_norm": 1.3626281023025513, "learning_rate": 0.00027916161399223514, "loss": 5.2841, "step": 199600 }, { "epoch": 1.2038566828668251, "grad_norm": 1.651572585105896, "learning_rate": 0.00027915126588680194, "loss": 4.8787, "step": 199650 }, { "epoch": 1.204158174648159, "grad_norm": 2.070716619491577, "learning_rate": 0.0002791409154045239, "loss": 4.7548, "step": 199700 }, { "epoch": 1.204459666429493, "grad_norm": 0.6999363303184509, "learning_rate": 0.0002791305625455917, "loss": 5.2275, "step": 199750 }, { "epoch": 1.204761158210827, "grad_norm": 0.5373135805130005, "learning_rate": 0.00027912020731019575, "loss": 4.8093, "step": 199800 }, { "epoch": 1.2050626499921613, "grad_norm": 1.3968935012817383, "learning_rate": 0.0002791098496985267, "loss": 4.5865, "step": 199850 }, { "epoch": 1.2053641417734953, "grad_norm": 1.2249414920806885, "learning_rate": 0.00027909948971077507, "loss": 5.4357, "step": 199900 }, { "epoch": 1.2056656335548293, "grad_norm": 1.2183185815811157, "learning_rate": 0.00027908912734713154, "loss": 5.5646, "step": 199950 }, { "epoch": 1.2059671253361635, "grad_norm": 0.7852365374565125, "learning_rate": 0.00027907876260778686, "loss": 4.9328, "step": 200000 }, { "epoch": 1.2062686171174974, "grad_norm": 1.028454303741455, "learning_rate": 0.0002790683954929318, "loss": 5.2428, "step": 200050 }, { "epoch": 1.2065701088988314, "grad_norm": 1.7258104085922241, "learning_rate": 0.0002790580260027571, "loss": 5.3429, "step": 200100 }, { "epoch": 1.2068716006801654, "grad_norm": 1.1054739952087402, "learning_rate": 0.00027904765413745355, "loss": 5.4272, "step": 200150 }, { "epoch": 1.2071730924614994, "grad_norm": 1.2874600887298584, "learning_rate": 0.00027903727989721206, "loss": 5.2792, "step": 200200 }, { "epoch": 1.2074745842428336, "grad_norm": 1.5879977941513062, "learning_rate": 0.0002790269032822237, "loss": 5.5049, "step": 200250 }, { "epoch": 1.2077760760241676, "grad_norm": 1.5369139909744263, "learning_rate": 0.00027901652429267924, "loss": 4.8896, "step": 200300 }, { "epoch": 1.2080775678055016, "grad_norm": 1.174109697341919, "learning_rate": 0.0002790061429287697, "loss": 4.796, "step": 200350 }, { "epoch": 1.2083790595868358, "grad_norm": 1.064683437347412, "learning_rate": 0.00027899575919068637, "loss": 5.1292, "step": 200400 }, { "epoch": 1.2086805513681698, "grad_norm": 1.4356403350830078, "learning_rate": 0.00027898537307862007, "loss": 4.9046, "step": 200450 }, { "epoch": 1.2089820431495037, "grad_norm": 0.6487130522727966, "learning_rate": 0.0002789749845927621, "loss": 4.9436, "step": 200500 }, { "epoch": 1.2092835349308377, "grad_norm": 1.5606940984725952, "learning_rate": 0.00027896459373330355, "loss": 4.7756, "step": 200550 }, { "epoch": 1.2095850267121717, "grad_norm": 1.0849956274032593, "learning_rate": 0.0002789542005004357, "loss": 4.9753, "step": 200600 }, { "epoch": 1.209886518493506, "grad_norm": 1.3975332975387573, "learning_rate": 0.0002789438048943499, "loss": 5.4066, "step": 200650 }, { "epoch": 1.21018801027484, "grad_norm": 0.517184853553772, "learning_rate": 0.00027893340691523727, "loss": 5.0492, "step": 200700 }, { "epoch": 1.210489502056174, "grad_norm": 1.2259601354599, "learning_rate": 0.00027892300656328934, "loss": 4.7455, "step": 200750 }, { "epoch": 1.210790993837508, "grad_norm": 1.7443584203720093, "learning_rate": 0.00027891260383869744, "loss": 5.4498, "step": 200800 }, { "epoch": 1.211092485618842, "grad_norm": 1.5777207612991333, "learning_rate": 0.00027890219874165305, "loss": 5.1943, "step": 200850 }, { "epoch": 1.211393977400176, "grad_norm": 1.6973271369934082, "learning_rate": 0.00027889179127234764, "loss": 5.2302, "step": 200900 }, { "epoch": 1.21169546918151, "grad_norm": 0.6573458909988403, "learning_rate": 0.0002788813814309727, "loss": 5.2444, "step": 200950 }, { "epoch": 1.211996960962844, "grad_norm": 1.316014051437378, "learning_rate": 0.00027887096921771993, "loss": 5.1381, "step": 201000 }, { "epoch": 1.211996960962844, "eval_loss": 5.610567092895508, "eval_runtime": 38.8939, "eval_samples_per_second": 13.164, "eval_steps_per_second": 6.582, "eval_tts_loss": 7.164809658727491, "step": 201000 }, { "epoch": 1.2122984527441782, "grad_norm": 0.4939590394496918, "learning_rate": 0.00027886055463278086, "loss": 5.2794, "step": 201050 }, { "epoch": 1.2125999445255122, "grad_norm": 0.351769357919693, "learning_rate": 0.0002788501376763472, "loss": 5.1556, "step": 201100 }, { "epoch": 1.2129014363068462, "grad_norm": 1.7360363006591797, "learning_rate": 0.00027883971834861054, "loss": 4.8585, "step": 201150 }, { "epoch": 1.2132029280881804, "grad_norm": 1.1009595394134521, "learning_rate": 0.00027882929664976283, "loss": 5.0476, "step": 201200 }, { "epoch": 1.2135044198695144, "grad_norm": 1.334452509880066, "learning_rate": 0.00027881887257999574, "loss": 5.0101, "step": 201250 }, { "epoch": 1.2138059116508484, "grad_norm": 0.39953649044036865, "learning_rate": 0.0002788084461395011, "loss": 4.9757, "step": 201300 }, { "epoch": 1.2141074034321824, "grad_norm": 1.5614832639694214, "learning_rate": 0.0002787980173284709, "loss": 4.743, "step": 201350 }, { "epoch": 1.2144088952135164, "grad_norm": 1.9400019645690918, "learning_rate": 0.0002787875861470969, "loss": 5.7431, "step": 201400 }, { "epoch": 1.2147103869948506, "grad_norm": 1.1605840921401978, "learning_rate": 0.00027877715259557123, "loss": 4.5586, "step": 201450 }, { "epoch": 1.2150118787761846, "grad_norm": 1.3811156749725342, "learning_rate": 0.0002787667166740858, "loss": 4.9918, "step": 201500 }, { "epoch": 1.2153133705575185, "grad_norm": 1.3645176887512207, "learning_rate": 0.00027875627838283275, "loss": 5.1326, "step": 201550 }, { "epoch": 1.2156148623388527, "grad_norm": 1.1409026384353638, "learning_rate": 0.0002787458377220041, "loss": 5.4958, "step": 201600 }, { "epoch": 1.2159163541201867, "grad_norm": 1.1713063716888428, "learning_rate": 0.00027873539469179206, "loss": 5.1126, "step": 201650 }, { "epoch": 1.2162178459015207, "grad_norm": 1.1002086400985718, "learning_rate": 0.00027872494929238883, "loss": 5.1673, "step": 201700 }, { "epoch": 1.2165193376828547, "grad_norm": 1.5463171005249023, "learning_rate": 0.00027871450152398657, "loss": 5.4836, "step": 201750 }, { "epoch": 1.216820829464189, "grad_norm": 1.2755597829818726, "learning_rate": 0.0002787040513867776, "loss": 5.3164, "step": 201800 }, { "epoch": 1.2171223212455229, "grad_norm": 1.073371410369873, "learning_rate": 0.00027869359888095426, "loss": 5.5443, "step": 201850 }, { "epoch": 1.2174238130268569, "grad_norm": 1.1248096227645874, "learning_rate": 0.00027868314400670885, "loss": 4.9934, "step": 201900 }, { "epoch": 1.2177253048081909, "grad_norm": 0.38973554968833923, "learning_rate": 0.0002786726867642338, "loss": 5.3615, "step": 201950 }, { "epoch": 1.218026796589525, "grad_norm": 0.7290513515472412, "learning_rate": 0.00027866222715372166, "loss": 4.9335, "step": 202000 }, { "epoch": 1.218328288370859, "grad_norm": 1.6411811113357544, "learning_rate": 0.00027865176517536473, "loss": 5.2235, "step": 202050 }, { "epoch": 1.218629780152193, "grad_norm": 1.132874846458435, "learning_rate": 0.0002786413008293557, "loss": 5.027, "step": 202100 }, { "epoch": 1.218931271933527, "grad_norm": 1.1892305612564087, "learning_rate": 0.0002786308341158872, "loss": 5.1978, "step": 202150 }, { "epoch": 1.2192327637148612, "grad_norm": 1.653830647468567, "learning_rate": 0.0002786203650351516, "loss": 4.8499, "step": 202200 }, { "epoch": 1.2195342554961952, "grad_norm": 1.5322942733764648, "learning_rate": 0.00027860989358734185, "loss": 4.8137, "step": 202250 }, { "epoch": 1.2198357472775292, "grad_norm": 0.9737940430641174, "learning_rate": 0.00027859941977265054, "loss": 5.1919, "step": 202300 }, { "epoch": 1.2201372390588632, "grad_norm": 2.180877208709717, "learning_rate": 0.00027858894359127036, "loss": 4.8786, "step": 202350 }, { "epoch": 1.2204387308401974, "grad_norm": 1.4726362228393555, "learning_rate": 0.00027857846504339423, "loss": 4.9691, "step": 202400 }, { "epoch": 1.2207402226215314, "grad_norm": 1.6977293491363525, "learning_rate": 0.000278567984129215, "loss": 5.3128, "step": 202450 }, { "epoch": 1.2210417144028654, "grad_norm": 1.127580165863037, "learning_rate": 0.0002785575008489254, "loss": 5.4103, "step": 202500 }, { "epoch": 1.2213432061841993, "grad_norm": 2.0822813510894775, "learning_rate": 0.00027854701520271845, "loss": 5.6477, "step": 202550 }, { "epoch": 1.2216446979655335, "grad_norm": 1.669156551361084, "learning_rate": 0.0002785365271907872, "loss": 5.3548, "step": 202600 }, { "epoch": 1.2219461897468675, "grad_norm": 1.7291301488876343, "learning_rate": 0.00027852603681332455, "loss": 4.9879, "step": 202650 }, { "epoch": 1.2222476815282015, "grad_norm": 1.640163540840149, "learning_rate": 0.00027851554407052367, "loss": 5.1319, "step": 202700 }, { "epoch": 1.2225491733095355, "grad_norm": 1.5263394117355347, "learning_rate": 0.00027850504896257755, "loss": 5.4467, "step": 202750 }, { "epoch": 1.2228506650908697, "grad_norm": 1.5231869220733643, "learning_rate": 0.00027849455148967945, "loss": 4.9019, "step": 202800 }, { "epoch": 1.2231521568722037, "grad_norm": 2.495295763015747, "learning_rate": 0.0002784840516520224, "loss": 5.1208, "step": 202850 }, { "epoch": 1.2234536486535377, "grad_norm": 1.1975852251052856, "learning_rate": 0.00027847354944979984, "loss": 5.56, "step": 202900 }, { "epoch": 1.2237551404348717, "grad_norm": 1.4176875352859497, "learning_rate": 0.0002784630448832049, "loss": 5.3476, "step": 202950 }, { "epoch": 1.2240566322162059, "grad_norm": 1.0174578428268433, "learning_rate": 0.0002784525379524309, "loss": 5.2163, "step": 203000 }, { "epoch": 1.2243581239975398, "grad_norm": 1.3756146430969238, "learning_rate": 0.00027844202865767134, "loss": 5.2212, "step": 203050 }, { "epoch": 1.2246596157788738, "grad_norm": 2.163670778274536, "learning_rate": 0.00027843151699911945, "loss": 5.2855, "step": 203100 }, { "epoch": 1.2249611075602078, "grad_norm": 1.0338332653045654, "learning_rate": 0.0002784210029769688, "loss": 4.8016, "step": 203150 }, { "epoch": 1.225262599341542, "grad_norm": 1.306002140045166, "learning_rate": 0.00027841048659141293, "loss": 5.5731, "step": 203200 }, { "epoch": 1.225564091122876, "grad_norm": 2.0822911262512207, "learning_rate": 0.00027839996784264525, "loss": 4.8781, "step": 203250 }, { "epoch": 1.22586558290421, "grad_norm": 0.8561975955963135, "learning_rate": 0.00027838944673085937, "loss": 4.971, "step": 203300 }, { "epoch": 1.226167074685544, "grad_norm": 2.075460910797119, "learning_rate": 0.000278378923256249, "loss": 4.8347, "step": 203350 }, { "epoch": 1.2264685664668782, "grad_norm": 1.5368574857711792, "learning_rate": 0.0002783683974190077, "loss": 5.1339, "step": 203400 }, { "epoch": 1.2267700582482122, "grad_norm": 0.2927395701408386, "learning_rate": 0.00027835786921932924, "loss": 4.9104, "step": 203450 }, { "epoch": 1.2270715500295462, "grad_norm": 2.024500608444214, "learning_rate": 0.00027834733865740737, "loss": 5.007, "step": 203500 }, { "epoch": 1.2273730418108801, "grad_norm": 1.9947818517684937, "learning_rate": 0.00027833680573343593, "loss": 5.2535, "step": 203550 }, { "epoch": 1.2276745335922143, "grad_norm": 1.3243554830551147, "learning_rate": 0.0002783262704476087, "loss": 5.2401, "step": 203600 }, { "epoch": 1.2279760253735483, "grad_norm": 1.7350068092346191, "learning_rate": 0.00027831573280011964, "loss": 4.6277, "step": 203650 }, { "epoch": 1.2282775171548823, "grad_norm": 1.3964577913284302, "learning_rate": 0.0002783051927911626, "loss": 5.298, "step": 203700 }, { "epoch": 1.2285790089362165, "grad_norm": 1.8915940523147583, "learning_rate": 0.0002782946504209316, "loss": 4.9608, "step": 203750 }, { "epoch": 1.2288805007175505, "grad_norm": 0.5730720162391663, "learning_rate": 0.00027828410568962064, "loss": 5.2216, "step": 203800 }, { "epoch": 1.2291819924988845, "grad_norm": 1.5411735773086548, "learning_rate": 0.0002782735585974238, "loss": 5.4995, "step": 203850 }, { "epoch": 1.2294834842802185, "grad_norm": 0.9780676364898682, "learning_rate": 0.0002782630091445351, "loss": 4.9094, "step": 203900 }, { "epoch": 1.2297849760615525, "grad_norm": 1.0904004573822021, "learning_rate": 0.00027825245733114885, "loss": 5.0252, "step": 203950 }, { "epoch": 1.2300864678428867, "grad_norm": 0.39180076122283936, "learning_rate": 0.00027824190315745907, "loss": 4.9938, "step": 204000 }, { "epoch": 1.2300864678428867, "eval_loss": 5.600615978240967, "eval_runtime": 38.875, "eval_samples_per_second": 13.17, "eval_steps_per_second": 6.585, "eval_tts_loss": 7.1502071409291705, "step": 204000 }, { "epoch": 1.2303879596242207, "grad_norm": 1.5331178903579712, "learning_rate": 0.00027823134662366015, "loss": 5.259, "step": 204050 }, { "epoch": 1.2306894514055546, "grad_norm": 1.1109789609909058, "learning_rate": 0.00027822078772994624, "loss": 5.3751, "step": 204100 }, { "epoch": 1.2309909431868888, "grad_norm": 1.1629517078399658, "learning_rate": 0.00027821022647651173, "loss": 5.0605, "step": 204150 }, { "epoch": 1.2312924349682228, "grad_norm": 1.4277786016464233, "learning_rate": 0.00027819966286355097, "loss": 5.27, "step": 204200 }, { "epoch": 1.2315939267495568, "grad_norm": 1.5083136558532715, "learning_rate": 0.00027818909689125834, "loss": 5.0548, "step": 204250 }, { "epoch": 1.2318954185308908, "grad_norm": 1.2304859161376953, "learning_rate": 0.00027817852855982836, "loss": 4.9399, "step": 204300 }, { "epoch": 1.2321969103122248, "grad_norm": 1.1741873025894165, "learning_rate": 0.0002781679578694555, "loss": 5.1329, "step": 204350 }, { "epoch": 1.232498402093559, "grad_norm": 2.5969269275665283, "learning_rate": 0.00027815738482033414, "loss": 5.3069, "step": 204400 }, { "epoch": 1.232799893874893, "grad_norm": 1.2605071067810059, "learning_rate": 0.00027814680941265913, "loss": 5.4668, "step": 204450 }, { "epoch": 1.233101385656227, "grad_norm": 1.6773204803466797, "learning_rate": 0.00027813623164662496, "loss": 4.8386, "step": 204500 }, { "epoch": 1.2334028774375612, "grad_norm": 1.2901933193206787, "learning_rate": 0.0002781256515224263, "loss": 5.5563, "step": 204550 }, { "epoch": 1.2337043692188951, "grad_norm": 0.5682030916213989, "learning_rate": 0.0002781150690402579, "loss": 4.8739, "step": 204600 }, { "epoch": 1.2340058610002291, "grad_norm": 1.3593738079071045, "learning_rate": 0.0002781044842003144, "loss": 5.1537, "step": 204650 }, { "epoch": 1.2343073527815631, "grad_norm": 1.4241583347320557, "learning_rate": 0.0002780938970027907, "loss": 5.5118, "step": 204700 }, { "epoch": 1.234608844562897, "grad_norm": 1.6107465028762817, "learning_rate": 0.00027808330744788165, "loss": 5.2396, "step": 204750 }, { "epoch": 1.2349103363442313, "grad_norm": 1.8389331102371216, "learning_rate": 0.0002780727155357821, "loss": 5.4035, "step": 204800 }, { "epoch": 1.2352118281255653, "grad_norm": 2.1054012775421143, "learning_rate": 0.00027806212126668705, "loss": 5.2921, "step": 204850 }, { "epoch": 1.2355133199068993, "grad_norm": 0.5866756439208984, "learning_rate": 0.00027805152464079134, "loss": 5.6652, "step": 204900 }, { "epoch": 1.2358148116882335, "grad_norm": 0.9524097442626953, "learning_rate": 0.0002780409256582901, "loss": 4.7316, "step": 204950 }, { "epoch": 1.2361163034695675, "grad_norm": 0.6687307953834534, "learning_rate": 0.0002780303243193783, "loss": 4.9344, "step": 205000 }, { "epoch": 1.2364177952509015, "grad_norm": 1.360164999961853, "learning_rate": 0.00027801972062425114, "loss": 5.0343, "step": 205050 }, { "epoch": 1.2367192870322354, "grad_norm": 0.9524866938591003, "learning_rate": 0.0002780091145731037, "loss": 4.8634, "step": 205100 }, { "epoch": 1.2370207788135694, "grad_norm": 0.7788587212562561, "learning_rate": 0.00027799850616613115, "loss": 4.8792, "step": 205150 }, { "epoch": 1.2373222705949036, "grad_norm": 1.1323885917663574, "learning_rate": 0.00027798789540352877, "loss": 5.0506, "step": 205200 }, { "epoch": 1.2376237623762376, "grad_norm": 1.5136306285858154, "learning_rate": 0.00027797728228549186, "loss": 5.678, "step": 205250 }, { "epoch": 1.2379252541575716, "grad_norm": 1.5983235836029053, "learning_rate": 0.0002779666668122157, "loss": 4.9643, "step": 205300 }, { "epoch": 1.2382267459389058, "grad_norm": 1.0107414722442627, "learning_rate": 0.0002779560489838956, "loss": 4.9655, "step": 205350 }, { "epoch": 1.2385282377202398, "grad_norm": 2.571505308151245, "learning_rate": 0.0002779454288007271, "loss": 5.3775, "step": 205400 }, { "epoch": 1.2388297295015738, "grad_norm": 1.597826361656189, "learning_rate": 0.00027793480626290553, "loss": 4.9921, "step": 205450 }, { "epoch": 1.2391312212829078, "grad_norm": 1.4351414442062378, "learning_rate": 0.0002779241813706264, "loss": 5.2429, "step": 205500 }, { "epoch": 1.239432713064242, "grad_norm": 1.4615594148635864, "learning_rate": 0.0002779135541240853, "loss": 5.0023, "step": 205550 }, { "epoch": 1.239734204845576, "grad_norm": 1.0470993518829346, "learning_rate": 0.0002779029245234778, "loss": 5.1228, "step": 205600 }, { "epoch": 1.24003569662691, "grad_norm": 1.16365385055542, "learning_rate": 0.0002778922925689995, "loss": 4.8613, "step": 205650 }, { "epoch": 1.240337188408244, "grad_norm": 1.1541776657104492, "learning_rate": 0.000277881658260846, "loss": 4.8764, "step": 205700 }, { "epoch": 1.2406386801895781, "grad_norm": 1.4874532222747803, "learning_rate": 0.0002778710215992131, "loss": 5.2537, "step": 205750 }, { "epoch": 1.240940171970912, "grad_norm": 1.52251398563385, "learning_rate": 0.0002778603825842966, "loss": 5.4521, "step": 205800 }, { "epoch": 1.241241663752246, "grad_norm": 0.6149612069129944, "learning_rate": 0.00027784974121629216, "loss": 5.0244, "step": 205850 }, { "epoch": 1.24154315553358, "grad_norm": 1.2898327112197876, "learning_rate": 0.0002778390974953957, "loss": 5.4134, "step": 205900 }, { "epoch": 1.2418446473149143, "grad_norm": 1.016901969909668, "learning_rate": 0.0002778284514218031, "loss": 4.9719, "step": 205950 }, { "epoch": 1.2421461390962483, "grad_norm": 1.9282540082931519, "learning_rate": 0.0002778178029957102, "loss": 4.9329, "step": 206000 }, { "epoch": 1.2424476308775823, "grad_norm": 1.8269407749176025, "learning_rate": 0.0002778071522173131, "loss": 4.9121, "step": 206050 }, { "epoch": 1.2427491226589162, "grad_norm": 1.684919834136963, "learning_rate": 0.00027779649908680776, "loss": 5.1355, "step": 206100 }, { "epoch": 1.2430506144402504, "grad_norm": 1.8880549669265747, "learning_rate": 0.00027778584360439024, "loss": 5.344, "step": 206150 }, { "epoch": 1.2433521062215844, "grad_norm": 1.1934624910354614, "learning_rate": 0.0002777751857702566, "loss": 5.0392, "step": 206200 }, { "epoch": 1.2436535980029184, "grad_norm": 1.8462460041046143, "learning_rate": 0.000277764525584603, "loss": 4.6182, "step": 206250 }, { "epoch": 1.2439550897842524, "grad_norm": 1.5487391948699951, "learning_rate": 0.00027775386304762566, "loss": 4.9043, "step": 206300 }, { "epoch": 1.2442565815655866, "grad_norm": 1.1164790391921997, "learning_rate": 0.0002777431981595208, "loss": 5.2758, "step": 206350 }, { "epoch": 1.2445580733469206, "grad_norm": 1.1381663084030151, "learning_rate": 0.0002777325309204847, "loss": 5.2081, "step": 206400 }, { "epoch": 1.2448595651282546, "grad_norm": 2.0282742977142334, "learning_rate": 0.0002777218613307136, "loss": 5.2083, "step": 206450 }, { "epoch": 1.2451610569095886, "grad_norm": 1.5635404586791992, "learning_rate": 0.00027771118939040395, "loss": 5.3351, "step": 206500 }, { "epoch": 1.2454625486909228, "grad_norm": 1.7686184644699097, "learning_rate": 0.00027770051509975205, "loss": 5.2992, "step": 206550 }, { "epoch": 1.2457640404722568, "grad_norm": 1.4275490045547485, "learning_rate": 0.00027768983845895446, "loss": 5.1769, "step": 206600 }, { "epoch": 1.2460655322535907, "grad_norm": 1.176891565322876, "learning_rate": 0.0002776791594682076, "loss": 5.3696, "step": 206650 }, { "epoch": 1.2463670240349247, "grad_norm": 1.3519089221954346, "learning_rate": 0.000277668478127708, "loss": 4.5445, "step": 206700 }, { "epoch": 1.246668515816259, "grad_norm": 1.2062259912490845, "learning_rate": 0.0002776577944376523, "loss": 4.9641, "step": 206750 }, { "epoch": 1.246970007597593, "grad_norm": 1.1593654155731201, "learning_rate": 0.000277647108398237, "loss": 5.233, "step": 206800 }, { "epoch": 1.247271499378927, "grad_norm": 1.2615725994110107, "learning_rate": 0.0002776364200096589, "loss": 4.7172, "step": 206850 }, { "epoch": 1.2475729911602609, "grad_norm": 0.3588660955429077, "learning_rate": 0.0002776257292721146, "loss": 4.4517, "step": 206900 }, { "epoch": 1.247874482941595, "grad_norm": 1.391624093055725, "learning_rate": 0.0002776150361858009, "loss": 5.0638, "step": 206950 }, { "epoch": 1.248175974722929, "grad_norm": 1.385370135307312, "learning_rate": 0.0002776043407509146, "loss": 5.0392, "step": 207000 }, { "epoch": 1.248175974722929, "eval_loss": 5.607182025909424, "eval_runtime": 39.2845, "eval_samples_per_second": 13.033, "eval_steps_per_second": 6.517, "eval_tts_loss": 7.205583007790454, "step": 207000 }, { "epoch": 1.248477466504263, "grad_norm": 1.861032485961914, "learning_rate": 0.00027759364296765245, "loss": 5.0075, "step": 207050 }, { "epoch": 1.248778958285597, "grad_norm": 1.3439393043518066, "learning_rate": 0.0002775829428362114, "loss": 5.3072, "step": 207100 }, { "epoch": 1.2490804500669312, "grad_norm": 1.5163649320602417, "learning_rate": 0.0002775722403567884, "loss": 4.848, "step": 207150 }, { "epoch": 1.2493819418482652, "grad_norm": 1.4893027544021606, "learning_rate": 0.0002775615355295803, "loss": 5.2304, "step": 207200 }, { "epoch": 1.2496834336295992, "grad_norm": 1.9447423219680786, "learning_rate": 0.0002775508283547843, "loss": 4.836, "step": 207250 }, { "epoch": 1.2499849254109332, "grad_norm": 1.3013006448745728, "learning_rate": 0.00027754011883259725, "loss": 5.0549, "step": 207300 }, { "epoch": 1.2502864171922674, "grad_norm": 1.7709988355636597, "learning_rate": 0.0002775294069632163, "loss": 5.145, "step": 207350 }, { "epoch": 1.2505879089736014, "grad_norm": 0.7651171088218689, "learning_rate": 0.00027751869274683866, "loss": 5.1942, "step": 207400 }, { "epoch": 1.2508894007549354, "grad_norm": 1.513644814491272, "learning_rate": 0.00027750797618366144, "loss": 5.0797, "step": 207450 }, { "epoch": 1.2511908925362696, "grad_norm": 1.3958675861358643, "learning_rate": 0.00027749725727388186, "loss": 5.4909, "step": 207500 }, { "epoch": 1.2514923843176036, "grad_norm": 1.6301908493041992, "learning_rate": 0.00027748653601769726, "loss": 5.0492, "step": 207550 }, { "epoch": 1.2517938760989376, "grad_norm": 0.5820406675338745, "learning_rate": 0.00027747581241530485, "loss": 5.2387, "step": 207600 }, { "epoch": 1.2520953678802715, "grad_norm": 1.6748027801513672, "learning_rate": 0.000277465086466902, "loss": 5.1682, "step": 207650 }, { "epoch": 1.2523968596616055, "grad_norm": 1.7677158117294312, "learning_rate": 0.0002774543581726862, "loss": 4.8764, "step": 207700 }, { "epoch": 1.2526983514429397, "grad_norm": 0.6980496048927307, "learning_rate": 0.00027744362753285483, "loss": 4.9701, "step": 207750 }, { "epoch": 1.2529998432242737, "grad_norm": 1.272247076034546, "learning_rate": 0.00027743289454760535, "loss": 5.6662, "step": 207800 }, { "epoch": 1.2533013350056077, "grad_norm": 1.2138922214508057, "learning_rate": 0.0002774221592171353, "loss": 5.4164, "step": 207850 }, { "epoch": 1.253602826786942, "grad_norm": 1.3916090726852417, "learning_rate": 0.00027741142154164223, "loss": 5.0656, "step": 207900 }, { "epoch": 1.2539043185682759, "grad_norm": 1.0995906591415405, "learning_rate": 0.00027740068152132377, "loss": 4.8616, "step": 207950 }, { "epoch": 1.2542058103496099, "grad_norm": 1.2145978212356567, "learning_rate": 0.00027738993915637757, "loss": 4.9059, "step": 208000 }, { "epoch": 1.2545073021309439, "grad_norm": 1.6542911529541016, "learning_rate": 0.0002773791944470014, "loss": 5.2763, "step": 208050 }, { "epoch": 1.2548087939122778, "grad_norm": 1.4385219812393188, "learning_rate": 0.0002773684473933929, "loss": 5.7209, "step": 208100 }, { "epoch": 1.255110285693612, "grad_norm": 1.6432586908340454, "learning_rate": 0.0002773576979957499, "loss": 4.7554, "step": 208150 }, { "epoch": 1.255411777474946, "grad_norm": 0.6720762252807617, "learning_rate": 0.0002773469462542702, "loss": 4.5064, "step": 208200 }, { "epoch": 1.25571326925628, "grad_norm": 1.2898931503295898, "learning_rate": 0.0002773361921691517, "loss": 5.5004, "step": 208250 }, { "epoch": 1.2560147610376142, "grad_norm": 1.2233535051345825, "learning_rate": 0.00027732543574059224, "loss": 5.119, "step": 208300 }, { "epoch": 1.2563162528189482, "grad_norm": 0.5252984166145325, "learning_rate": 0.0002773146769687899, "loss": 4.7274, "step": 208350 }, { "epoch": 1.2566177446002822, "grad_norm": 1.640654444694519, "learning_rate": 0.00027730391585394266, "loss": 5.3043, "step": 208400 }, { "epoch": 1.2569192363816162, "grad_norm": 1.1840641498565674, "learning_rate": 0.00027729315239624847, "loss": 5.1171, "step": 208450 }, { "epoch": 1.2572207281629502, "grad_norm": 1.275545597076416, "learning_rate": 0.00027728238659590547, "loss": 4.9813, "step": 208500 }, { "epoch": 1.2575222199442844, "grad_norm": 1.5242714881896973, "learning_rate": 0.00027727161845311174, "loss": 4.6719, "step": 208550 }, { "epoch": 1.2578237117256184, "grad_norm": 1.146813988685608, "learning_rate": 0.0002772608479680656, "loss": 4.9604, "step": 208600 }, { "epoch": 1.2581252035069523, "grad_norm": 1.420428991317749, "learning_rate": 0.00027725007514096505, "loss": 5.1116, "step": 208650 }, { "epoch": 1.2584266952882865, "grad_norm": 2.355515241622925, "learning_rate": 0.00027723929997200853, "loss": 4.9031, "step": 208700 }, { "epoch": 1.2587281870696205, "grad_norm": 1.027391791343689, "learning_rate": 0.00027722852246139427, "loss": 5.1146, "step": 208750 }, { "epoch": 1.2590296788509545, "grad_norm": 1.5545212030410767, "learning_rate": 0.0002772177426093206, "loss": 5.1377, "step": 208800 }, { "epoch": 1.2593311706322885, "grad_norm": 1.3876115083694458, "learning_rate": 0.00027720696041598595, "loss": 5.2238, "step": 208850 }, { "epoch": 1.2596326624136225, "grad_norm": 1.2765148878097534, "learning_rate": 0.0002771961758815887, "loss": 5.3121, "step": 208900 }, { "epoch": 1.2599341541949567, "grad_norm": 1.188167929649353, "learning_rate": 0.0002771853890063274, "loss": 5.3216, "step": 208950 }, { "epoch": 1.2602356459762907, "grad_norm": 1.6560152769088745, "learning_rate": 0.0002771745997904005, "loss": 5.2026, "step": 209000 }, { "epoch": 1.2605371377576247, "grad_norm": 0.9158074259757996, "learning_rate": 0.00027716380823400655, "loss": 5.156, "step": 209050 }, { "epoch": 1.2608386295389589, "grad_norm": 1.2659351825714111, "learning_rate": 0.00027715301433734417, "loss": 4.9521, "step": 209100 }, { "epoch": 1.2611401213202929, "grad_norm": 1.8248804807662964, "learning_rate": 0.0002771422181006121, "loss": 5.4069, "step": 209150 }, { "epoch": 1.2614416131016268, "grad_norm": 0.7829437851905823, "learning_rate": 0.00027713141952400886, "loss": 5.3763, "step": 209200 }, { "epoch": 1.2617431048829608, "grad_norm": 1.7142683267593384, "learning_rate": 0.00027712061860773335, "loss": 5.2276, "step": 209250 }, { "epoch": 1.2620445966642948, "grad_norm": 1.4362176656723022, "learning_rate": 0.00027710981535198416, "loss": 5.1242, "step": 209300 }, { "epoch": 1.262346088445629, "grad_norm": 1.5321404933929443, "learning_rate": 0.0002770990097569603, "loss": 5.4324, "step": 209350 }, { "epoch": 1.262647580226963, "grad_norm": 1.524880051612854, "learning_rate": 0.0002770882018228605, "loss": 5.4735, "step": 209400 }, { "epoch": 1.262949072008297, "grad_norm": 1.2886754274368286, "learning_rate": 0.00027707739154988375, "loss": 4.8442, "step": 209450 }, { "epoch": 1.2632505637896312, "grad_norm": 1.0909090042114258, "learning_rate": 0.0002770665789382289, "loss": 5.2202, "step": 209500 }, { "epoch": 1.2635520555709652, "grad_norm": 1.4603092670440674, "learning_rate": 0.0002770557639880951, "loss": 5.046, "step": 209550 }, { "epoch": 1.2638535473522992, "grad_norm": 1.1527103185653687, "learning_rate": 0.00027704494669968116, "loss": 5.4595, "step": 209600 }, { "epoch": 1.2641550391336331, "grad_norm": 1.4186060428619385, "learning_rate": 0.0002770341270731864, "loss": 5.2551, "step": 209650 }, { "epoch": 1.2644565309149671, "grad_norm": 1.5360093116760254, "learning_rate": 0.00027702330510880974, "loss": 5.2579, "step": 209700 }, { "epoch": 1.2647580226963013, "grad_norm": 0.7484563589096069, "learning_rate": 0.0002770124808067504, "loss": 5.696, "step": 209750 }, { "epoch": 1.2650595144776353, "grad_norm": 0.31515005230903625, "learning_rate": 0.0002770016541672077, "loss": 4.9374, "step": 209800 }, { "epoch": 1.2653610062589693, "grad_norm": 2.003580331802368, "learning_rate": 0.0002769908251903807, "loss": 5.1459, "step": 209850 }, { "epoch": 1.2656624980403035, "grad_norm": 0.4668184220790863, "learning_rate": 0.0002769799938764688, "loss": 5.2045, "step": 209900 }, { "epoch": 1.2659639898216375, "grad_norm": 1.1750504970550537, "learning_rate": 0.0002769691602256714, "loss": 5.1137, "step": 209950 }, { "epoch": 1.2662654816029715, "grad_norm": 1.192126750946045, "learning_rate": 0.00027695832423818773, "loss": 5.3537, "step": 210000 }, { "epoch": 1.2662654816029715, "eval_loss": 5.619024753570557, "eval_runtime": 39.207, "eval_samples_per_second": 13.059, "eval_steps_per_second": 6.529, "eval_tts_loss": 7.178812910517944, "step": 210000 }, { "epoch": 1.2665669733843055, "grad_norm": 1.1697582006454468, "learning_rate": 0.00027694748591421727, "loss": 5.3721, "step": 210050 }, { "epoch": 1.2668684651656397, "grad_norm": 1.4786019325256348, "learning_rate": 0.00027693664525395956, "loss": 5.5069, "step": 210100 }, { "epoch": 1.2671699569469737, "grad_norm": 1.2678474187850952, "learning_rate": 0.00027692580225761404, "loss": 4.8911, "step": 210150 }, { "epoch": 1.2674714487283076, "grad_norm": 0.9324401617050171, "learning_rate": 0.0002769149569253802, "loss": 5.2696, "step": 210200 }, { "epoch": 1.2677729405096416, "grad_norm": 2.3007164001464844, "learning_rate": 0.00027690410925745775, "loss": 5.4596, "step": 210250 }, { "epoch": 1.2680744322909758, "grad_norm": 1.090295672416687, "learning_rate": 0.00027689325925404624, "loss": 5.1643, "step": 210300 }, { "epoch": 1.2683759240723098, "grad_norm": 1.6650323867797852, "learning_rate": 0.0002768824069153454, "loss": 5.2797, "step": 210350 }, { "epoch": 1.2686774158536438, "grad_norm": 1.7119560241699219, "learning_rate": 0.0002768715522415549, "loss": 4.969, "step": 210400 }, { "epoch": 1.268978907634978, "grad_norm": 1.6938581466674805, "learning_rate": 0.00027686069523287455, "loss": 4.8011, "step": 210450 }, { "epoch": 1.269280399416312, "grad_norm": 1.1164864301681519, "learning_rate": 0.00027684983588950417, "loss": 4.9418, "step": 210500 }, { "epoch": 1.269581891197646, "grad_norm": 1.8624069690704346, "learning_rate": 0.00027683897421164355, "loss": 5.2217, "step": 210550 }, { "epoch": 1.26988338297898, "grad_norm": 0.7956475019454956, "learning_rate": 0.00027682811019949267, "loss": 4.7802, "step": 210600 }, { "epoch": 1.270184874760314, "grad_norm": 0.6266791224479675, "learning_rate": 0.00027681724385325134, "loss": 5.0473, "step": 210650 }, { "epoch": 1.2704863665416481, "grad_norm": 1.784896969795227, "learning_rate": 0.00027680637517311967, "loss": 5.3323, "step": 210700 }, { "epoch": 1.2707878583229821, "grad_norm": 1.7992373704910278, "learning_rate": 0.0002767955041592976, "loss": 5.3856, "step": 210750 }, { "epoch": 1.2710893501043161, "grad_norm": 3.6672699451446533, "learning_rate": 0.0002767846308119853, "loss": 4.9518, "step": 210800 }, { "epoch": 1.2713908418856503, "grad_norm": 1.9298007488250732, "learning_rate": 0.00027677375513138273, "loss": 5.3001, "step": 210850 }, { "epoch": 1.2716923336669843, "grad_norm": 1.2070778608322144, "learning_rate": 0.0002767628771176901, "loss": 5.2977, "step": 210900 }, { "epoch": 1.2719938254483183, "grad_norm": 1.1372851133346558, "learning_rate": 0.0002767519967711076, "loss": 4.9866, "step": 210950 }, { "epoch": 1.2722953172296523, "grad_norm": 1.87486732006073, "learning_rate": 0.00027674111409183553, "loss": 4.7555, "step": 211000 }, { "epoch": 1.2725968090109863, "grad_norm": 0.3645171523094177, "learning_rate": 0.0002767302290800741, "loss": 4.8589, "step": 211050 }, { "epoch": 1.2728983007923205, "grad_norm": 1.3553392887115479, "learning_rate": 0.0002767193417360237, "loss": 5.1546, "step": 211100 }, { "epoch": 1.2731997925736545, "grad_norm": 1.138318419456482, "learning_rate": 0.0002767084520598846, "loss": 5.2654, "step": 211150 }, { "epoch": 1.2735012843549884, "grad_norm": 1.5286965370178223, "learning_rate": 0.00027669756005185733, "loss": 4.8083, "step": 211200 }, { "epoch": 1.2738027761363226, "grad_norm": 1.1711761951446533, "learning_rate": 0.0002766866657121422, "loss": 5.3109, "step": 211250 }, { "epoch": 1.2741042679176566, "grad_norm": 0.832142174243927, "learning_rate": 0.00027667576904093986, "loss": 5.2792, "step": 211300 }, { "epoch": 1.2744057596989906, "grad_norm": 1.0791499614715576, "learning_rate": 0.0002766648700384507, "loss": 5.5038, "step": 211350 }, { "epoch": 1.2747072514803246, "grad_norm": 1.1831506490707397, "learning_rate": 0.0002766539687048754, "loss": 5.3187, "step": 211400 }, { "epoch": 1.2750087432616586, "grad_norm": 1.4998561143875122, "learning_rate": 0.0002766430650404145, "loss": 5.0655, "step": 211450 }, { "epoch": 1.2753102350429928, "grad_norm": 2.678424119949341, "learning_rate": 0.0002766321590452688, "loss": 5.1773, "step": 211500 }, { "epoch": 1.2756117268243268, "grad_norm": 2.1245996952056885, "learning_rate": 0.00027662125071963885, "loss": 5.0399, "step": 211550 }, { "epoch": 1.2759132186056608, "grad_norm": 1.132633924484253, "learning_rate": 0.0002766103400637255, "loss": 4.9816, "step": 211600 }, { "epoch": 1.276214710386995, "grad_norm": 0.513213038444519, "learning_rate": 0.0002765994270777295, "loss": 5.5026, "step": 211650 }, { "epoch": 1.276516202168329, "grad_norm": 1.044446587562561, "learning_rate": 0.0002765885117618518, "loss": 4.8285, "step": 211700 }, { "epoch": 1.276817693949663, "grad_norm": 1.120420217514038, "learning_rate": 0.0002765775941162931, "loss": 5.7389, "step": 211750 }, { "epoch": 1.277119185730997, "grad_norm": 1.3901910781860352, "learning_rate": 0.0002765666741412545, "loss": 5.6252, "step": 211800 }, { "epoch": 1.277420677512331, "grad_norm": 1.4976966381072998, "learning_rate": 0.00027655575183693683, "loss": 5.1653, "step": 211850 }, { "epoch": 1.277722169293665, "grad_norm": 0.7566967606544495, "learning_rate": 0.00027654482720354117, "loss": 4.8762, "step": 211900 }, { "epoch": 1.278023661074999, "grad_norm": 1.1916908025741577, "learning_rate": 0.0002765339002412685, "loss": 5.0513, "step": 211950 }, { "epoch": 1.278325152856333, "grad_norm": 1.4446054697036743, "learning_rate": 0.00027652297095032, "loss": 4.9939, "step": 212000 }, { "epoch": 1.2786266446376673, "grad_norm": 1.6369491815567017, "learning_rate": 0.0002765120393308968, "loss": 5.325, "step": 212050 }, { "epoch": 1.2789281364190013, "grad_norm": 0.9732106328010559, "learning_rate": 0.00027650110538320007, "loss": 5.2123, "step": 212100 }, { "epoch": 1.2792296282003353, "grad_norm": 1.1773344278335571, "learning_rate": 0.000276490169107431, "loss": 4.7783, "step": 212150 }, { "epoch": 1.2795311199816692, "grad_norm": 1.0634468793869019, "learning_rate": 0.0002764792305037909, "loss": 5.7363, "step": 212200 }, { "epoch": 1.2798326117630032, "grad_norm": 0.9507513046264648, "learning_rate": 0.000276468289572481, "loss": 5.3633, "step": 212250 }, { "epoch": 1.2801341035443374, "grad_norm": 1.161782145500183, "learning_rate": 0.0002764573463137028, "loss": 5.3394, "step": 212300 }, { "epoch": 1.2804355953256714, "grad_norm": 1.171125888824463, "learning_rate": 0.00027644640072765756, "loss": 5.3354, "step": 212350 }, { "epoch": 1.2807370871070054, "grad_norm": 1.0859720706939697, "learning_rate": 0.0002764354528145467, "loss": 4.8205, "step": 212400 }, { "epoch": 1.2810385788883396, "grad_norm": 1.9226337671279907, "learning_rate": 0.00027642450257457185, "loss": 4.9041, "step": 212450 }, { "epoch": 1.2813400706696736, "grad_norm": 1.536464810371399, "learning_rate": 0.0002764135500079344, "loss": 5.043, "step": 212500 }, { "epoch": 1.2816415624510076, "grad_norm": 1.1956754922866821, "learning_rate": 0.000276402595114836, "loss": 4.8281, "step": 212550 }, { "epoch": 1.2819430542323416, "grad_norm": 1.1972094774246216, "learning_rate": 0.0002763916378954782, "loss": 5.4443, "step": 212600 }, { "epoch": 1.2822445460136755, "grad_norm": 1.2634422779083252, "learning_rate": 0.0002763806783500627, "loss": 4.9998, "step": 212650 }, { "epoch": 1.2825460377950098, "grad_norm": 1.4444918632507324, "learning_rate": 0.0002763697164787911, "loss": 5.2313, "step": 212700 }, { "epoch": 1.2828475295763437, "grad_norm": 0.911903977394104, "learning_rate": 0.0002763587522818653, "loss": 5.4346, "step": 212750 }, { "epoch": 1.2831490213576777, "grad_norm": 1.2587319612503052, "learning_rate": 0.00027634778575948697, "loss": 5.0247, "step": 212800 }, { "epoch": 1.283450513139012, "grad_norm": 1.5496034622192383, "learning_rate": 0.0002763368169118579, "loss": 5.4167, "step": 212850 }, { "epoch": 1.283752004920346, "grad_norm": 1.1519099473953247, "learning_rate": 0.00027632584573918003, "loss": 5.1308, "step": 212900 }, { "epoch": 1.28405349670168, "grad_norm": 0.9305027723312378, "learning_rate": 0.00027631487224165526, "loss": 5.3133, "step": 212950 }, { "epoch": 1.2843549884830139, "grad_norm": 1.3339946269989014, "learning_rate": 0.0002763038964194855, "loss": 5.2561, "step": 213000 }, { "epoch": 1.2843549884830139, "eval_loss": 5.599466323852539, "eval_runtime": 39.3384, "eval_samples_per_second": 13.015, "eval_steps_per_second": 6.508, "eval_tts_loss": 7.240734812545815, "step": 213000 }, { "epoch": 1.2846564802643479, "grad_norm": 1.7338310480117798, "learning_rate": 0.0002762929182728728, "loss": 5.5162, "step": 213050 }, { "epoch": 1.284957972045682, "grad_norm": 0.9422967433929443, "learning_rate": 0.00027628193780201917, "loss": 5.2613, "step": 213100 }, { "epoch": 1.285259463827016, "grad_norm": 1.6321359872817993, "learning_rate": 0.0002762709550071266, "loss": 5.1244, "step": 213150 }, { "epoch": 1.28556095560835, "grad_norm": 1.136759638786316, "learning_rate": 0.00027625996988839746, "loss": 5.2379, "step": 213200 }, { "epoch": 1.2858624473896842, "grad_norm": 1.1780829429626465, "learning_rate": 0.00027624898244603364, "loss": 5.1622, "step": 213250 }, { "epoch": 1.2861639391710182, "grad_norm": 1.3057527542114258, "learning_rate": 0.0002762379926802375, "loss": 5.1542, "step": 213300 }, { "epoch": 1.2864654309523522, "grad_norm": 1.1415355205535889, "learning_rate": 0.00027622700059121125, "loss": 5.0648, "step": 213350 }, { "epoch": 1.2867669227336862, "grad_norm": 1.353996753692627, "learning_rate": 0.0002762160061791572, "loss": 5.4618, "step": 213400 }, { "epoch": 1.2870684145150202, "grad_norm": 2.0700817108154297, "learning_rate": 0.0002762050094442777, "loss": 5.3849, "step": 213450 }, { "epoch": 1.2873699062963544, "grad_norm": 1.706700086593628, "learning_rate": 0.00027619401038677505, "loss": 4.9905, "step": 213500 }, { "epoch": 1.2876713980776884, "grad_norm": 1.8972009420394897, "learning_rate": 0.0002761830090068518, "loss": 5.1125, "step": 213550 }, { "epoch": 1.2879728898590224, "grad_norm": 1.0760949850082397, "learning_rate": 0.00027617200530471027, "loss": 5.1749, "step": 213600 }, { "epoch": 1.2882743816403566, "grad_norm": 1.122373342514038, "learning_rate": 0.00027616099928055307, "loss": 5.2111, "step": 213650 }, { "epoch": 1.2885758734216906, "grad_norm": 2.00179123878479, "learning_rate": 0.0002761499909345827, "loss": 4.8952, "step": 213700 }, { "epoch": 1.2888773652030245, "grad_norm": 1.0692052841186523, "learning_rate": 0.0002761389802670018, "loss": 4.7176, "step": 213750 }, { "epoch": 1.2891788569843585, "grad_norm": 1.459024429321289, "learning_rate": 0.000276127967278013, "loss": 5.2576, "step": 213800 }, { "epoch": 1.2894803487656927, "grad_norm": 1.220977783203125, "learning_rate": 0.0002761169519678189, "loss": 5.1849, "step": 213850 }, { "epoch": 1.2897818405470267, "grad_norm": 2.3121423721313477, "learning_rate": 0.0002761059343366223, "loss": 5.4129, "step": 213900 }, { "epoch": 1.2900833323283607, "grad_norm": 1.4415782690048218, "learning_rate": 0.0002760949143846259, "loss": 5.2114, "step": 213950 }, { "epoch": 1.2903848241096947, "grad_norm": 1.3922748565673828, "learning_rate": 0.0002760838921120326, "loss": 5.1376, "step": 214000 }, { "epoch": 1.290686315891029, "grad_norm": 2.112678050994873, "learning_rate": 0.00027607286751904524, "loss": 5.1851, "step": 214050 }, { "epoch": 1.2909878076723629, "grad_norm": 2.80306339263916, "learning_rate": 0.0002760618406058666, "loss": 4.9175, "step": 214100 }, { "epoch": 1.2912892994536969, "grad_norm": 1.206876277923584, "learning_rate": 0.00027605081137269973, "loss": 4.6478, "step": 214150 }, { "epoch": 1.291590791235031, "grad_norm": 1.537032127380371, "learning_rate": 0.00027603977981974746, "loss": 4.9879, "step": 214200 }, { "epoch": 1.291892283016365, "grad_norm": 1.6045047044754028, "learning_rate": 0.000276028745947213, "loss": 4.9571, "step": 214250 }, { "epoch": 1.292193774797699, "grad_norm": 1.675628423690796, "learning_rate": 0.00027601770975529933, "loss": 5.3722, "step": 214300 }, { "epoch": 1.292495266579033, "grad_norm": 1.3861558437347412, "learning_rate": 0.0002760066712442095, "loss": 5.5189, "step": 214350 }, { "epoch": 1.292796758360367, "grad_norm": 0.4654754102230072, "learning_rate": 0.00027599563041414674, "loss": 5.5897, "step": 214400 }, { "epoch": 1.2930982501417012, "grad_norm": 1.08905827999115, "learning_rate": 0.00027598458726531424, "loss": 5.2987, "step": 214450 }, { "epoch": 1.2933997419230352, "grad_norm": 1.4381070137023926, "learning_rate": 0.0002759735417979151, "loss": 5.226, "step": 214500 }, { "epoch": 1.2937012337043692, "grad_norm": 1.8033134937286377, "learning_rate": 0.00027596249401215273, "loss": 5.0525, "step": 214550 }, { "epoch": 1.2940027254857034, "grad_norm": 1.1835622787475586, "learning_rate": 0.00027595144390823046, "loss": 4.8591, "step": 214600 }, { "epoch": 1.2943042172670374, "grad_norm": 1.5957645177841187, "learning_rate": 0.00027594039148635165, "loss": 4.9202, "step": 214650 }, { "epoch": 1.2946057090483714, "grad_norm": 1.215847373008728, "learning_rate": 0.00027592933674671955, "loss": 5.238, "step": 214700 }, { "epoch": 1.2949072008297053, "grad_norm": 2.3205056190490723, "learning_rate": 0.0002759182796895378, "loss": 4.6101, "step": 214750 }, { "epoch": 1.2952086926110393, "grad_norm": 1.9661738872528076, "learning_rate": 0.00027590722031500977, "loss": 4.8289, "step": 214800 }, { "epoch": 1.2955101843923735, "grad_norm": 2.3732147216796875, "learning_rate": 0.00027589615862333905, "loss": 5.1329, "step": 214850 }, { "epoch": 1.2958116761737075, "grad_norm": 1.319786787033081, "learning_rate": 0.0002758850946147292, "loss": 5.3084, "step": 214900 }, { "epoch": 1.2961131679550415, "grad_norm": 1.08543062210083, "learning_rate": 0.00027587402828938376, "loss": 5.1763, "step": 214950 }, { "epoch": 1.2964146597363757, "grad_norm": 1.7266048192977905, "learning_rate": 0.0002758629596475065, "loss": 5.4491, "step": 215000 }, { "epoch": 1.2967161515177097, "grad_norm": 1.5532195568084717, "learning_rate": 0.0002758518886893011, "loss": 5.1772, "step": 215050 }, { "epoch": 1.2970176432990437, "grad_norm": 0.972000241279602, "learning_rate": 0.0002758408154149713, "loss": 5.291, "step": 215100 }, { "epoch": 1.2973191350803777, "grad_norm": 0.6895864605903625, "learning_rate": 0.00027582973982472086, "loss": 4.9216, "step": 215150 }, { "epoch": 1.2976206268617116, "grad_norm": 1.0085631608963013, "learning_rate": 0.00027581866191875363, "loss": 4.6883, "step": 215200 }, { "epoch": 1.2979221186430459, "grad_norm": 0.4423551857471466, "learning_rate": 0.0002758075816972735, "loss": 5.2266, "step": 215250 }, { "epoch": 1.2982236104243798, "grad_norm": 1.194225549697876, "learning_rate": 0.00027579649916048435, "loss": 5.51, "step": 215300 }, { "epoch": 1.2985251022057138, "grad_norm": 1.9582480192184448, "learning_rate": 0.0002757854143085901, "loss": 5.1556, "step": 215350 }, { "epoch": 1.298826593987048, "grad_norm": 2.199399471282959, "learning_rate": 0.0002757743271417949, "loss": 5.3045, "step": 215400 }, { "epoch": 1.299128085768382, "grad_norm": 1.5727758407592773, "learning_rate": 0.0002757632376603026, "loss": 5.0065, "step": 215450 }, { "epoch": 1.299429577549716, "grad_norm": 2.800666093826294, "learning_rate": 0.00027575214586431746, "loss": 5.0975, "step": 215500 }, { "epoch": 1.29973106933105, "grad_norm": 1.4331401586532593, "learning_rate": 0.0002757410517540435, "loss": 5.4892, "step": 215550 }, { "epoch": 1.300032561112384, "grad_norm": 1.5175269842147827, "learning_rate": 0.000275729955329685, "loss": 5.1267, "step": 215600 }, { "epoch": 1.3003340528937182, "grad_norm": 1.2554329633712769, "learning_rate": 0.000275718856591446, "loss": 5.4576, "step": 215650 }, { "epoch": 1.3006355446750522, "grad_norm": 1.749381184577942, "learning_rate": 0.00027570775553953093, "loss": 5.4699, "step": 215700 }, { "epoch": 1.3009370364563861, "grad_norm": 1.2989132404327393, "learning_rate": 0.0002756966521741439, "loss": 4.6906, "step": 215750 }, { "epoch": 1.3012385282377203, "grad_norm": 3.014988422393799, "learning_rate": 0.0002756855464954895, "loss": 4.7792, "step": 215800 }, { "epoch": 1.3015400200190543, "grad_norm": 1.464569330215454, "learning_rate": 0.0002756744385037719, "loss": 5.2713, "step": 215850 }, { "epoch": 1.3018415118003883, "grad_norm": 1.278063178062439, "learning_rate": 0.0002756633281991956, "loss": 5.5585, "step": 215900 }, { "epoch": 1.3021430035817223, "grad_norm": 1.70377516746521, "learning_rate": 0.00027565221558196513, "loss": 5.5077, "step": 215950 }, { "epoch": 1.3024444953630563, "grad_norm": 1.7035630941390991, "learning_rate": 0.0002756411006522849, "loss": 5.0601, "step": 216000 }, { "epoch": 1.3024444953630563, "eval_loss": 5.594174861907959, "eval_runtime": 39.0327, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.559, "eval_tts_loss": 7.302802365381115, "step": 216000 }, { "epoch": 1.3027459871443905, "grad_norm": 1.59882652759552, "learning_rate": 0.00027562998341035955, "loss": 4.9088, "step": 216050 }, { "epoch": 1.3030474789257245, "grad_norm": 1.5453481674194336, "learning_rate": 0.0002756188638563936, "loss": 5.339, "step": 216100 }, { "epoch": 1.3033489707070585, "grad_norm": 1.4357348680496216, "learning_rate": 0.0002756077419905917, "loss": 4.9319, "step": 216150 }, { "epoch": 1.3036504624883927, "grad_norm": 2.3530991077423096, "learning_rate": 0.0002755966178131587, "loss": 5.3719, "step": 216200 }, { "epoch": 1.3039519542697267, "grad_norm": 2.537184476852417, "learning_rate": 0.000275585491324299, "loss": 5.0527, "step": 216250 }, { "epoch": 1.3042534460510606, "grad_norm": 1.0928198099136353, "learning_rate": 0.0002755743625242177, "loss": 5.4363, "step": 216300 }, { "epoch": 1.3045549378323946, "grad_norm": 1.2949228286743164, "learning_rate": 0.0002755632314131194, "loss": 5.3556, "step": 216350 }, { "epoch": 1.3048564296137286, "grad_norm": 1.1190526485443115, "learning_rate": 0.00027555209799120904, "loss": 4.3648, "step": 216400 }, { "epoch": 1.3051579213950628, "grad_norm": 0.8058116436004639, "learning_rate": 0.00027554096225869146, "loss": 4.8037, "step": 216450 }, { "epoch": 1.3054594131763968, "grad_norm": 1.183022379875183, "learning_rate": 0.00027552982421577164, "loss": 5.418, "step": 216500 }, { "epoch": 1.3057609049577308, "grad_norm": 1.5691287517547607, "learning_rate": 0.00027551868386265453, "loss": 5.1793, "step": 216550 }, { "epoch": 1.306062396739065, "grad_norm": 1.0980854034423828, "learning_rate": 0.00027550754119954517, "loss": 5.042, "step": 216600 }, { "epoch": 1.306363888520399, "grad_norm": 1.3705475330352783, "learning_rate": 0.0002754963962266486, "loss": 5.5147, "step": 216650 }, { "epoch": 1.306665380301733, "grad_norm": 1.2175995111465454, "learning_rate": 0.00027548524894417003, "loss": 5.1694, "step": 216700 }, { "epoch": 1.306966872083067, "grad_norm": 1.3512771129608154, "learning_rate": 0.00027547409935231446, "loss": 5.0705, "step": 216750 }, { "epoch": 1.307268363864401, "grad_norm": 1.3618495464324951, "learning_rate": 0.0002754629474512872, "loss": 5.5908, "step": 216800 }, { "epoch": 1.3075698556457351, "grad_norm": 1.9969284534454346, "learning_rate": 0.0002754517932412934, "loss": 5.0024, "step": 216850 }, { "epoch": 1.3078713474270691, "grad_norm": 1.3115708827972412, "learning_rate": 0.00027544063672253836, "loss": 5.1721, "step": 216900 }, { "epoch": 1.308172839208403, "grad_norm": 1.6495615243911743, "learning_rate": 0.00027542947789522745, "loss": 5.1321, "step": 216950 }, { "epoch": 1.3084743309897373, "grad_norm": 0.5923617482185364, "learning_rate": 0.000275418316759566, "loss": 5.5286, "step": 217000 }, { "epoch": 1.3087758227710713, "grad_norm": 1.594437599182129, "learning_rate": 0.0002754071533157594, "loss": 4.8624, "step": 217050 }, { "epoch": 1.3090773145524053, "grad_norm": 0.8454617857933044, "learning_rate": 0.0002753959875640131, "loss": 4.6678, "step": 217100 }, { "epoch": 1.3093788063337393, "grad_norm": 1.3292059898376465, "learning_rate": 0.00027538481950453264, "loss": 4.7353, "step": 217150 }, { "epoch": 1.3096802981150732, "grad_norm": 0.8657404780387878, "learning_rate": 0.0002753736491375235, "loss": 5.3106, "step": 217200 }, { "epoch": 1.3099817898964075, "grad_norm": 0.6757892370223999, "learning_rate": 0.0002753624764631912, "loss": 4.6674, "step": 217250 }, { "epoch": 1.3102832816777414, "grad_norm": 1.4782980680465698, "learning_rate": 0.0002753513014817415, "loss": 4.8345, "step": 217300 }, { "epoch": 1.3105847734590754, "grad_norm": 1.5647149085998535, "learning_rate": 0.0002753401241933799, "loss": 4.732, "step": 217350 }, { "epoch": 1.3108862652404096, "grad_norm": 1.116824746131897, "learning_rate": 0.00027532894459831225, "loss": 5.2456, "step": 217400 }, { "epoch": 1.3111877570217436, "grad_norm": 1.5781092643737793, "learning_rate": 0.0002753177626967442, "loss": 5.4976, "step": 217450 }, { "epoch": 1.3114892488030776, "grad_norm": 0.6275713443756104, "learning_rate": 0.0002753065784888816, "loss": 5.1331, "step": 217500 }, { "epoch": 1.3117907405844118, "grad_norm": 0.3542298674583435, "learning_rate": 0.00027529539197493027, "loss": 5.091, "step": 217550 }, { "epoch": 1.3120922323657458, "grad_norm": 1.5183560848236084, "learning_rate": 0.000275284203155096, "loss": 4.4954, "step": 217600 }, { "epoch": 1.3123937241470798, "grad_norm": 1.233657717704773, "learning_rate": 0.00027527301202958477, "loss": 5.0584, "step": 217650 }, { "epoch": 1.3126952159284138, "grad_norm": 1.3598617315292358, "learning_rate": 0.00027526181859860254, "loss": 4.9049, "step": 217700 }, { "epoch": 1.3129967077097477, "grad_norm": 1.5993871688842773, "learning_rate": 0.0002752506228623553, "loss": 4.5889, "step": 217750 }, { "epoch": 1.313298199491082, "grad_norm": 1.5163087844848633, "learning_rate": 0.00027523942482104904, "loss": 5.2548, "step": 217800 }, { "epoch": 1.313599691272416, "grad_norm": 1.085030198097229, "learning_rate": 0.0002752282244748899, "loss": 5.4131, "step": 217850 }, { "epoch": 1.31390118305375, "grad_norm": 1.2277367115020752, "learning_rate": 0.000275217021824084, "loss": 5.3415, "step": 217900 }, { "epoch": 1.3142026748350841, "grad_norm": 0.7875383496284485, "learning_rate": 0.0002752058168688376, "loss": 5.0087, "step": 217950 }, { "epoch": 1.3145041666164181, "grad_norm": 0.6613393425941467, "learning_rate": 0.00027519460960935674, "loss": 5.1017, "step": 218000 }, { "epoch": 1.314805658397752, "grad_norm": 1.1766436100006104, "learning_rate": 0.00027518340004584776, "loss": 5.0152, "step": 218050 }, { "epoch": 1.315107150179086, "grad_norm": 1.310745120048523, "learning_rate": 0.00027517218817851695, "loss": 5.2431, "step": 218100 }, { "epoch": 1.31540864196042, "grad_norm": 1.3005717992782593, "learning_rate": 0.00027516097400757066, "loss": 5.494, "step": 218150 }, { "epoch": 1.3157101337417543, "grad_norm": 1.691750168800354, "learning_rate": 0.0002751497575332152, "loss": 5.5039, "step": 218200 }, { "epoch": 1.3160116255230883, "grad_norm": 1.7528107166290283, "learning_rate": 0.00027513853875565713, "loss": 5.3759, "step": 218250 }, { "epoch": 1.3163131173044222, "grad_norm": 1.8352484703063965, "learning_rate": 0.0002751273176751028, "loss": 5.1986, "step": 218300 }, { "epoch": 1.3166146090857564, "grad_norm": 1.5434483289718628, "learning_rate": 0.00027511609429175875, "loss": 5.3763, "step": 218350 }, { "epoch": 1.3169161008670904, "grad_norm": 1.749005675315857, "learning_rate": 0.00027510486860583155, "loss": 5.1419, "step": 218400 }, { "epoch": 1.3172175926484244, "grad_norm": 1.1849520206451416, "learning_rate": 0.0002750936406175278, "loss": 4.9053, "step": 218450 }, { "epoch": 1.3175190844297584, "grad_norm": 1.4771448373794556, "learning_rate": 0.0002750824103270541, "loss": 4.9931, "step": 218500 }, { "epoch": 1.3178205762110924, "grad_norm": 1.1483092308044434, "learning_rate": 0.0002750711777346171, "loss": 5.0814, "step": 218550 }, { "epoch": 1.3181220679924266, "grad_norm": 2.0638880729675293, "learning_rate": 0.0002750599428404236, "loss": 5.2703, "step": 218600 }, { "epoch": 1.3184235597737606, "grad_norm": 1.08284592628479, "learning_rate": 0.0002750487056446803, "loss": 5.3673, "step": 218650 }, { "epoch": 1.3187250515550946, "grad_norm": 1.6887060403823853, "learning_rate": 0.00027503746614759405, "loss": 5.2056, "step": 218700 }, { "epoch": 1.3190265433364288, "grad_norm": 1.6733343601226807, "learning_rate": 0.00027502622434937165, "loss": 5.2571, "step": 218750 }, { "epoch": 1.3193280351177628, "grad_norm": 1.25657057762146, "learning_rate": 0.00027501498025022005, "loss": 4.9723, "step": 218800 }, { "epoch": 1.3196295268990967, "grad_norm": 1.0589383840560913, "learning_rate": 0.00027500373385034606, "loss": 4.9494, "step": 218850 }, { "epoch": 1.3199310186804307, "grad_norm": 1.0789077281951904, "learning_rate": 0.0002749924851499568, "loss": 5.0649, "step": 218900 }, { "epoch": 1.3202325104617647, "grad_norm": 1.6461882591247559, "learning_rate": 0.0002749812341492592, "loss": 5.1977, "step": 218950 }, { "epoch": 1.320534002243099, "grad_norm": 1.1782540082931519, "learning_rate": 0.00027496998084846036, "loss": 5.17, "step": 219000 }, { "epoch": 1.320534002243099, "eval_loss": 5.597347259521484, "eval_runtime": 39.0123, "eval_samples_per_second": 13.124, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.260924700953286, "step": 219000 }, { "epoch": 1.320835494024433, "grad_norm": 1.7778422832489014, "learning_rate": 0.00027495872524776734, "loss": 4.8326, "step": 219050 }, { "epoch": 1.3211369858057669, "grad_norm": 1.4042809009552002, "learning_rate": 0.00027494746734738735, "loss": 5.3083, "step": 219100 }, { "epoch": 1.321438477587101, "grad_norm": 1.9499980211257935, "learning_rate": 0.0002749362071475275, "loss": 5.2841, "step": 219150 }, { "epoch": 1.321739969368435, "grad_norm": 1.1259026527404785, "learning_rate": 0.0002749249446483951, "loss": 5.1012, "step": 219200 }, { "epoch": 1.322041461149769, "grad_norm": 0.658138632774353, "learning_rate": 0.00027491367985019735, "loss": 5.2013, "step": 219250 }, { "epoch": 1.322342952931103, "grad_norm": 1.8473076820373535, "learning_rate": 0.00027490241275314155, "loss": 5.4032, "step": 219300 }, { "epoch": 1.322644444712437, "grad_norm": 1.2782071828842163, "learning_rate": 0.0002748911433574351, "loss": 5.291, "step": 219350 }, { "epoch": 1.3229459364937712, "grad_norm": 1.1695828437805176, "learning_rate": 0.0002748798716632854, "loss": 5.4595, "step": 219400 }, { "epoch": 1.3232474282751052, "grad_norm": 1.2191298007965088, "learning_rate": 0.0002748685976708999, "loss": 4.7599, "step": 219450 }, { "epoch": 1.3235489200564392, "grad_norm": 1.9031862020492554, "learning_rate": 0.000274857321380486, "loss": 5.0459, "step": 219500 }, { "epoch": 1.3238504118377734, "grad_norm": 0.9159022569656372, "learning_rate": 0.00027484604279225133, "loss": 5.0408, "step": 219550 }, { "epoch": 1.3241519036191074, "grad_norm": 1.4506257772445679, "learning_rate": 0.00027483476190640337, "loss": 5.4901, "step": 219600 }, { "epoch": 1.3244533954004414, "grad_norm": 1.1104960441589355, "learning_rate": 0.00027482347872314977, "loss": 5.1208, "step": 219650 }, { "epoch": 1.3247548871817754, "grad_norm": 0.685017466545105, "learning_rate": 0.0002748121932426982, "loss": 5.4797, "step": 219700 }, { "epoch": 1.3250563789631093, "grad_norm": 1.5075000524520874, "learning_rate": 0.00027480090546525633, "loss": 5.0211, "step": 219750 }, { "epoch": 1.3253578707444436, "grad_norm": 2.137244701385498, "learning_rate": 0.00027478961539103186, "loss": 5.614, "step": 219800 }, { "epoch": 1.3256593625257775, "grad_norm": 1.1600087881088257, "learning_rate": 0.0002747783230202326, "loss": 4.7936, "step": 219850 }, { "epoch": 1.3259608543071115, "grad_norm": 1.5270277261734009, "learning_rate": 0.00027476702835306646, "loss": 5.3514, "step": 219900 }, { "epoch": 1.3262623460884457, "grad_norm": 1.5458048582077026, "learning_rate": 0.00027475573138974115, "loss": 4.9688, "step": 219950 }, { "epoch": 1.3265638378697797, "grad_norm": 1.5404722690582275, "learning_rate": 0.00027474443213046466, "loss": 5.0279, "step": 220000 }, { "epoch": 1.3268653296511137, "grad_norm": 1.327260136604309, "learning_rate": 0.0002747331305754449, "loss": 4.9058, "step": 220050 }, { "epoch": 1.3271668214324477, "grad_norm": 1.2281889915466309, "learning_rate": 0.00027472182672488984, "loss": 5.5017, "step": 220100 }, { "epoch": 1.3274683132137817, "grad_norm": 0.916061520576477, "learning_rate": 0.0002747105205790076, "loss": 4.9133, "step": 220150 }, { "epoch": 1.3277698049951159, "grad_norm": 1.2490403652191162, "learning_rate": 0.0002746992121380062, "loss": 5.2547, "step": 220200 }, { "epoch": 1.3280712967764499, "grad_norm": 0.7164002656936646, "learning_rate": 0.0002746879014020937, "loss": 5.1725, "step": 220250 }, { "epoch": 1.3283727885577838, "grad_norm": 2.7584822177886963, "learning_rate": 0.0002746765883714784, "loss": 5.1134, "step": 220300 }, { "epoch": 1.328674280339118, "grad_norm": 1.6409095525741577, "learning_rate": 0.00027466527304636834, "loss": 5.4071, "step": 220350 }, { "epoch": 1.328975772120452, "grad_norm": 1.2634035348892212, "learning_rate": 0.0002746539554269719, "loss": 5.0823, "step": 220400 }, { "epoch": 1.329277263901786, "grad_norm": 1.647263526916504, "learning_rate": 0.0002746426355134972, "loss": 5.2257, "step": 220450 }, { "epoch": 1.32957875568312, "grad_norm": 1.6832455396652222, "learning_rate": 0.0002746313133061528, "loss": 5.1925, "step": 220500 }, { "epoch": 1.329880247464454, "grad_norm": 2.690070152282715, "learning_rate": 0.00027461998880514685, "loss": 5.2155, "step": 220550 }, { "epoch": 1.3301817392457882, "grad_norm": 1.3446481227874756, "learning_rate": 0.00027460866201068783, "loss": 4.9489, "step": 220600 }, { "epoch": 1.3304832310271222, "grad_norm": 1.4025312662124634, "learning_rate": 0.0002745973329229843, "loss": 5.2756, "step": 220650 }, { "epoch": 1.3307847228084562, "grad_norm": 1.9331897497177124, "learning_rate": 0.00027458600154224463, "loss": 5.2153, "step": 220700 }, { "epoch": 1.3310862145897904, "grad_norm": 0.7531061172485352, "learning_rate": 0.0002745746678686774, "loss": 4.7962, "step": 220750 }, { "epoch": 1.3313877063711244, "grad_norm": 1.0948486328125, "learning_rate": 0.00027456333190249113, "loss": 4.7059, "step": 220800 }, { "epoch": 1.3316891981524583, "grad_norm": 1.3369019031524658, "learning_rate": 0.00027455199364389454, "loss": 5.5842, "step": 220850 }, { "epoch": 1.3319906899337923, "grad_norm": 1.1220308542251587, "learning_rate": 0.00027454065309309624, "loss": 5.003, "step": 220900 }, { "epoch": 1.3322921817151263, "grad_norm": 1.5734611749649048, "learning_rate": 0.00027452931025030497, "loss": 4.6601, "step": 220950 }, { "epoch": 1.3325936734964605, "grad_norm": 1.1654906272888184, "learning_rate": 0.0002745179651157294, "loss": 5.3779, "step": 221000 }, { "epoch": 1.3328951652777945, "grad_norm": 1.6199949979782104, "learning_rate": 0.00027450661768957846, "loss": 5.4136, "step": 221050 }, { "epoch": 1.3331966570591285, "grad_norm": 1.286673665046692, "learning_rate": 0.00027449526797206086, "loss": 5.0555, "step": 221100 }, { "epoch": 1.3334981488404627, "grad_norm": 1.773094892501831, "learning_rate": 0.00027448391596338555, "loss": 4.9133, "step": 221150 }, { "epoch": 1.3337996406217967, "grad_norm": 1.5270987749099731, "learning_rate": 0.0002744725616637614, "loss": 4.9127, "step": 221200 }, { "epoch": 1.3341011324031307, "grad_norm": 1.1188170909881592, "learning_rate": 0.0002744612050733973, "loss": 5.0668, "step": 221250 }, { "epoch": 1.3344026241844649, "grad_norm": 1.325176477432251, "learning_rate": 0.0002744498461925025, "loss": 5.1851, "step": 221300 }, { "epoch": 1.3347041159657989, "grad_norm": 1.1859517097473145, "learning_rate": 0.0002744384850212857, "loss": 5.2121, "step": 221350 }, { "epoch": 1.3350056077471328, "grad_norm": 1.0597196817398071, "learning_rate": 0.00027442712155995627, "loss": 4.5524, "step": 221400 }, { "epoch": 1.3353070995284668, "grad_norm": 0.8031468987464905, "learning_rate": 0.00027441575580872325, "loss": 4.5977, "step": 221450 }, { "epoch": 1.3356085913098008, "grad_norm": 1.0787891149520874, "learning_rate": 0.00027440438776779575, "loss": 5.2834, "step": 221500 }, { "epoch": 1.335910083091135, "grad_norm": 1.040008306503296, "learning_rate": 0.000274393017437383, "loss": 4.9845, "step": 221550 }, { "epoch": 1.336211574872469, "grad_norm": 1.3678394556045532, "learning_rate": 0.0002743816448176944, "loss": 5.3033, "step": 221600 }, { "epoch": 1.336513066653803, "grad_norm": 1.1906851530075073, "learning_rate": 0.000274370269908939, "loss": 5.0513, "step": 221650 }, { "epoch": 1.3368145584351372, "grad_norm": 1.5512456893920898, "learning_rate": 0.00027435889271132635, "loss": 4.7289, "step": 221700 }, { "epoch": 1.3371160502164712, "grad_norm": 1.3253490924835205, "learning_rate": 0.0002743475132250657, "loss": 5.3133, "step": 221750 }, { "epoch": 1.3374175419978052, "grad_norm": 0.6713650822639465, "learning_rate": 0.00027433613145036657, "loss": 5.4004, "step": 221800 }, { "epoch": 1.3377190337791391, "grad_norm": 0.3121047914028168, "learning_rate": 0.0002743247473874384, "loss": 4.9615, "step": 221850 }, { "epoch": 1.3380205255604731, "grad_norm": 1.778960943222046, "learning_rate": 0.0002743133610364906, "loss": 4.9168, "step": 221900 }, { "epoch": 1.3383220173418073, "grad_norm": 1.541167140007019, "learning_rate": 0.0002743019723977328, "loss": 4.9429, "step": 221950 }, { "epoch": 1.3386235091231413, "grad_norm": 1.5150057077407837, "learning_rate": 0.0002742905814713746, "loss": 5.1145, "step": 222000 }, { "epoch": 1.3386235091231413, "eval_loss": 5.608868598937988, "eval_runtime": 39.0357, "eval_samples_per_second": 13.116, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.211417713795409, "step": 222000 }, { "epoch": 1.3389250009044753, "grad_norm": 2.2495665550231934, "learning_rate": 0.00027427918825762563, "loss": 5.1647, "step": 222050 }, { "epoch": 1.3392264926858095, "grad_norm": 2.375173807144165, "learning_rate": 0.00027426779275669554, "loss": 4.9392, "step": 222100 }, { "epoch": 1.3395279844671435, "grad_norm": 2.197911500930786, "learning_rate": 0.0002742563949687941, "loss": 5.3144, "step": 222150 }, { "epoch": 1.3398294762484775, "grad_norm": 1.2435030937194824, "learning_rate": 0.00027424499489413104, "loss": 4.9406, "step": 222200 }, { "epoch": 1.3401309680298115, "grad_norm": 0.9185370802879333, "learning_rate": 0.0002742335925329161, "loss": 4.9693, "step": 222250 }, { "epoch": 1.3404324598111454, "grad_norm": 1.1491748094558716, "learning_rate": 0.00027422218788535925, "loss": 5.188, "step": 222300 }, { "epoch": 1.3407339515924797, "grad_norm": 2.0285942554473877, "learning_rate": 0.00027421078095167024, "loss": 4.6437, "step": 222350 }, { "epoch": 1.3410354433738136, "grad_norm": 1.2406710386276245, "learning_rate": 0.00027419937173205907, "loss": 5.3492, "step": 222400 }, { "epoch": 1.3413369351551476, "grad_norm": 1.4557982683181763, "learning_rate": 0.0002741879602267357, "loss": 5.1968, "step": 222450 }, { "epoch": 1.3416384269364818, "grad_norm": 0.320151686668396, "learning_rate": 0.0002741765464359102, "loss": 5.0101, "step": 222500 }, { "epoch": 1.3419399187178158, "grad_norm": 1.7074512243270874, "learning_rate": 0.00027416513035979256, "loss": 4.9268, "step": 222550 }, { "epoch": 1.3422414104991498, "grad_norm": 2.103299140930176, "learning_rate": 0.00027415371199859286, "loss": 5.1612, "step": 222600 }, { "epoch": 1.3425429022804838, "grad_norm": 2.085836887359619, "learning_rate": 0.00027414229135252127, "loss": 4.8116, "step": 222650 }, { "epoch": 1.3428443940618178, "grad_norm": 0.9852523803710938, "learning_rate": 0.000274130868421788, "loss": 4.9151, "step": 222700 }, { "epoch": 1.343145885843152, "grad_norm": 2.177987575531006, "learning_rate": 0.0002741194432066032, "loss": 4.9268, "step": 222750 }, { "epoch": 1.343447377624486, "grad_norm": 1.871716856956482, "learning_rate": 0.0002741080157071772, "loss": 4.9756, "step": 222800 }, { "epoch": 1.34374886940582, "grad_norm": 1.1638503074645996, "learning_rate": 0.00027409658592372034, "loss": 5.2525, "step": 222850 }, { "epoch": 1.3440503611871542, "grad_norm": 2.030867338180542, "learning_rate": 0.00027408515385644286, "loss": 5.0328, "step": 222900 }, { "epoch": 1.3443518529684881, "grad_norm": 1.2177934646606445, "learning_rate": 0.0002740737195055552, "loss": 5.1325, "step": 222950 }, { "epoch": 1.3446533447498221, "grad_norm": 1.7099088430404663, "learning_rate": 0.00027406228287126774, "loss": 4.7263, "step": 223000 }, { "epoch": 1.344954836531156, "grad_norm": 1.866519570350647, "learning_rate": 0.0002740508439537911, "loss": 4.9211, "step": 223050 }, { "epoch": 1.34525632831249, "grad_norm": 0.7742995023727417, "learning_rate": 0.00027403940275333563, "loss": 5.2202, "step": 223100 }, { "epoch": 1.3455578200938243, "grad_norm": 2.240826368331909, "learning_rate": 0.000274027959270112, "loss": 5.3248, "step": 223150 }, { "epoch": 1.3458593118751583, "grad_norm": 1.7205276489257812, "learning_rate": 0.0002740165135043308, "loss": 5.3793, "step": 223200 }, { "epoch": 1.3461608036564923, "grad_norm": 1.1256473064422607, "learning_rate": 0.00027400506545620265, "loss": 4.9229, "step": 223250 }, { "epoch": 1.3464622954378265, "grad_norm": 2.2352569103240967, "learning_rate": 0.0002739936151259382, "loss": 5.3979, "step": 223300 }, { "epoch": 1.3467637872191605, "grad_norm": 1.5294733047485352, "learning_rate": 0.0002739821625137482, "loss": 5.0691, "step": 223350 }, { "epoch": 1.3470652790004944, "grad_norm": 1.0314134359359741, "learning_rate": 0.0002739707076198435, "loss": 5.0613, "step": 223400 }, { "epoch": 1.3473667707818284, "grad_norm": 0.8808208703994751, "learning_rate": 0.0002739592504444348, "loss": 5.2427, "step": 223450 }, { "epoch": 1.3476682625631624, "grad_norm": 1.513940453529358, "learning_rate": 0.00027394779098773297, "loss": 5.1657, "step": 223500 }, { "epoch": 1.3479697543444966, "grad_norm": 1.1443023681640625, "learning_rate": 0.00027393632924994893, "loss": 4.7399, "step": 223550 }, { "epoch": 1.3482712461258306, "grad_norm": 1.4114012718200684, "learning_rate": 0.00027392486523129366, "loss": 5.2232, "step": 223600 }, { "epoch": 1.3485727379071646, "grad_norm": 1.409002423286438, "learning_rate": 0.0002739133989319781, "loss": 5.3163, "step": 223650 }, { "epoch": 1.3488742296884988, "grad_norm": 0.5802708864212036, "learning_rate": 0.00027390193035221324, "loss": 5.1699, "step": 223700 }, { "epoch": 1.3491757214698328, "grad_norm": 2.2356529235839844, "learning_rate": 0.0002738904594922101, "loss": 4.7334, "step": 223750 }, { "epoch": 1.3494772132511668, "grad_norm": 0.5436963438987732, "learning_rate": 0.00027387898635217997, "loss": 4.9567, "step": 223800 }, { "epoch": 1.3497787050325007, "grad_norm": 0.4188934564590454, "learning_rate": 0.0002738675109323339, "loss": 5.0743, "step": 223850 }, { "epoch": 1.3500801968138347, "grad_norm": 1.2224432229995728, "learning_rate": 0.0002738560332328829, "loss": 5.0355, "step": 223900 }, { "epoch": 1.350381688595169, "grad_norm": 1.2119958400726318, "learning_rate": 0.0002738445532540385, "loss": 5.3941, "step": 223950 }, { "epoch": 1.350683180376503, "grad_norm": 1.8555084466934204, "learning_rate": 0.0002738330709960118, "loss": 5.4983, "step": 224000 }, { "epoch": 1.350984672157837, "grad_norm": 1.5950047969818115, "learning_rate": 0.00027382158645901416, "loss": 5.1084, "step": 224050 }, { "epoch": 1.3512861639391711, "grad_norm": 1.1955785751342773, "learning_rate": 0.0002738100996432569, "loss": 5.3464, "step": 224100 }, { "epoch": 1.351587655720505, "grad_norm": 2.2475359439849854, "learning_rate": 0.00027379861054895147, "loss": 4.7757, "step": 224150 }, { "epoch": 1.351889147501839, "grad_norm": 0.5206892490386963, "learning_rate": 0.0002737871191763092, "loss": 5.0315, "step": 224200 }, { "epoch": 1.352190639283173, "grad_norm": 1.5442918539047241, "learning_rate": 0.0002737756255255417, "loss": 4.9885, "step": 224250 }, { "epoch": 1.352492131064507, "grad_norm": 1.9625293016433716, "learning_rate": 0.0002737641295968605, "loss": 5.3289, "step": 224300 }, { "epoch": 1.3527936228458413, "grad_norm": 2.385158061981201, "learning_rate": 0.0002737526313904771, "loss": 4.7494, "step": 224350 }, { "epoch": 1.3530951146271752, "grad_norm": 1.1699086427688599, "learning_rate": 0.0002737411309066031, "loss": 5.0797, "step": 224400 }, { "epoch": 1.3533966064085092, "grad_norm": 2.2201433181762695, "learning_rate": 0.00027372962814545016, "loss": 5.1377, "step": 224450 }, { "epoch": 1.3536980981898434, "grad_norm": 1.2916827201843262, "learning_rate": 0.00027371812310723003, "loss": 5.2832, "step": 224500 }, { "epoch": 1.3539995899711774, "grad_norm": 1.6151080131530762, "learning_rate": 0.0002737066157921543, "loss": 4.9634, "step": 224550 }, { "epoch": 1.3543010817525114, "grad_norm": 1.0283995866775513, "learning_rate": 0.00027369510620043496, "loss": 5.4501, "step": 224600 }, { "epoch": 1.3546025735338454, "grad_norm": 1.1117937564849854, "learning_rate": 0.0002736835943322837, "loss": 5.2774, "step": 224650 }, { "epoch": 1.3549040653151794, "grad_norm": 1.2438994646072388, "learning_rate": 0.00027367208018791233, "loss": 4.8626, "step": 224700 }, { "epoch": 1.3552055570965136, "grad_norm": 1.806861400604248, "learning_rate": 0.0002736605637675328, "loss": 5.4299, "step": 224750 }, { "epoch": 1.3555070488778476, "grad_norm": 1.2602382898330688, "learning_rate": 0.0002736490450713572, "loss": 5.3876, "step": 224800 }, { "epoch": 1.3558085406591815, "grad_norm": 1.594226360321045, "learning_rate": 0.00027363752409959724, "loss": 4.9951, "step": 224850 }, { "epoch": 1.3561100324405158, "grad_norm": 1.2926623821258545, "learning_rate": 0.00027362600085246514, "loss": 5.401, "step": 224900 }, { "epoch": 1.3564115242218497, "grad_norm": 1.5031996965408325, "learning_rate": 0.00027361447533017294, "loss": 5.1291, "step": 224950 }, { "epoch": 1.3567130160031837, "grad_norm": 1.5838762521743774, "learning_rate": 0.00027360294753293266, "loss": 5.4069, "step": 225000 }, { "epoch": 1.3567130160031837, "eval_loss": 5.56874942779541, "eval_runtime": 39.1113, "eval_samples_per_second": 13.091, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.249835831524486, "step": 225000 }, { "epoch": 1.357014507784518, "grad_norm": 2.5891590118408203, "learning_rate": 0.00027359141746095665, "loss": 4.7437, "step": 225050 }, { "epoch": 1.357315999565852, "grad_norm": 1.285713791847229, "learning_rate": 0.00027357988511445686, "loss": 4.7624, "step": 225100 }, { "epoch": 1.357617491347186, "grad_norm": 1.7137178182601929, "learning_rate": 0.0002735683504936457, "loss": 5.3264, "step": 225150 }, { "epoch": 1.3579189831285199, "grad_norm": 1.173270583152771, "learning_rate": 0.00027355681359873536, "loss": 4.943, "step": 225200 }, { "epoch": 1.3582204749098539, "grad_norm": 1.4593825340270996, "learning_rate": 0.0002735452744299382, "loss": 5.1719, "step": 225250 }, { "epoch": 1.358521966691188, "grad_norm": 1.5434519052505493, "learning_rate": 0.0002735337329874666, "loss": 5.5276, "step": 225300 }, { "epoch": 1.358823458472522, "grad_norm": 1.8469812870025635, "learning_rate": 0.0002735221892715329, "loss": 5.2503, "step": 225350 }, { "epoch": 1.359124950253856, "grad_norm": 0.8233856558799744, "learning_rate": 0.00027351064328234956, "loss": 4.8198, "step": 225400 }, { "epoch": 1.3594264420351903, "grad_norm": 1.1562913656234741, "learning_rate": 0.00027349909502012915, "loss": 5.1802, "step": 225450 }, { "epoch": 1.3597279338165242, "grad_norm": 1.5255217552185059, "learning_rate": 0.0002734875444850841, "loss": 5.2413, "step": 225500 }, { "epoch": 1.3600294255978582, "grad_norm": 1.4792274236679077, "learning_rate": 0.00027347599167742704, "loss": 5.462, "step": 225550 }, { "epoch": 1.3603309173791922, "grad_norm": 0.8384031653404236, "learning_rate": 0.00027346443659737057, "loss": 5.0774, "step": 225600 }, { "epoch": 1.3606324091605262, "grad_norm": 0.509813129901886, "learning_rate": 0.00027345287924512726, "loss": 5.3876, "step": 225650 }, { "epoch": 1.3609339009418604, "grad_norm": 1.3863600492477417, "learning_rate": 0.00027344131962091, "loss": 5.216, "step": 225700 }, { "epoch": 1.3612353927231944, "grad_norm": 1.2103321552276611, "learning_rate": 0.00027342975772493136, "loss": 5.0699, "step": 225750 }, { "epoch": 1.3615368845045284, "grad_norm": 1.0227593183517456, "learning_rate": 0.00027341819355740414, "loss": 5.1938, "step": 225800 }, { "epoch": 1.3618383762858626, "grad_norm": 1.393632411956787, "learning_rate": 0.00027340662711854125, "loss": 5.0105, "step": 225850 }, { "epoch": 1.3621398680671966, "grad_norm": 1.2776296138763428, "learning_rate": 0.00027339505840855545, "loss": 5.1017, "step": 225900 }, { "epoch": 1.3624413598485305, "grad_norm": 2.1416876316070557, "learning_rate": 0.00027338348742765974, "loss": 5.0677, "step": 225950 }, { "epoch": 1.3627428516298645, "grad_norm": 0.4556427299976349, "learning_rate": 0.000273371914176067, "loss": 4.6788, "step": 226000 }, { "epoch": 1.3630443434111985, "grad_norm": 1.867712140083313, "learning_rate": 0.0002733603386539902, "loss": 5.1332, "step": 226050 }, { "epoch": 1.3633458351925327, "grad_norm": 1.751389741897583, "learning_rate": 0.0002733487608616425, "loss": 4.9538, "step": 226100 }, { "epoch": 1.3636473269738667, "grad_norm": 1.3741674423217773, "learning_rate": 0.0002733371807992368, "loss": 4.9874, "step": 226150 }, { "epoch": 1.3639488187552007, "grad_norm": 0.6766153573989868, "learning_rate": 0.0002733255984669864, "loss": 4.6848, "step": 226200 }, { "epoch": 1.364250310536535, "grad_norm": 1.2321810722351074, "learning_rate": 0.0002733140138651043, "loss": 5.1724, "step": 226250 }, { "epoch": 1.3645518023178689, "grad_norm": 0.9886924028396606, "learning_rate": 0.0002733024269938037, "loss": 4.7917, "step": 226300 }, { "epoch": 1.3648532940992029, "grad_norm": 1.245564579963684, "learning_rate": 0.00027329083785329795, "loss": 4.8498, "step": 226350 }, { "epoch": 1.3651547858805368, "grad_norm": 1.3390930891036987, "learning_rate": 0.0002732792464438003, "loss": 4.9907, "step": 226400 }, { "epoch": 1.3654562776618708, "grad_norm": 1.173039197921753, "learning_rate": 0.000273267652765524, "loss": 5.4323, "step": 226450 }, { "epoch": 1.365757769443205, "grad_norm": 1.1895700693130493, "learning_rate": 0.00027325605681868245, "loss": 5.2544, "step": 226500 }, { "epoch": 1.366059261224539, "grad_norm": 2.077873468399048, "learning_rate": 0.00027324445860348913, "loss": 5.264, "step": 226550 }, { "epoch": 1.366360753005873, "grad_norm": 1.2215019464492798, "learning_rate": 0.00027323285812015733, "loss": 4.8378, "step": 226600 }, { "epoch": 1.3666622447872072, "grad_norm": 1.1318132877349854, "learning_rate": 0.0002732212553689007, "loss": 4.8714, "step": 226650 }, { "epoch": 1.3669637365685412, "grad_norm": 1.6368038654327393, "learning_rate": 0.00027320965034993277, "loss": 4.7768, "step": 226700 }, { "epoch": 1.3672652283498752, "grad_norm": 1.1214951276779175, "learning_rate": 0.00027319804306346696, "loss": 5.3849, "step": 226750 }, { "epoch": 1.3675667201312092, "grad_norm": 1.7120193243026733, "learning_rate": 0.000273186433509717, "loss": 5.3381, "step": 226800 }, { "epoch": 1.3678682119125432, "grad_norm": 0.5526238083839417, "learning_rate": 0.0002731748216888966, "loss": 5.3223, "step": 226850 }, { "epoch": 1.3681697036938774, "grad_norm": 2.0778722763061523, "learning_rate": 0.0002731632076012193, "loss": 5.1217, "step": 226900 }, { "epoch": 1.3684711954752113, "grad_norm": 1.508967399597168, "learning_rate": 0.000273151591246899, "loss": 5.2433, "step": 226950 }, { "epoch": 1.3687726872565453, "grad_norm": 1.6312127113342285, "learning_rate": 0.00027313997262614937, "loss": 5.5283, "step": 227000 }, { "epoch": 1.3690741790378795, "grad_norm": 1.421813726425171, "learning_rate": 0.00027312835173918425, "loss": 5.0538, "step": 227050 }, { "epoch": 1.3693756708192135, "grad_norm": 1.067888617515564, "learning_rate": 0.0002731167285862175, "loss": 4.6401, "step": 227100 }, { "epoch": 1.3696771626005475, "grad_norm": 0.8223515152931213, "learning_rate": 0.0002731051031674631, "loss": 5.1469, "step": 227150 }, { "epoch": 1.3699786543818815, "grad_norm": 0.9135720729827881, "learning_rate": 0.00027309347548313504, "loss": 5.0347, "step": 227200 }, { "epoch": 1.3702801461632155, "grad_norm": 1.963561773300171, "learning_rate": 0.00027308184553344713, "loss": 5.2222, "step": 227250 }, { "epoch": 1.3705816379445497, "grad_norm": 0.7976226806640625, "learning_rate": 0.0002730702133186136, "loss": 5.17, "step": 227300 }, { "epoch": 1.3708831297258837, "grad_norm": 2.7074880599975586, "learning_rate": 0.00027305857883884834, "loss": 5.3488, "step": 227350 }, { "epoch": 1.3711846215072176, "grad_norm": 1.3372770547866821, "learning_rate": 0.0002730469420943655, "loss": 4.9537, "step": 227400 }, { "epoch": 1.3714861132885519, "grad_norm": 1.324307918548584, "learning_rate": 0.00027303530308537936, "loss": 5.0685, "step": 227450 }, { "epoch": 1.3717876050698858, "grad_norm": 1.526253342628479, "learning_rate": 0.00027302366181210404, "loss": 5.371, "step": 227500 }, { "epoch": 1.3720890968512198, "grad_norm": 1.5173006057739258, "learning_rate": 0.00027301201827475383, "loss": 5.0247, "step": 227550 }, { "epoch": 1.3723905886325538, "grad_norm": 1.5463910102844238, "learning_rate": 0.0002730003724735429, "loss": 4.977, "step": 227600 }, { "epoch": 1.3726920804138878, "grad_norm": 1.090740442276001, "learning_rate": 0.0002729887244086857, "loss": 5.3176, "step": 227650 }, { "epoch": 1.372993572195222, "grad_norm": 0.9507352113723755, "learning_rate": 0.0002729770740803965, "loss": 5.0302, "step": 227700 }, { "epoch": 1.373295063976556, "grad_norm": 1.9688767194747925, "learning_rate": 0.00027296542148888976, "loss": 5.4637, "step": 227750 }, { "epoch": 1.37359655575789, "grad_norm": 1.477952480316162, "learning_rate": 0.0002729537666343799, "loss": 5.1356, "step": 227800 }, { "epoch": 1.3738980475392242, "grad_norm": 1.4989049434661865, "learning_rate": 0.0002729421095170814, "loss": 4.8509, "step": 227850 }, { "epoch": 1.3741995393205582, "grad_norm": 1.2812143564224243, "learning_rate": 0.0002729304501372089, "loss": 4.9145, "step": 227900 }, { "epoch": 1.3745010311018921, "grad_norm": 1.1842848062515259, "learning_rate": 0.0002729187884949769, "loss": 5.0153, "step": 227950 }, { "epoch": 1.3748025228832261, "grad_norm": 2.168867588043213, "learning_rate": 0.00027290712459059996, "loss": 4.9904, "step": 228000 }, { "epoch": 1.3748025228832261, "eval_loss": 5.580165863037109, "eval_runtime": 39.1056, "eval_samples_per_second": 13.093, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.267948893553318, "step": 228000 }, { "epoch": 1.3751040146645601, "grad_norm": 1.2212517261505127, "learning_rate": 0.0002728954584242929, "loss": 4.8456, "step": 228050 }, { "epoch": 1.3754055064458943, "grad_norm": 0.6016877293586731, "learning_rate": 0.00027288378999627014, "loss": 4.8265, "step": 228100 }, { "epoch": 1.3757069982272283, "grad_norm": 1.3743107318878174, "learning_rate": 0.0002728721193067467, "loss": 4.9097, "step": 228150 }, { "epoch": 1.3760084900085623, "grad_norm": 1.1500507593154907, "learning_rate": 0.0002728604463559372, "loss": 5.29, "step": 228200 }, { "epoch": 1.3763099817898965, "grad_norm": 1.5982612371444702, "learning_rate": 0.00027284877114405656, "loss": 4.7317, "step": 228250 }, { "epoch": 1.3766114735712305, "grad_norm": 1.6632741689682007, "learning_rate": 0.00027283709367131957, "loss": 5.2513, "step": 228300 }, { "epoch": 1.3769129653525645, "grad_norm": 1.7747865915298462, "learning_rate": 0.0002728254139379412, "loss": 4.9882, "step": 228350 }, { "epoch": 1.3772144571338985, "grad_norm": 2.0323755741119385, "learning_rate": 0.00027281373194413633, "loss": 5.2779, "step": 228400 }, { "epoch": 1.3775159489152324, "grad_norm": 1.6261186599731445, "learning_rate": 0.00027280204769012, "loss": 5.0368, "step": 228450 }, { "epoch": 1.3778174406965666, "grad_norm": 1.4064849615097046, "learning_rate": 0.0002727903611761072, "loss": 4.9575, "step": 228500 }, { "epoch": 1.3781189324779006, "grad_norm": 1.202348232269287, "learning_rate": 0.00027277867240231313, "loss": 4.9844, "step": 228550 }, { "epoch": 1.3784204242592346, "grad_norm": 1.6196060180664062, "learning_rate": 0.0002727669813689527, "loss": 5.4768, "step": 228600 }, { "epoch": 1.3787219160405688, "grad_norm": 1.3952529430389404, "learning_rate": 0.00027275528807624126, "loss": 4.7888, "step": 228650 }, { "epoch": 1.3790234078219028, "grad_norm": 1.2248295545578003, "learning_rate": 0.0002727435925243939, "loss": 5.2788, "step": 228700 }, { "epoch": 1.3793248996032368, "grad_norm": 1.1881588697433472, "learning_rate": 0.00027273189471362585, "loss": 5.0514, "step": 228750 }, { "epoch": 1.379626391384571, "grad_norm": 2.3359761238098145, "learning_rate": 0.00027272019464415247, "loss": 4.7189, "step": 228800 }, { "epoch": 1.379927883165905, "grad_norm": 1.1579822301864624, "learning_rate": 0.000272708492316189, "loss": 5.4049, "step": 228850 }, { "epoch": 1.380229374947239, "grad_norm": 1.6192387342453003, "learning_rate": 0.0002726967877299509, "loss": 5.0846, "step": 228900 }, { "epoch": 1.380530866728573, "grad_norm": 1.14531672000885, "learning_rate": 0.0002726850808856535, "loss": 5.1897, "step": 228950 }, { "epoch": 1.380832358509907, "grad_norm": 1.23465895652771, "learning_rate": 0.0002726733717835122, "loss": 5.2257, "step": 229000 }, { "epoch": 1.3811338502912411, "grad_norm": 1.800634503364563, "learning_rate": 0.0002726616604237426, "loss": 5.1501, "step": 229050 }, { "epoch": 1.3814353420725751, "grad_norm": 1.6334593296051025, "learning_rate": 0.00027264994680656015, "loss": 4.8557, "step": 229100 }, { "epoch": 1.381736833853909, "grad_norm": 1.2275217771530151, "learning_rate": 0.0002726382309321805, "loss": 5.3319, "step": 229150 }, { "epoch": 1.3820383256352433, "grad_norm": 1.3497613668441772, "learning_rate": 0.00027262651280081927, "loss": 5.169, "step": 229200 }, { "epoch": 1.3823398174165773, "grad_norm": 1.2381441593170166, "learning_rate": 0.000272614792412692, "loss": 5.0958, "step": 229250 }, { "epoch": 1.3826413091979113, "grad_norm": 0.6785418391227722, "learning_rate": 0.0002726030697680145, "loss": 4.9821, "step": 229300 }, { "epoch": 1.3829428009792453, "grad_norm": 1.1575241088867188, "learning_rate": 0.00027259134486700245, "loss": 5.1828, "step": 229350 }, { "epoch": 1.3832442927605793, "grad_norm": 1.5900882482528687, "learning_rate": 0.00027257961770987165, "loss": 5.0506, "step": 229400 }, { "epoch": 1.3835457845419135, "grad_norm": 1.785941481590271, "learning_rate": 0.00027256788829683787, "loss": 5.0784, "step": 229450 }, { "epoch": 1.3838472763232474, "grad_norm": 2.096524238586426, "learning_rate": 0.0002725561566281171, "loss": 5.1466, "step": 229500 }, { "epoch": 1.3841487681045814, "grad_norm": 1.9149900674819946, "learning_rate": 0.0002725444227039251, "loss": 4.9191, "step": 229550 }, { "epoch": 1.3844502598859156, "grad_norm": 1.0957324504852295, "learning_rate": 0.00027253268652447786, "loss": 5.1391, "step": 229600 }, { "epoch": 1.3847517516672496, "grad_norm": 1.291882038116455, "learning_rate": 0.00027252094808999144, "loss": 5.2627, "step": 229650 }, { "epoch": 1.3850532434485836, "grad_norm": 2.295942783355713, "learning_rate": 0.00027250920740068185, "loss": 5.1301, "step": 229700 }, { "epoch": 1.3853547352299176, "grad_norm": 1.2021843194961548, "learning_rate": 0.00027249746445676506, "loss": 4.8188, "step": 229750 }, { "epoch": 1.3856562270112516, "grad_norm": 2.460407018661499, "learning_rate": 0.0002724857192584573, "loss": 5.1546, "step": 229800 }, { "epoch": 1.3859577187925858, "grad_norm": 1.1791561841964722, "learning_rate": 0.00027247397180597463, "loss": 5.2529, "step": 229850 }, { "epoch": 1.3862592105739198, "grad_norm": 2.8335607051849365, "learning_rate": 0.0002724622220995333, "loss": 5.2265, "step": 229900 }, { "epoch": 1.3865607023552537, "grad_norm": 2.1263649463653564, "learning_rate": 0.0002724504701393495, "loss": 4.9241, "step": 229950 }, { "epoch": 1.386862194136588, "grad_norm": 1.5323045253753662, "learning_rate": 0.00027243871592563964, "loss": 4.9491, "step": 230000 }, { "epoch": 1.387163685917922, "grad_norm": 1.5458924770355225, "learning_rate": 0.00027242695945861985, "loss": 5.1032, "step": 230050 }, { "epoch": 1.387465177699256, "grad_norm": 2.004079818725586, "learning_rate": 0.0002724152007385067, "loss": 5.5732, "step": 230100 }, { "epoch": 1.38776666948059, "grad_norm": 0.9832176566123962, "learning_rate": 0.0002724034397655163, "loss": 5.2309, "step": 230150 }, { "epoch": 1.388068161261924, "grad_norm": 1.5866488218307495, "learning_rate": 0.00027239167653986544, "loss": 5.1628, "step": 230200 }, { "epoch": 1.388369653043258, "grad_norm": 2.369722843170166, "learning_rate": 0.00027237991106177035, "loss": 5.1852, "step": 230250 }, { "epoch": 1.388671144824592, "grad_norm": 1.1799594163894653, "learning_rate": 0.00027236814333144767, "loss": 5.1685, "step": 230300 }, { "epoch": 1.388972636605926, "grad_norm": 1.8031679391860962, "learning_rate": 0.0002723563733491139, "loss": 4.9467, "step": 230350 }, { "epoch": 1.3892741283872603, "grad_norm": 1.049599289894104, "learning_rate": 0.0002723446011149857, "loss": 5.0261, "step": 230400 }, { "epoch": 1.3895756201685943, "grad_norm": 0.5818116664886475, "learning_rate": 0.00027233282662927975, "loss": 5.2023, "step": 230450 }, { "epoch": 1.3898771119499282, "grad_norm": 1.8406929969787598, "learning_rate": 0.00027232104989221266, "loss": 5.3083, "step": 230500 }, { "epoch": 1.3901786037312622, "grad_norm": 0.5533864498138428, "learning_rate": 0.00027230927090400126, "loss": 5.2712, "step": 230550 }, { "epoch": 1.3904800955125962, "grad_norm": 1.276361346244812, "learning_rate": 0.0002722974896648622, "loss": 4.6551, "step": 230600 }, { "epoch": 1.3907815872939304, "grad_norm": 1.656570315361023, "learning_rate": 0.00027228570617501244, "loss": 4.9506, "step": 230650 }, { "epoch": 1.3910830790752644, "grad_norm": 1.14115309715271, "learning_rate": 0.0002722739204346687, "loss": 5.5267, "step": 230700 }, { "epoch": 1.3913845708565984, "grad_norm": 1.7634910345077515, "learning_rate": 0.00027226213244404803, "loss": 5.6986, "step": 230750 }, { "epoch": 1.3916860626379326, "grad_norm": 1.7691346406936646, "learning_rate": 0.0002722503422033673, "loss": 5.0267, "step": 230800 }, { "epoch": 1.3919875544192666, "grad_norm": 1.2971692085266113, "learning_rate": 0.0002722385497128434, "loss": 5.196, "step": 230850 }, { "epoch": 1.3922890462006006, "grad_norm": 1.2759850025177002, "learning_rate": 0.0002722267549726935, "loss": 4.8819, "step": 230900 }, { "epoch": 1.3925905379819346, "grad_norm": 1.8966647386550903, "learning_rate": 0.00027221495798313456, "loss": 5.268, "step": 230950 }, { "epoch": 1.3928920297632685, "grad_norm": 1.311041235923767, "learning_rate": 0.00027220315874438374, "loss": 5.0564, "step": 231000 }, { "epoch": 1.3928920297632685, "eval_loss": 5.574944972991943, "eval_runtime": 38.9317, "eval_samples_per_second": 13.151, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.260835861584848, "step": 231000 }, { "epoch": 1.3931935215446027, "grad_norm": 1.3372092247009277, "learning_rate": 0.00027219135725665815, "loss": 5.1887, "step": 231050 }, { "epoch": 1.3934950133259367, "grad_norm": 1.836166501045227, "learning_rate": 0.00027217955352017506, "loss": 5.0785, "step": 231100 }, { "epoch": 1.3937965051072707, "grad_norm": 1.6164129972457886, "learning_rate": 0.0002721677475351516, "loss": 5.2779, "step": 231150 }, { "epoch": 1.394097996888605, "grad_norm": 2.2427315711975098, "learning_rate": 0.00027215593930180517, "loss": 4.953, "step": 231200 }, { "epoch": 1.394399488669939, "grad_norm": 1.1873877048492432, "learning_rate": 0.0002721441288203529, "loss": 5.2276, "step": 231250 }, { "epoch": 1.3947009804512729, "grad_norm": 2.049828290939331, "learning_rate": 0.00027213231609101227, "loss": 5.363, "step": 231300 }, { "epoch": 1.3950024722326069, "grad_norm": 1.5927152633666992, "learning_rate": 0.0002721205011140007, "loss": 4.9915, "step": 231350 }, { "epoch": 1.3953039640139409, "grad_norm": 1.3993254899978638, "learning_rate": 0.00027210868388953557, "loss": 4.7556, "step": 231400 }, { "epoch": 1.395605455795275, "grad_norm": 1.1285396814346313, "learning_rate": 0.00027209686441783434, "loss": 5.3752, "step": 231450 }, { "epoch": 1.395906947576609, "grad_norm": 1.8958381414413452, "learning_rate": 0.00027208504269911457, "loss": 5.2844, "step": 231500 }, { "epoch": 1.396208439357943, "grad_norm": 1.1539057493209839, "learning_rate": 0.00027207321873359385, "loss": 5.3716, "step": 231550 }, { "epoch": 1.3965099311392772, "grad_norm": 1.4324142932891846, "learning_rate": 0.0002720613925214897, "loss": 4.9566, "step": 231600 }, { "epoch": 1.3968114229206112, "grad_norm": 1.4493651390075684, "learning_rate": 0.0002720495640630199, "loss": 5.1303, "step": 231650 }, { "epoch": 1.3971129147019452, "grad_norm": 1.632641315460205, "learning_rate": 0.00027203773335840193, "loss": 4.9626, "step": 231700 }, { "epoch": 1.3974144064832792, "grad_norm": 1.128981113433838, "learning_rate": 0.0002720259004078537, "loss": 5.3307, "step": 231750 }, { "epoch": 1.3977158982646132, "grad_norm": 1.640318512916565, "learning_rate": 0.0002720140652115929, "loss": 5.0423, "step": 231800 }, { "epoch": 1.3980173900459474, "grad_norm": 0.826790988445282, "learning_rate": 0.0002720022277698373, "loss": 5.1357, "step": 231850 }, { "epoch": 1.3983188818272814, "grad_norm": 1.187997817993164, "learning_rate": 0.0002719903880828049, "loss": 5.0923, "step": 231900 }, { "epoch": 1.3986203736086154, "grad_norm": 0.7209887504577637, "learning_rate": 0.0002719785461507135, "loss": 5.1531, "step": 231950 }, { "epoch": 1.3989218653899496, "grad_norm": 1.9956411123275757, "learning_rate": 0.0002719667019737809, "loss": 4.7614, "step": 232000 }, { "epoch": 1.3992233571712835, "grad_norm": 1.3589656352996826, "learning_rate": 0.00027195485555222534, "loss": 4.8908, "step": 232050 }, { "epoch": 1.3995248489526175, "grad_norm": 1.1657987833023071, "learning_rate": 0.00027194300688626464, "loss": 5.1393, "step": 232100 }, { "epoch": 1.3998263407339515, "grad_norm": 1.5473543405532837, "learning_rate": 0.00027193115597611695, "loss": 5.1604, "step": 232150 }, { "epoch": 1.4001278325152855, "grad_norm": 1.27736496925354, "learning_rate": 0.00027191930282200035, "loss": 4.9674, "step": 232200 }, { "epoch": 1.4004293242966197, "grad_norm": 1.6174359321594238, "learning_rate": 0.00027190744742413295, "loss": 5.2463, "step": 232250 }, { "epoch": 1.4007308160779537, "grad_norm": 1.0776398181915283, "learning_rate": 0.00027189558978273293, "loss": 5.0602, "step": 232300 }, { "epoch": 1.4010323078592877, "grad_norm": 0.8911868333816528, "learning_rate": 0.00027188372989801854, "loss": 5.285, "step": 232350 }, { "epoch": 1.4013337996406219, "grad_norm": 1.1827452182769775, "learning_rate": 0.00027187186777020804, "loss": 4.662, "step": 232400 }, { "epoch": 1.4016352914219559, "grad_norm": 0.8314592242240906, "learning_rate": 0.0002718600033995197, "loss": 5.052, "step": 232450 }, { "epoch": 1.4019367832032898, "grad_norm": 0.6103248000144958, "learning_rate": 0.000271848136786172, "loss": 5.1245, "step": 232500 }, { "epoch": 1.402238274984624, "grad_norm": 0.7901971340179443, "learning_rate": 0.00027183626793038314, "loss": 4.9855, "step": 232550 }, { "epoch": 1.402539766765958, "grad_norm": 1.5353739261627197, "learning_rate": 0.0002718243968323717, "loss": 5.1039, "step": 232600 }, { "epoch": 1.402841258547292, "grad_norm": 2.33150315284729, "learning_rate": 0.00027181252349235604, "loss": 5.4579, "step": 232650 }, { "epoch": 1.403142750328626, "grad_norm": 1.6802843809127808, "learning_rate": 0.00027180064791055474, "loss": 5.2302, "step": 232700 }, { "epoch": 1.40344424210996, "grad_norm": 1.140775203704834, "learning_rate": 0.00027178877008718637, "loss": 5.1704, "step": 232750 }, { "epoch": 1.4037457338912942, "grad_norm": 1.7116297483444214, "learning_rate": 0.00027177689002246946, "loss": 5.4243, "step": 232800 }, { "epoch": 1.4040472256726282, "grad_norm": 1.849892497062683, "learning_rate": 0.0002717650077166226, "loss": 5.0124, "step": 232850 }, { "epoch": 1.4043487174539622, "grad_norm": 1.4090533256530762, "learning_rate": 0.0002717531231698647, "loss": 5.2613, "step": 232900 }, { "epoch": 1.4046502092352964, "grad_norm": 1.2294495105743408, "learning_rate": 0.00027174123638241417, "loss": 4.9962, "step": 232950 }, { "epoch": 1.4049517010166304, "grad_norm": 1.500586986541748, "learning_rate": 0.00027172934735448997, "loss": 4.9824, "step": 233000 }, { "epoch": 1.4052531927979643, "grad_norm": 1.088972568511963, "learning_rate": 0.00027171745608631085, "loss": 4.8809, "step": 233050 }, { "epoch": 1.4055546845792983, "grad_norm": 1.281103253364563, "learning_rate": 0.00027170556257809567, "loss": 5.1806, "step": 233100 }, { "epoch": 1.4058561763606323, "grad_norm": 1.7952821254730225, "learning_rate": 0.00027169366683006327, "loss": 5.2815, "step": 233150 }, { "epoch": 1.4061576681419665, "grad_norm": 1.4695031642913818, "learning_rate": 0.0002716817688424326, "loss": 5.7089, "step": 233200 }, { "epoch": 1.4064591599233005, "grad_norm": 1.5663745403289795, "learning_rate": 0.00027166986861542263, "loss": 4.8446, "step": 233250 }, { "epoch": 1.4067606517046345, "grad_norm": 1.1907685995101929, "learning_rate": 0.00027165796614925236, "loss": 5.2574, "step": 233300 }, { "epoch": 1.4070621434859687, "grad_norm": 1.472060203552246, "learning_rate": 0.00027164606144414085, "loss": 4.8471, "step": 233350 }, { "epoch": 1.4073636352673027, "grad_norm": 1.3163859844207764, "learning_rate": 0.00027163415450030717, "loss": 5.1928, "step": 233400 }, { "epoch": 1.4076651270486367, "grad_norm": 1.6738356351852417, "learning_rate": 0.00027162224531797043, "loss": 5.1101, "step": 233450 }, { "epoch": 1.4079666188299707, "grad_norm": 1.6261392831802368, "learning_rate": 0.00027161033389734986, "loss": 4.9713, "step": 233500 }, { "epoch": 1.4082681106113046, "grad_norm": 1.6563637256622314, "learning_rate": 0.0002715984202386647, "loss": 4.9689, "step": 233550 }, { "epoch": 1.4085696023926388, "grad_norm": 1.6479169130325317, "learning_rate": 0.00027158650434213407, "loss": 5.1684, "step": 233600 }, { "epoch": 1.4088710941739728, "grad_norm": 0.8222907185554504, "learning_rate": 0.00027157458620797735, "loss": 5.43, "step": 233650 }, { "epoch": 1.4091725859553068, "grad_norm": 1.2542953491210938, "learning_rate": 0.0002715626658364139, "loss": 4.7891, "step": 233700 }, { "epoch": 1.409474077736641, "grad_norm": 1.3205424547195435, "learning_rate": 0.00027155074322766297, "loss": 5.1103, "step": 233750 }, { "epoch": 1.409775569517975, "grad_norm": 1.4480654001235962, "learning_rate": 0.0002715388183819441, "loss": 5.002, "step": 233800 }, { "epoch": 1.410077061299309, "grad_norm": 2.0871829986572266, "learning_rate": 0.0002715268912994768, "loss": 5.1106, "step": 233850 }, { "epoch": 1.410378553080643, "grad_norm": 1.7734838724136353, "learning_rate": 0.0002715149619804805, "loss": 4.9916, "step": 233900 }, { "epoch": 1.410680044861977, "grad_norm": 0.49278464913368225, "learning_rate": 0.0002715030304251747, "loss": 4.9063, "step": 233950 }, { "epoch": 1.4109815366433112, "grad_norm": 0.22115002572536469, "learning_rate": 0.00027149109663377895, "loss": 5.6571, "step": 234000 }, { "epoch": 1.4109815366433112, "eval_loss": 5.584138870239258, "eval_runtime": 39.1926, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 7.286531894303887, "step": 234000 }, { "epoch": 1.4112830284246451, "grad_norm": 2.6118078231811523, "learning_rate": 0.000271479160606513, "loss": 5.1309, "step": 234050 }, { "epoch": 1.4115845202059791, "grad_norm": 2.2133851051330566, "learning_rate": 0.0002714672223435965, "loss": 5.2792, "step": 234100 }, { "epoch": 1.4118860119873133, "grad_norm": 1.2023875713348389, "learning_rate": 0.00027145528184524905, "loss": 4.9921, "step": 234150 }, { "epoch": 1.4121875037686473, "grad_norm": 1.4892524480819702, "learning_rate": 0.0002714433391116905, "loss": 5.6185, "step": 234200 }, { "epoch": 1.4124889955499813, "grad_norm": 0.5274794697761536, "learning_rate": 0.0002714313941431406, "loss": 4.6174, "step": 234250 }, { "epoch": 1.4127904873313153, "grad_norm": 1.8133339881896973, "learning_rate": 0.00027141944693981915, "loss": 4.8798, "step": 234300 }, { "epoch": 1.4130919791126493, "grad_norm": 2.1353402137756348, "learning_rate": 0.0002714074975019461, "loss": 5.1752, "step": 234350 }, { "epoch": 1.4133934708939835, "grad_norm": 1.603299856185913, "learning_rate": 0.0002713955458297413, "loss": 4.4899, "step": 234400 }, { "epoch": 1.4136949626753175, "grad_norm": 1.3896679878234863, "learning_rate": 0.0002713835919234247, "loss": 5.4869, "step": 234450 }, { "epoch": 1.4139964544566515, "grad_norm": 1.332124948501587, "learning_rate": 0.00027137163578321634, "loss": 5.5821, "step": 234500 }, { "epoch": 1.4142979462379857, "grad_norm": 1.4184950590133667, "learning_rate": 0.00027135967740933624, "loss": 5.2081, "step": 234550 }, { "epoch": 1.4145994380193196, "grad_norm": 0.7718331813812256, "learning_rate": 0.00027134771680200443, "loss": 5.1522, "step": 234600 }, { "epoch": 1.4149009298006536, "grad_norm": 1.2056747674942017, "learning_rate": 0.00027133575396144107, "loss": 5.2558, "step": 234650 }, { "epoch": 1.4152024215819876, "grad_norm": 1.4313381910324097, "learning_rate": 0.00027132378888786635, "loss": 5.4113, "step": 234700 }, { "epoch": 1.4155039133633216, "grad_norm": 1.1928551197052002, "learning_rate": 0.00027131182158150036, "loss": 5.4563, "step": 234750 }, { "epoch": 1.4158054051446558, "grad_norm": 1.4387494325637817, "learning_rate": 0.00027129985204256353, "loss": 5.3636, "step": 234800 }, { "epoch": 1.4161068969259898, "grad_norm": 1.4097894430160522, "learning_rate": 0.00027128788027127595, "loss": 5.0937, "step": 234850 }, { "epoch": 1.4164083887073238, "grad_norm": 0.5675707459449768, "learning_rate": 0.000271275906267858, "loss": 5.3708, "step": 234900 }, { "epoch": 1.416709880488658, "grad_norm": 1.7506314516067505, "learning_rate": 0.0002712639300325301, "loss": 4.9062, "step": 234950 }, { "epoch": 1.417011372269992, "grad_norm": 1.7729347944259644, "learning_rate": 0.00027125195156551266, "loss": 5.2542, "step": 235000 }, { "epoch": 1.417312864051326, "grad_norm": 1.7119109630584717, "learning_rate": 0.00027123997086702604, "loss": 5.4773, "step": 235050 }, { "epoch": 1.41761435583266, "grad_norm": 1.2563586235046387, "learning_rate": 0.00027122798793729076, "loss": 5.2477, "step": 235100 }, { "epoch": 1.417915847613994, "grad_norm": 3.088099479675293, "learning_rate": 0.0002712160027765274, "loss": 4.8266, "step": 235150 }, { "epoch": 1.4182173393953281, "grad_norm": 0.343490332365036, "learning_rate": 0.0002712040153849565, "loss": 4.7137, "step": 235200 }, { "epoch": 1.418518831176662, "grad_norm": 1.1629756689071655, "learning_rate": 0.0002711920257627986, "loss": 4.9979, "step": 235250 }, { "epoch": 1.418820322957996, "grad_norm": 2.0401532649993896, "learning_rate": 0.0002711800339102745, "loss": 5.2998, "step": 235300 }, { "epoch": 1.4191218147393303, "grad_norm": 1.248362421989441, "learning_rate": 0.00027116803982760477, "loss": 5.5458, "step": 235350 }, { "epoch": 1.4194233065206643, "grad_norm": 0.6547353863716125, "learning_rate": 0.00027115604351501016, "loss": 5.1921, "step": 235400 }, { "epoch": 1.4197247983019983, "grad_norm": 1.2520325183868408, "learning_rate": 0.00027114404497271145, "loss": 5.4289, "step": 235450 }, { "epoch": 1.4200262900833323, "grad_norm": 0.7399041652679443, "learning_rate": 0.00027113204420092953, "loss": 4.7183, "step": 235500 }, { "epoch": 1.4203277818646662, "grad_norm": 1.4999523162841797, "learning_rate": 0.0002711200411998852, "loss": 5.7017, "step": 235550 }, { "epoch": 1.4206292736460004, "grad_norm": 1.1563944816589355, "learning_rate": 0.00027110803596979934, "loss": 5.0102, "step": 235600 }, { "epoch": 1.4209307654273344, "grad_norm": 1.6046361923217773, "learning_rate": 0.00027109602851089294, "loss": 5.1513, "step": 235650 }, { "epoch": 1.4212322572086684, "grad_norm": 1.7668917179107666, "learning_rate": 0.0002710840188233869, "loss": 5.2999, "step": 235700 }, { "epoch": 1.4215337489900026, "grad_norm": 1.157639503479004, "learning_rate": 0.00027107200690750226, "loss": 5.5743, "step": 235750 }, { "epoch": 1.4218352407713366, "grad_norm": 2.198112726211548, "learning_rate": 0.0002710599927634602, "loss": 5.1759, "step": 235800 }, { "epoch": 1.4221367325526706, "grad_norm": 1.3193752765655518, "learning_rate": 0.00027104797639148167, "loss": 5.2759, "step": 235850 }, { "epoch": 1.4224382243340046, "grad_norm": 1.186824083328247, "learning_rate": 0.000271035957791788, "loss": 4.6263, "step": 235900 }, { "epoch": 1.4227397161153386, "grad_norm": 1.1671233177185059, "learning_rate": 0.0002710239369646001, "loss": 5.3297, "step": 235950 }, { "epoch": 1.4230412078966728, "grad_norm": 1.460819125175476, "learning_rate": 0.0002710119139101394, "loss": 4.8, "step": 236000 }, { "epoch": 1.4233426996780068, "grad_norm": 1.2343558073043823, "learning_rate": 0.00027099988862862713, "loss": 5.1471, "step": 236050 }, { "epoch": 1.4236441914593407, "grad_norm": 1.546074390411377, "learning_rate": 0.0002709878611202846, "loss": 5.4843, "step": 236100 }, { "epoch": 1.423945683240675, "grad_norm": 1.2679699659347534, "learning_rate": 0.0002709758313853331, "loss": 5.2198, "step": 236150 }, { "epoch": 1.424247175022009, "grad_norm": 1.7903715372085571, "learning_rate": 0.0002709637994239941, "loss": 4.5887, "step": 236200 }, { "epoch": 1.424548666803343, "grad_norm": 1.6275932788848877, "learning_rate": 0.00027095176523648905, "loss": 5.6363, "step": 236250 }, { "epoch": 1.4248501585846771, "grad_norm": 1.4066789150238037, "learning_rate": 0.0002709397288230393, "loss": 5.0671, "step": 236300 }, { "epoch": 1.425151650366011, "grad_norm": 1.9292576313018799, "learning_rate": 0.0002709276901838664, "loss": 5.066, "step": 236350 }, { "epoch": 1.425453142147345, "grad_norm": 0.7973955869674683, "learning_rate": 0.000270915649319192, "loss": 5.5852, "step": 236400 }, { "epoch": 1.425754633928679, "grad_norm": 1.4572886228561401, "learning_rate": 0.0002709036062292376, "loss": 5.0817, "step": 236450 }, { "epoch": 1.426056125710013, "grad_norm": 1.1486443281173706, "learning_rate": 0.00027089156091422484, "loss": 5.0714, "step": 236500 }, { "epoch": 1.4263576174913473, "grad_norm": 1.3721927404403687, "learning_rate": 0.0002708795133743755, "loss": 4.8977, "step": 236550 }, { "epoch": 1.4266591092726812, "grad_norm": 1.3158771991729736, "learning_rate": 0.00027086746360991114, "loss": 5.0318, "step": 236600 }, { "epoch": 1.4269606010540152, "grad_norm": 1.7410911321640015, "learning_rate": 0.00027085541162105367, "loss": 5.1216, "step": 236650 }, { "epoch": 1.4272620928353494, "grad_norm": 1.982667326927185, "learning_rate": 0.0002708433574080248, "loss": 5.147, "step": 236700 }, { "epoch": 1.4275635846166834, "grad_norm": 1.0664432048797607, "learning_rate": 0.00027083130097104634, "loss": 4.7606, "step": 236750 }, { "epoch": 1.4278650763980174, "grad_norm": 1.3480942249298096, "learning_rate": 0.00027081924231034025, "loss": 4.8236, "step": 236800 }, { "epoch": 1.4281665681793514, "grad_norm": 1.3317972421646118, "learning_rate": 0.00027080718142612843, "loss": 5.6856, "step": 236850 }, { "epoch": 1.4284680599606854, "grad_norm": 1.0103461742401123, "learning_rate": 0.00027079511831863283, "loss": 5.4037, "step": 236900 }, { "epoch": 1.4287695517420196, "grad_norm": 1.058576226234436, "learning_rate": 0.0002707830529880755, "loss": 4.9627, "step": 236950 }, { "epoch": 1.4290710435233536, "grad_norm": 1.6707446575164795, "learning_rate": 0.0002707709854346784, "loss": 5.1444, "step": 237000 }, { "epoch": 1.4290710435233536, "eval_loss": 5.569713592529297, "eval_runtime": 39.1172, "eval_samples_per_second": 13.089, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.237600077956082, "step": 237000 }, { "epoch": 1.4293725353046876, "grad_norm": 1.1449854373931885, "learning_rate": 0.0002707589156586637, "loss": 5.5421, "step": 237050 }, { "epoch": 1.4296740270860218, "grad_norm": 1.3228200674057007, "learning_rate": 0.0002707468436602534, "loss": 4.7756, "step": 237100 }, { "epoch": 1.4299755188673557, "grad_norm": 1.0266494750976562, "learning_rate": 0.00027073476943966987, "loss": 5.384, "step": 237150 }, { "epoch": 1.4302770106486897, "grad_norm": 1.8663108348846436, "learning_rate": 0.00027072269299713514, "loss": 4.7253, "step": 237200 }, { "epoch": 1.4305785024300237, "grad_norm": 1.1819754838943481, "learning_rate": 0.0002707106143328716, "loss": 5.1053, "step": 237250 }, { "epoch": 1.4308799942113577, "grad_norm": 1.52288818359375, "learning_rate": 0.00027069853344710137, "loss": 5.1192, "step": 237300 }, { "epoch": 1.431181485992692, "grad_norm": 2.077826738357544, "learning_rate": 0.0002706864503400469, "loss": 5.2214, "step": 237350 }, { "epoch": 1.4314829777740259, "grad_norm": 1.224460482597351, "learning_rate": 0.00027067436501193056, "loss": 5.1251, "step": 237400 }, { "epoch": 1.4317844695553599, "grad_norm": 1.3902413845062256, "learning_rate": 0.00027066227746297476, "loss": 4.9099, "step": 237450 }, { "epoch": 1.432085961336694, "grad_norm": 2.0132710933685303, "learning_rate": 0.0002706501876934019, "loss": 5.437, "step": 237500 }, { "epoch": 1.432387453118028, "grad_norm": 1.6096925735473633, "learning_rate": 0.00027063809570343456, "loss": 5.2374, "step": 237550 }, { "epoch": 1.432688944899362, "grad_norm": 1.1016671657562256, "learning_rate": 0.00027062600149329523, "loss": 4.4528, "step": 237600 }, { "epoch": 1.432990436680696, "grad_norm": 1.8973363637924194, "learning_rate": 0.0002706139050632064, "loss": 5.232, "step": 237650 }, { "epoch": 1.43329192846203, "grad_norm": 0.32655489444732666, "learning_rate": 0.00027060180641339085, "loss": 4.9686, "step": 237700 }, { "epoch": 1.4335934202433642, "grad_norm": 1.6391537189483643, "learning_rate": 0.00027058970554407115, "loss": 5.4246, "step": 237750 }, { "epoch": 1.4338949120246982, "grad_norm": 1.0676230192184448, "learning_rate": 0.00027057760245547, "loss": 5.0062, "step": 237800 }, { "epoch": 1.4341964038060322, "grad_norm": 1.5726439952850342, "learning_rate": 0.0002705654971478101, "loss": 5.1948, "step": 237850 }, { "epoch": 1.4344978955873664, "grad_norm": 1.0414689779281616, "learning_rate": 0.00027055338962131436, "loss": 5.3534, "step": 237900 }, { "epoch": 1.4347993873687004, "grad_norm": 1.1510307788848877, "learning_rate": 0.0002705412798762055, "loss": 4.9025, "step": 237950 }, { "epoch": 1.4351008791500344, "grad_norm": 1.129345178604126, "learning_rate": 0.0002705291679127064, "loss": 4.6327, "step": 238000 }, { "epoch": 1.4354023709313684, "grad_norm": 1.8104513883590698, "learning_rate": 0.00027051705373104, "loss": 5.079, "step": 238050 }, { "epoch": 1.4357038627127023, "grad_norm": 1.0888478755950928, "learning_rate": 0.0002705049373314292, "loss": 5.3422, "step": 238100 }, { "epoch": 1.4360053544940365, "grad_norm": 2.111921787261963, "learning_rate": 0.00027049281871409695, "loss": 5.2547, "step": 238150 }, { "epoch": 1.4363068462753705, "grad_norm": 0.5491183400154114, "learning_rate": 0.00027048069787926634, "loss": 5.1807, "step": 238200 }, { "epoch": 1.4366083380567045, "grad_norm": 1.188043236732483, "learning_rate": 0.0002704685748271604, "loss": 5.182, "step": 238250 }, { "epoch": 1.4369098298380387, "grad_norm": 1.1078566312789917, "learning_rate": 0.00027045644955800234, "loss": 5.0793, "step": 238300 }, { "epoch": 1.4372113216193727, "grad_norm": 1.700154423713684, "learning_rate": 0.00027044432207201516, "loss": 4.5927, "step": 238350 }, { "epoch": 1.4375128134007067, "grad_norm": 1.5624289512634277, "learning_rate": 0.0002704321923694222, "loss": 5.2993, "step": 238400 }, { "epoch": 1.4378143051820407, "grad_norm": 1.6166925430297852, "learning_rate": 0.00027042006045044646, "loss": 5.0832, "step": 238450 }, { "epoch": 1.4381157969633747, "grad_norm": 1.7678762674331665, "learning_rate": 0.00027040792631531147, "loss": 4.7938, "step": 238500 }, { "epoch": 1.4384172887447089, "grad_norm": 1.1225738525390625, "learning_rate": 0.0002703957899642404, "loss": 5.1469, "step": 238550 }, { "epoch": 1.4387187805260429, "grad_norm": 1.2229835987091064, "learning_rate": 0.00027038365139745666, "loss": 4.6582, "step": 238600 }, { "epoch": 1.4390202723073768, "grad_norm": 1.2372584342956543, "learning_rate": 0.00027037151061518353, "loss": 4.7268, "step": 238650 }, { "epoch": 1.439321764088711, "grad_norm": 1.1857270002365112, "learning_rate": 0.00027035936761764457, "loss": 5.4136, "step": 238700 }, { "epoch": 1.439623255870045, "grad_norm": 1.4288828372955322, "learning_rate": 0.0002703472224050632, "loss": 4.9934, "step": 238750 }, { "epoch": 1.439924747651379, "grad_norm": 1.2200862169265747, "learning_rate": 0.00027033507497766304, "loss": 4.9026, "step": 238800 }, { "epoch": 1.440226239432713, "grad_norm": 1.0155038833618164, "learning_rate": 0.00027032292533566744, "loss": 4.9131, "step": 238850 }, { "epoch": 1.440527731214047, "grad_norm": 1.2187587022781372, "learning_rate": 0.00027031077347930015, "loss": 5.3887, "step": 238900 }, { "epoch": 1.4408292229953812, "grad_norm": 1.9606400728225708, "learning_rate": 0.0002702986194087847, "loss": 5.054, "step": 238950 }, { "epoch": 1.4411307147767152, "grad_norm": 1.816355586051941, "learning_rate": 0.000270286463124345, "loss": 5.065, "step": 239000 }, { "epoch": 1.4414322065580492, "grad_norm": 2.1328227519989014, "learning_rate": 0.0002702743046262044, "loss": 5.1754, "step": 239050 }, { "epoch": 1.4417336983393834, "grad_norm": 1.3387951850891113, "learning_rate": 0.00027026214391458696, "loss": 4.8279, "step": 239100 }, { "epoch": 1.4420351901207173, "grad_norm": 1.0727940797805786, "learning_rate": 0.00027024998098971645, "loss": 5.0868, "step": 239150 }, { "epoch": 1.4423366819020513, "grad_norm": 1.4479460716247559, "learning_rate": 0.00027023781585181655, "loss": 5.3326, "step": 239200 }, { "epoch": 1.4426381736833853, "grad_norm": 1.0687549114227295, "learning_rate": 0.0002702256485011113, "loss": 5.3127, "step": 239250 }, { "epoch": 1.4429396654647193, "grad_norm": 1.2893435955047607, "learning_rate": 0.0002702134789378245, "loss": 5.1293, "step": 239300 }, { "epoch": 1.4432411572460535, "grad_norm": 1.8511439561843872, "learning_rate": 0.0002702013071621802, "loss": 4.8294, "step": 239350 }, { "epoch": 1.4435426490273875, "grad_norm": 1.3113986253738403, "learning_rate": 0.0002701891331744024, "loss": 5.0477, "step": 239400 }, { "epoch": 1.4438441408087215, "grad_norm": 1.719482660293579, "learning_rate": 0.0002701769569747151, "loss": 4.6989, "step": 239450 }, { "epoch": 1.4441456325900557, "grad_norm": 1.542165756225586, "learning_rate": 0.0002701647785633424, "loss": 5.2104, "step": 239500 }, { "epoch": 1.4444471243713897, "grad_norm": 2.260002851486206, "learning_rate": 0.0002701525979405085, "loss": 4.5221, "step": 239550 }, { "epoch": 1.4447486161527237, "grad_norm": 1.1501930952072144, "learning_rate": 0.00027014041510643744, "loss": 4.9365, "step": 239600 }, { "epoch": 1.4450501079340576, "grad_norm": 1.3256815671920776, "learning_rate": 0.0002701282300613535, "loss": 4.8733, "step": 239650 }, { "epoch": 1.4453515997153916, "grad_norm": 1.4008033275604248, "learning_rate": 0.0002701160428054809, "loss": 4.9063, "step": 239700 }, { "epoch": 1.4456530914967258, "grad_norm": 1.559951663017273, "learning_rate": 0.000270103853339044, "loss": 5.6055, "step": 239750 }, { "epoch": 1.4459545832780598, "grad_norm": 1.6592098474502563, "learning_rate": 0.00027009166166226705, "loss": 5.1253, "step": 239800 }, { "epoch": 1.4462560750593938, "grad_norm": 1.224100112915039, "learning_rate": 0.0002700794677753744, "loss": 5.5493, "step": 239850 }, { "epoch": 1.446557566840728, "grad_norm": 1.976806879043579, "learning_rate": 0.00027006727167859057, "loss": 4.9548, "step": 239900 }, { "epoch": 1.446859058622062, "grad_norm": 1.6091058254241943, "learning_rate": 0.00027005507337213996, "loss": 5.2162, "step": 239950 }, { "epoch": 1.447160550403396, "grad_norm": 1.378793478012085, "learning_rate": 0.000270042872856247, "loss": 5.2417, "step": 240000 }, { "epoch": 1.447160550403396, "eval_loss": 5.586062908172607, "eval_runtime": 39.2424, "eval_samples_per_second": 13.047, "eval_steps_per_second": 6.524, "eval_tts_loss": 7.270881838286578, "step": 240000 }, { "epoch": 1.4474620421847302, "grad_norm": 1.3474164009094238, "learning_rate": 0.0002700306701311363, "loss": 5.2884, "step": 240050 }, { "epoch": 1.4477635339660642, "grad_norm": 1.5053210258483887, "learning_rate": 0.0002700184651970324, "loss": 5.3667, "step": 240100 }, { "epoch": 1.4480650257473981, "grad_norm": 1.7445566654205322, "learning_rate": 0.0002700062580541599, "loss": 5.0912, "step": 240150 }, { "epoch": 1.4483665175287321, "grad_norm": 1.7397491931915283, "learning_rate": 0.0002699940487027435, "loss": 5.0953, "step": 240200 }, { "epoch": 1.4486680093100661, "grad_norm": 1.198474407196045, "learning_rate": 0.0002699818371430079, "loss": 5.1257, "step": 240250 }, { "epoch": 1.4489695010914003, "grad_norm": 1.3785136938095093, "learning_rate": 0.0002699696233751778, "loss": 5.0462, "step": 240300 }, { "epoch": 1.4492709928727343, "grad_norm": 1.059618592262268, "learning_rate": 0.00026995740739947795, "loss": 5.2443, "step": 240350 }, { "epoch": 1.4495724846540683, "grad_norm": 1.2682691812515259, "learning_rate": 0.00026994518921613327, "loss": 5.0257, "step": 240400 }, { "epoch": 1.4498739764354025, "grad_norm": 1.7740464210510254, "learning_rate": 0.0002699329688253685, "loss": 5.0549, "step": 240450 }, { "epoch": 1.4501754682167365, "grad_norm": 1.4043420553207397, "learning_rate": 0.00026992074622740863, "loss": 4.913, "step": 240500 }, { "epoch": 1.4504769599980705, "grad_norm": 1.6248681545257568, "learning_rate": 0.00026990852142247856, "loss": 5.5515, "step": 240550 }, { "epoch": 1.4507784517794045, "grad_norm": 1.697767972946167, "learning_rate": 0.00026989629441080324, "loss": 5.1517, "step": 240600 }, { "epoch": 1.4510799435607384, "grad_norm": 1.7011079788208008, "learning_rate": 0.0002698840651926077, "loss": 5.1109, "step": 240650 }, { "epoch": 1.4513814353420726, "grad_norm": 1.4908826351165771, "learning_rate": 0.0002698718337681171, "loss": 5.2635, "step": 240700 }, { "epoch": 1.4516829271234066, "grad_norm": 1.5713540315628052, "learning_rate": 0.0002698596001375564, "loss": 5.1736, "step": 240750 }, { "epoch": 1.4519844189047406, "grad_norm": 1.9161856174468994, "learning_rate": 0.00026984736430115086, "loss": 5.1953, "step": 240800 }, { "epoch": 1.4522859106860748, "grad_norm": 2.270031213760376, "learning_rate": 0.00026983512625912556, "loss": 5.3209, "step": 240850 }, { "epoch": 1.4525874024674088, "grad_norm": 1.350219488143921, "learning_rate": 0.0002698228860117058, "loss": 4.795, "step": 240900 }, { "epoch": 1.4528888942487428, "grad_norm": 0.49720823764801025, "learning_rate": 0.00026981064355911685, "loss": 5.3182, "step": 240950 }, { "epoch": 1.4531903860300768, "grad_norm": 1.333680272102356, "learning_rate": 0.00026979839890158393, "loss": 5.3888, "step": 241000 }, { "epoch": 1.4534918778114108, "grad_norm": 0.8663345575332642, "learning_rate": 0.00026978615203933245, "loss": 5.3902, "step": 241050 }, { "epoch": 1.453793369592745, "grad_norm": 2.5777342319488525, "learning_rate": 0.0002697739029725878, "loss": 5.1245, "step": 241100 }, { "epoch": 1.454094861374079, "grad_norm": 1.3993563652038574, "learning_rate": 0.00026976165170157544, "loss": 5.2231, "step": 241150 }, { "epoch": 1.454396353155413, "grad_norm": 1.2490451335906982, "learning_rate": 0.0002697493982265207, "loss": 5.613, "step": 241200 }, { "epoch": 1.4546978449367471, "grad_norm": 2.658761501312256, "learning_rate": 0.00026973714254764923, "loss": 5.6713, "step": 241250 }, { "epoch": 1.4549993367180811, "grad_norm": 1.612487554550171, "learning_rate": 0.00026972488466518653, "loss": 5.2131, "step": 241300 }, { "epoch": 1.4553008284994151, "grad_norm": 2.8895320892333984, "learning_rate": 0.00026971262457935814, "loss": 5.5604, "step": 241350 }, { "epoch": 1.455602320280749, "grad_norm": 1.7460496425628662, "learning_rate": 0.0002697003622903898, "loss": 4.9226, "step": 241400 }, { "epoch": 1.455903812062083, "grad_norm": 1.416326880455017, "learning_rate": 0.00026968809779850706, "loss": 5.5444, "step": 241450 }, { "epoch": 1.4562053038434173, "grad_norm": 1.3309859037399292, "learning_rate": 0.0002696758311039357, "loss": 5.1434, "step": 241500 }, { "epoch": 1.4565067956247513, "grad_norm": 1.5049649477005005, "learning_rate": 0.0002696635622069015, "loss": 4.8361, "step": 241550 }, { "epoch": 1.4568082874060853, "grad_norm": 1.7954357862472534, "learning_rate": 0.0002696512911076301, "loss": 5.4268, "step": 241600 }, { "epoch": 1.4571097791874195, "grad_norm": 1.114702582359314, "learning_rate": 0.0002696390178063475, "loss": 5.1625, "step": 241650 }, { "epoch": 1.4574112709687534, "grad_norm": 1.5098211765289307, "learning_rate": 0.0002696267423032795, "loss": 5.2763, "step": 241700 }, { "epoch": 1.4577127627500874, "grad_norm": 2.002081871032715, "learning_rate": 0.0002696144645986521, "loss": 4.5385, "step": 241750 }, { "epoch": 1.4580142545314214, "grad_norm": 1.620376706123352, "learning_rate": 0.0002696021846926911, "loss": 5.0004, "step": 241800 }, { "epoch": 1.4583157463127554, "grad_norm": 1.2932205200195312, "learning_rate": 0.0002695899025856226, "loss": 5.1144, "step": 241850 }, { "epoch": 1.4586172380940896, "grad_norm": 1.4388962984085083, "learning_rate": 0.0002695776182776726, "loss": 5.5375, "step": 241900 }, { "epoch": 1.4589187298754236, "grad_norm": 0.9421124458312988, "learning_rate": 0.0002695653317690672, "loss": 5.1328, "step": 241950 }, { "epoch": 1.4592202216567576, "grad_norm": 1.1496307849884033, "learning_rate": 0.00026955304306003244, "loss": 5.1182, "step": 242000 }, { "epoch": 1.4595217134380918, "grad_norm": 1.179880142211914, "learning_rate": 0.0002695407521507946, "loss": 5.4664, "step": 242050 }, { "epoch": 1.4598232052194258, "grad_norm": 1.3349618911743164, "learning_rate": 0.00026952845904157976, "loss": 5.0372, "step": 242100 }, { "epoch": 1.4601246970007598, "grad_norm": 1.343213438987732, "learning_rate": 0.00026951616373261424, "loss": 5.003, "step": 242150 }, { "epoch": 1.4604261887820937, "grad_norm": 1.4305955171585083, "learning_rate": 0.00026950386622412426, "loss": 5.1332, "step": 242200 }, { "epoch": 1.4607276805634277, "grad_norm": 1.2908860445022583, "learning_rate": 0.00026949156651633613, "loss": 5.3319, "step": 242250 }, { "epoch": 1.461029172344762, "grad_norm": 1.4929475784301758, "learning_rate": 0.00026947926460947635, "loss": 5.7985, "step": 242300 }, { "epoch": 1.461330664126096, "grad_norm": 0.975675106048584, "learning_rate": 0.0002694669605037711, "loss": 5.1128, "step": 242350 }, { "epoch": 1.46163215590743, "grad_norm": 1.3685688972473145, "learning_rate": 0.000269454654199447, "loss": 5.3272, "step": 242400 }, { "epoch": 1.461933647688764, "grad_norm": 1.1330523490905762, "learning_rate": 0.0002694423456967304, "loss": 5.1231, "step": 242450 }, { "epoch": 1.462235139470098, "grad_norm": 0.2129719853401184, "learning_rate": 0.000269430034995848, "loss": 5.2258, "step": 242500 }, { "epoch": 1.462536631251432, "grad_norm": 0.9688673615455627, "learning_rate": 0.0002694177220970261, "loss": 5.0646, "step": 242550 }, { "epoch": 1.462838123032766, "grad_norm": 1.1817368268966675, "learning_rate": 0.00026940540700049154, "loss": 5.1103, "step": 242600 }, { "epoch": 1.4631396148141, "grad_norm": 1.8962982892990112, "learning_rate": 0.0002693930897064708, "loss": 5.4442, "step": 242650 }, { "epoch": 1.4634411065954342, "grad_norm": 1.2195281982421875, "learning_rate": 0.00026938077021519067, "loss": 5.1863, "step": 242700 }, { "epoch": 1.4637425983767682, "grad_norm": 0.6270005106925964, "learning_rate": 0.00026936844852687783, "loss": 4.9415, "step": 242750 }, { "epoch": 1.4640440901581022, "grad_norm": 1.1089617013931274, "learning_rate": 0.00026935612464175905, "loss": 4.8753, "step": 242800 }, { "epoch": 1.4643455819394364, "grad_norm": 0.24333453178405762, "learning_rate": 0.000269343798560061, "loss": 4.8368, "step": 242850 }, { "epoch": 1.4646470737207704, "grad_norm": 0.5240609645843506, "learning_rate": 0.00026933147028201086, "loss": 5.3086, "step": 242900 }, { "epoch": 1.4649485655021044, "grad_norm": 1.1065869331359863, "learning_rate": 0.00026931913980783514, "loss": 5.4077, "step": 242950 }, { "epoch": 1.4652500572834384, "grad_norm": 0.5450378060340881, "learning_rate": 0.000269306807137761, "loss": 4.9077, "step": 243000 }, { "epoch": 1.4652500572834384, "eval_loss": 5.579868316650391, "eval_runtime": 39.039, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.215405977609811, "step": 243000 }, { "epoch": 1.4655515490647724, "grad_norm": 1.2876368761062622, "learning_rate": 0.00026929447227201533, "loss": 5.1221, "step": 243050 }, { "epoch": 1.4658530408461066, "grad_norm": 1.1868305206298828, "learning_rate": 0.00026928213521082515, "loss": 4.8983, "step": 243100 }, { "epoch": 1.4661545326274406, "grad_norm": 1.4596840143203735, "learning_rate": 0.0002692697959544175, "loss": 5.1266, "step": 243150 }, { "epoch": 1.4664560244087745, "grad_norm": 0.7659274339675903, "learning_rate": 0.00026925745450301944, "loss": 4.8253, "step": 243200 }, { "epoch": 1.4667575161901087, "grad_norm": 1.6383368968963623, "learning_rate": 0.0002692451108568581, "loss": 5.0443, "step": 243250 }, { "epoch": 1.4670590079714427, "grad_norm": 0.956576406955719, "learning_rate": 0.00026923276501616066, "loss": 4.627, "step": 243300 }, { "epoch": 1.4673604997527767, "grad_norm": 0.9625290036201477, "learning_rate": 0.0002692204169811544, "loss": 5.1667, "step": 243350 }, { "epoch": 1.4676619915341107, "grad_norm": 1.6502951383590698, "learning_rate": 0.00026920806675206644, "loss": 5.2535, "step": 243400 }, { "epoch": 1.4679634833154447, "grad_norm": 1.3738774061203003, "learning_rate": 0.0002691957143291242, "loss": 4.7396, "step": 243450 }, { "epoch": 1.468264975096779, "grad_norm": 2.6890182495117188, "learning_rate": 0.00026918335971255487, "loss": 5.6335, "step": 243500 }, { "epoch": 1.4685664668781129, "grad_norm": 2.186662197113037, "learning_rate": 0.0002691710029025859, "loss": 4.999, "step": 243550 }, { "epoch": 1.4688679586594469, "grad_norm": 1.2362161874771118, "learning_rate": 0.00026915864389944475, "loss": 5.1467, "step": 243600 }, { "epoch": 1.469169450440781, "grad_norm": 0.8057460188865662, "learning_rate": 0.0002691462827033587, "loss": 5.2058, "step": 243650 }, { "epoch": 1.469470942222115, "grad_norm": 1.5652416944503784, "learning_rate": 0.00026913391931455543, "loss": 5.1897, "step": 243700 }, { "epoch": 1.469772434003449, "grad_norm": 1.4179997444152832, "learning_rate": 0.00026912155373326235, "loss": 5.1165, "step": 243750 }, { "epoch": 1.4700739257847832, "grad_norm": 1.5801928043365479, "learning_rate": 0.00026910918595970707, "loss": 4.887, "step": 243800 }, { "epoch": 1.4703754175661172, "grad_norm": 1.2528555393218994, "learning_rate": 0.00026909681599411725, "loss": 5.3552, "step": 243850 }, { "epoch": 1.4706769093474512, "grad_norm": 1.8320313692092896, "learning_rate": 0.00026908444383672047, "loss": 5.4933, "step": 243900 }, { "epoch": 1.4709784011287852, "grad_norm": 2.8060078620910645, "learning_rate": 0.0002690720694877444, "loss": 5.0583, "step": 243950 }, { "epoch": 1.4712798929101192, "grad_norm": 1.4102596044540405, "learning_rate": 0.00026905969294741683, "loss": 4.7252, "step": 244000 }, { "epoch": 1.4715813846914534, "grad_norm": 0.9029082655906677, "learning_rate": 0.00026904731421596554, "loss": 5.1652, "step": 244050 }, { "epoch": 1.4718828764727874, "grad_norm": 1.4623414278030396, "learning_rate": 0.00026903493329361836, "loss": 4.8, "step": 244100 }, { "epoch": 1.4721843682541214, "grad_norm": 1.700419545173645, "learning_rate": 0.000269022550180603, "loss": 5.3186, "step": 244150 }, { "epoch": 1.4724858600354556, "grad_norm": 1.405966877937317, "learning_rate": 0.00026901016487714755, "loss": 5.0434, "step": 244200 }, { "epoch": 1.4727873518167895, "grad_norm": 0.5231946110725403, "learning_rate": 0.00026899777738347984, "loss": 5.2128, "step": 244250 }, { "epoch": 1.4730888435981235, "grad_norm": 1.924221396446228, "learning_rate": 0.0002689853876998278, "loss": 5.1027, "step": 244300 }, { "epoch": 1.4733903353794575, "grad_norm": 1.8878053426742554, "learning_rate": 0.00026897299582641956, "loss": 5.1221, "step": 244350 }, { "epoch": 1.4736918271607915, "grad_norm": 1.6171730756759644, "learning_rate": 0.0002689606017634831, "loss": 5.3585, "step": 244400 }, { "epoch": 1.4739933189421257, "grad_norm": 1.8776339292526245, "learning_rate": 0.0002689482055112465, "loss": 5.3114, "step": 244450 }, { "epoch": 1.4742948107234597, "grad_norm": 1.3071776628494263, "learning_rate": 0.00026893580706993794, "loss": 5.5729, "step": 244500 }, { "epoch": 1.4745963025047937, "grad_norm": 1.8965693712234497, "learning_rate": 0.0002689234064397856, "loss": 4.7066, "step": 244550 }, { "epoch": 1.4748977942861279, "grad_norm": 1.8272340297698975, "learning_rate": 0.0002689110036210177, "loss": 4.7408, "step": 244600 }, { "epoch": 1.4751992860674619, "grad_norm": 1.330878734588623, "learning_rate": 0.0002688985986138624, "loss": 5.2544, "step": 244650 }, { "epoch": 1.4755007778487959, "grad_norm": 2.0527474880218506, "learning_rate": 0.00026888619141854814, "loss": 5.2737, "step": 244700 }, { "epoch": 1.4758022696301298, "grad_norm": 1.7679064273834229, "learning_rate": 0.00026887378203530315, "loss": 4.8937, "step": 244750 }, { "epoch": 1.4761037614114638, "grad_norm": 2.4256367683410645, "learning_rate": 0.0002688613704643558, "loss": 4.9703, "step": 244800 }, { "epoch": 1.476405253192798, "grad_norm": 1.5104162693023682, "learning_rate": 0.00026884895670593466, "loss": 5.2044, "step": 244850 }, { "epoch": 1.476706744974132, "grad_norm": 1.315924882888794, "learning_rate": 0.000268836540760268, "loss": 5.0972, "step": 244900 }, { "epoch": 1.477008236755466, "grad_norm": 1.5141005516052246, "learning_rate": 0.00026882412262758444, "loss": 5.0352, "step": 244950 }, { "epoch": 1.4773097285368002, "grad_norm": 1.168424129486084, "learning_rate": 0.0002688117023081124, "loss": 5.6674, "step": 245000 }, { "epoch": 1.4776112203181342, "grad_norm": 0.9782376885414124, "learning_rate": 0.0002687992798020806, "loss": 5.1673, "step": 245050 }, { "epoch": 1.4779127120994682, "grad_norm": 1.104285717010498, "learning_rate": 0.0002687868551097176, "loss": 5.0375, "step": 245100 }, { "epoch": 1.4782142038808022, "grad_norm": 1.2819143533706665, "learning_rate": 0.00026877442823125205, "loss": 5.2836, "step": 245150 }, { "epoch": 1.4785156956621361, "grad_norm": 2.2486977577209473, "learning_rate": 0.0002687619991669126, "loss": 4.6621, "step": 245200 }, { "epoch": 1.4788171874434703, "grad_norm": 2.0774598121643066, "learning_rate": 0.00026874956791692804, "loss": 5.6473, "step": 245250 }, { "epoch": 1.4791186792248043, "grad_norm": 1.474021315574646, "learning_rate": 0.00026873713448152716, "loss": 4.891, "step": 245300 }, { "epoch": 1.4794201710061383, "grad_norm": 1.6193350553512573, "learning_rate": 0.00026872469886093876, "loss": 4.6649, "step": 245350 }, { "epoch": 1.4797216627874725, "grad_norm": 1.3512136936187744, "learning_rate": 0.00026871226105539177, "loss": 4.7716, "step": 245400 }, { "epoch": 1.4800231545688065, "grad_norm": 2.972100257873535, "learning_rate": 0.000268699821065115, "loss": 4.8622, "step": 245450 }, { "epoch": 1.4803246463501405, "grad_norm": 2.3247923851013184, "learning_rate": 0.0002686873788903374, "loss": 5.4187, "step": 245500 }, { "epoch": 1.4806261381314745, "grad_norm": 0.8780568838119507, "learning_rate": 0.0002686749345312879, "loss": 5.2259, "step": 245550 }, { "epoch": 1.4809276299128085, "grad_norm": 2.083179235458374, "learning_rate": 0.00026866248798819567, "loss": 5.0634, "step": 245600 }, { "epoch": 1.4812291216941427, "grad_norm": 1.6168657541275024, "learning_rate": 0.00026865003926128964, "loss": 4.6945, "step": 245650 }, { "epoch": 1.4815306134754767, "grad_norm": 1.4108105897903442, "learning_rate": 0.000268637588350799, "loss": 5.1355, "step": 245700 }, { "epoch": 1.4818321052568106, "grad_norm": 1.3749072551727295, "learning_rate": 0.00026862513525695287, "loss": 4.8542, "step": 245750 }, { "epoch": 1.4821335970381448, "grad_norm": 2.217130422592163, "learning_rate": 0.00026861267997998036, "loss": 5.1299, "step": 245800 }, { "epoch": 1.4824350888194788, "grad_norm": 2.058419942855835, "learning_rate": 0.00026860022252011076, "loss": 5.1983, "step": 245850 }, { "epoch": 1.4827365806008128, "grad_norm": 1.669350028038025, "learning_rate": 0.0002685877628775733, "loss": 5.5454, "step": 245900 }, { "epoch": 1.4830380723821468, "grad_norm": 1.9750633239746094, "learning_rate": 0.00026857530105259736, "loss": 5.2181, "step": 245950 }, { "epoch": 1.4833395641634808, "grad_norm": 1.2439823150634766, "learning_rate": 0.00026856283704541217, "loss": 4.8525, "step": 246000 }, { "epoch": 1.4833395641634808, "eval_loss": 5.5774736404418945, "eval_runtime": 38.9359, "eval_samples_per_second": 13.15, "eval_steps_per_second": 6.575, "eval_tts_loss": 7.215512286869289, "step": 246000 }, { "epoch": 1.483641055944815, "grad_norm": 1.4812560081481934, "learning_rate": 0.00026855037085624724, "loss": 4.5713, "step": 246050 }, { "epoch": 1.483942547726149, "grad_norm": 1.3452826738357544, "learning_rate": 0.0002685379024853318, "loss": 5.4048, "step": 246100 }, { "epoch": 1.484244039507483, "grad_norm": 2.8894503116607666, "learning_rate": 0.00026852543193289546, "loss": 4.9713, "step": 246150 }, { "epoch": 1.4845455312888172, "grad_norm": 1.9306645393371582, "learning_rate": 0.00026851295919916773, "loss": 4.7542, "step": 246200 }, { "epoch": 1.4848470230701512, "grad_norm": 0.4478348195552826, "learning_rate": 0.00026850048428437814, "loss": 5.4256, "step": 246250 }, { "epoch": 1.4851485148514851, "grad_norm": 1.3629028797149658, "learning_rate": 0.00026848800718875615, "loss": 5.5385, "step": 246300 }, { "epoch": 1.4854500066328191, "grad_norm": 0.5926460027694702, "learning_rate": 0.0002684755279125315, "loss": 5.2609, "step": 246350 }, { "epoch": 1.485751498414153, "grad_norm": 0.7026382684707642, "learning_rate": 0.00026846304645593386, "loss": 4.9221, "step": 246400 }, { "epoch": 1.4860529901954873, "grad_norm": 1.1429730653762817, "learning_rate": 0.0002684505628191929, "loss": 4.4759, "step": 246450 }, { "epoch": 1.4863544819768213, "grad_norm": 1.5704501867294312, "learning_rate": 0.00026843807700253837, "loss": 4.6128, "step": 246500 }, { "epoch": 1.4866559737581553, "grad_norm": 2.38492488861084, "learning_rate": 0.00026842558900620013, "loss": 5.1319, "step": 246550 }, { "epoch": 1.4869574655394895, "grad_norm": 1.3643051385879517, "learning_rate": 0.0002684130988304078, "loss": 4.9833, "step": 246600 }, { "epoch": 1.4872589573208235, "grad_norm": 1.8319456577301025, "learning_rate": 0.0002684006064753914, "loss": 5.4432, "step": 246650 }, { "epoch": 1.4875604491021575, "grad_norm": 2.787646532058716, "learning_rate": 0.00026838811194138085, "loss": 5.1444, "step": 246700 }, { "epoch": 1.4878619408834914, "grad_norm": 1.4523407220840454, "learning_rate": 0.000268375615228606, "loss": 5.2206, "step": 246750 }, { "epoch": 1.4881634326648254, "grad_norm": 1.1919804811477661, "learning_rate": 0.0002683631163372969, "loss": 4.8008, "step": 246800 }, { "epoch": 1.4884649244461596, "grad_norm": 1.3504871129989624, "learning_rate": 0.0002683506152676836, "loss": 5.444, "step": 246850 }, { "epoch": 1.4887664162274936, "grad_norm": 1.471798300743103, "learning_rate": 0.0002683381120199961, "loss": 5.1506, "step": 246900 }, { "epoch": 1.4890679080088276, "grad_norm": 1.2247775793075562, "learning_rate": 0.00026832560659446447, "loss": 5.1568, "step": 246950 }, { "epoch": 1.4893693997901618, "grad_norm": 1.1995197534561157, "learning_rate": 0.000268313098991319, "loss": 5.1985, "step": 247000 }, { "epoch": 1.4896708915714958, "grad_norm": 0.531264066696167, "learning_rate": 0.0002683005892107897, "loss": 4.7731, "step": 247050 }, { "epoch": 1.4899723833528298, "grad_norm": 1.9759753942489624, "learning_rate": 0.00026828807725310695, "loss": 4.9084, "step": 247100 }, { "epoch": 1.4902738751341638, "grad_norm": 1.0403705835342407, "learning_rate": 0.00026827556311850094, "loss": 4.9872, "step": 247150 }, { "epoch": 1.4905753669154977, "grad_norm": 1.2703450918197632, "learning_rate": 0.0002682630468072019, "loss": 5.1345, "step": 247200 }, { "epoch": 1.490876858696832, "grad_norm": 1.7248573303222656, "learning_rate": 0.0002682505283194403, "loss": 5.1073, "step": 247250 }, { "epoch": 1.491178350478166, "grad_norm": 1.3881303071975708, "learning_rate": 0.00026823800765544655, "loss": 5.0628, "step": 247300 }, { "epoch": 1.4914798422595, "grad_norm": 0.43059009313583374, "learning_rate": 0.0002682254848154509, "loss": 5.0155, "step": 247350 }, { "epoch": 1.4917813340408341, "grad_norm": 0.8507065773010254, "learning_rate": 0.00026821295979968394, "loss": 4.7598, "step": 247400 }, { "epoch": 1.4920828258221681, "grad_norm": 1.1904772520065308, "learning_rate": 0.0002682004326083762, "loss": 4.9395, "step": 247450 }, { "epoch": 1.492384317603502, "grad_norm": 1.5242570638656616, "learning_rate": 0.0002681879032417581, "loss": 5.093, "step": 247500 }, { "epoch": 1.4926858093848363, "grad_norm": 0.9632312655448914, "learning_rate": 0.0002681753717000604, "loss": 5.1283, "step": 247550 }, { "epoch": 1.4929873011661703, "grad_norm": 1.4422675371170044, "learning_rate": 0.0002681628379835136, "loss": 5.0911, "step": 247600 }, { "epoch": 1.4932887929475043, "grad_norm": 1.0568444728851318, "learning_rate": 0.00026815030209234836, "loss": 5.502, "step": 247650 }, { "epoch": 1.4935902847288383, "grad_norm": 1.1306309700012207, "learning_rate": 0.00026813776402679544, "loss": 4.8176, "step": 247700 }, { "epoch": 1.4938917765101722, "grad_norm": 1.798449158668518, "learning_rate": 0.00026812522378708557, "loss": 5.0184, "step": 247750 }, { "epoch": 1.4941932682915064, "grad_norm": 1.7015098333358765, "learning_rate": 0.0002681126813734495, "loss": 4.9163, "step": 247800 }, { "epoch": 1.4944947600728404, "grad_norm": 1.1867547035217285, "learning_rate": 0.00026810013678611814, "loss": 4.7881, "step": 247850 }, { "epoch": 1.4947962518541744, "grad_norm": 1.2063987255096436, "learning_rate": 0.00026808759002532224, "loss": 5.174, "step": 247900 }, { "epoch": 1.4950977436355086, "grad_norm": 1.3084698915481567, "learning_rate": 0.0002680750410912928, "loss": 5.3496, "step": 247950 }, { "epoch": 1.4953992354168426, "grad_norm": 1.4732075929641724, "learning_rate": 0.0002680624899842607, "loss": 5.1117, "step": 248000 }, { "epoch": 1.4957007271981766, "grad_norm": 2.0427744388580322, "learning_rate": 0.00026804993670445697, "loss": 5.3457, "step": 248050 }, { "epoch": 1.4960022189795106, "grad_norm": 2.07088565826416, "learning_rate": 0.00026803738125211267, "loss": 4.9439, "step": 248100 }, { "epoch": 1.4963037107608446, "grad_norm": 1.565977692604065, "learning_rate": 0.0002680248236274588, "loss": 5.0907, "step": 248150 }, { "epoch": 1.4966052025421788, "grad_norm": 1.226742148399353, "learning_rate": 0.00026801226383072645, "loss": 5.1954, "step": 248200 }, { "epoch": 1.4969066943235128, "grad_norm": 2.1837007999420166, "learning_rate": 0.0002679997018621468, "loss": 5.1692, "step": 248250 }, { "epoch": 1.4972081861048467, "grad_norm": 1.309983730316162, "learning_rate": 0.000267987137721951, "loss": 4.9776, "step": 248300 }, { "epoch": 1.497509677886181, "grad_norm": 0.6446052193641663, "learning_rate": 0.00026797457141037033, "loss": 4.974, "step": 248350 }, { "epoch": 1.497811169667515, "grad_norm": 2.321082592010498, "learning_rate": 0.0002679620029276361, "loss": 4.8543, "step": 248400 }, { "epoch": 1.498112661448849, "grad_norm": 1.5929356813430786, "learning_rate": 0.0002679494322739795, "loss": 5.4211, "step": 248450 }, { "epoch": 1.498414153230183, "grad_norm": 1.1854428052902222, "learning_rate": 0.0002679368594496319, "loss": 5.1386, "step": 248500 }, { "epoch": 1.4987156450115169, "grad_norm": 1.3071434497833252, "learning_rate": 0.0002679242844548247, "loss": 5.0047, "step": 248550 }, { "epoch": 1.499017136792851, "grad_norm": 1.203904628753662, "learning_rate": 0.0002679117072897894, "loss": 5.348, "step": 248600 }, { "epoch": 1.499318628574185, "grad_norm": 1.5721930265426636, "learning_rate": 0.0002678991279547573, "loss": 5.0818, "step": 248650 }, { "epoch": 1.499620120355519, "grad_norm": 1.6889588832855225, "learning_rate": 0.00026788654644996005, "loss": 5.1297, "step": 248700 }, { "epoch": 1.4999216121368533, "grad_norm": 1.0055404901504517, "learning_rate": 0.00026787396277562915, "loss": 4.7763, "step": 248750 }, { "epoch": 1.500223103918187, "grad_norm": 1.3842177391052246, "learning_rate": 0.0002678613769319962, "loss": 5.0376, "step": 248800 }, { "epoch": 1.5005245956995212, "grad_norm": 1.5473884344100952, "learning_rate": 0.0002678487889192928, "loss": 5.136, "step": 248850 }, { "epoch": 1.5008260874808552, "grad_norm": 1.1193110942840576, "learning_rate": 0.00026783619873775057, "loss": 4.9588, "step": 248900 }, { "epoch": 1.5011275792621892, "grad_norm": 1.2613270282745361, "learning_rate": 0.0002678236063876013, "loss": 4.9464, "step": 248950 }, { "epoch": 1.5014290710435234, "grad_norm": 1.4803158044815063, "learning_rate": 0.0002678110118690767, "loss": 4.588, "step": 249000 }, { "epoch": 1.5014290710435234, "eval_loss": 5.591412544250488, "eval_runtime": 39.0841, "eval_samples_per_second": 13.1, "eval_steps_per_second": 6.55, "eval_tts_loss": 7.201006490956142, "step": 249000 }, { "epoch": 1.5017305628248574, "grad_norm": 1.7339829206466675, "learning_rate": 0.00026779841518240855, "loss": 5.1734, "step": 249050 }, { "epoch": 1.5020320546061914, "grad_norm": 1.0774741172790527, "learning_rate": 0.00026778581632782867, "loss": 5.1536, "step": 249100 }, { "epoch": 1.5023335463875256, "grad_norm": 1.6778693199157715, "learning_rate": 0.0002677732153055689, "loss": 5.4347, "step": 249150 }, { "epoch": 1.5026350381688596, "grad_norm": 2.130854368209839, "learning_rate": 0.0002677606121158612, "loss": 5.1761, "step": 249200 }, { "epoch": 1.5029365299501936, "grad_norm": 1.7395298480987549, "learning_rate": 0.00026774800675893754, "loss": 5.2714, "step": 249250 }, { "epoch": 1.5032380217315278, "grad_norm": 2.013859510421753, "learning_rate": 0.00026773539923502976, "loss": 5.3755, "step": 249300 }, { "epoch": 1.5035395135128615, "grad_norm": 1.1508721113204956, "learning_rate": 0.00026772278954437, "loss": 4.7866, "step": 249350 }, { "epoch": 1.5038410052941957, "grad_norm": 0.39175984263420105, "learning_rate": 0.00026771017768719037, "loss": 5.0394, "step": 249400 }, { "epoch": 1.5041424970755297, "grad_norm": 1.4668056964874268, "learning_rate": 0.0002676975636637228, "loss": 5.0895, "step": 249450 }, { "epoch": 1.5044439888568637, "grad_norm": 2.1442501544952393, "learning_rate": 0.0002676849474741996, "loss": 4.8608, "step": 249500 }, { "epoch": 1.504745480638198, "grad_norm": 0.7786572575569153, "learning_rate": 0.0002676723291188529, "loss": 4.7136, "step": 249550 }, { "epoch": 1.505046972419532, "grad_norm": 0.6426277756690979, "learning_rate": 0.00026765970859791483, "loss": 5.099, "step": 249600 }, { "epoch": 1.5053484642008659, "grad_norm": 1.2322568893432617, "learning_rate": 0.0002676470859116178, "loss": 5.2687, "step": 249650 }, { "epoch": 1.5056499559822, "grad_norm": 1.3302698135375977, "learning_rate": 0.00026763446106019403, "loss": 4.7425, "step": 249700 }, { "epoch": 1.5059514477635338, "grad_norm": 1.4410076141357422, "learning_rate": 0.00026762183404387585, "loss": 5.3536, "step": 249750 }, { "epoch": 1.506252939544868, "grad_norm": 1.1614360809326172, "learning_rate": 0.0002676092048628957, "loss": 5.1976, "step": 249800 }, { "epoch": 1.506554431326202, "grad_norm": 1.8134245872497559, "learning_rate": 0.00026759657351748593, "loss": 5.3163, "step": 249850 }, { "epoch": 1.506855923107536, "grad_norm": 0.8812612295150757, "learning_rate": 0.0002675839400078791, "loss": 5.0165, "step": 249900 }, { "epoch": 1.5071574148888702, "grad_norm": 1.5527985095977783, "learning_rate": 0.0002675713043343076, "loss": 5.4512, "step": 249950 }, { "epoch": 1.5074589066702042, "grad_norm": 1.3754411935806274, "learning_rate": 0.0002675586664970041, "loss": 5.3312, "step": 250000 }, { "epoch": 1.5077603984515382, "grad_norm": 1.4160559177398682, "learning_rate": 0.0002675460264962011, "loss": 4.7035, "step": 250050 }, { "epoch": 1.5080618902328724, "grad_norm": 0.7280053496360779, "learning_rate": 0.00026753338433213114, "loss": 5.1213, "step": 250100 }, { "epoch": 1.5083633820142062, "grad_norm": 1.6796705722808838, "learning_rate": 0.0002675207400050271, "loss": 4.7724, "step": 250150 }, { "epoch": 1.5086648737955404, "grad_norm": 1.341314673423767, "learning_rate": 0.00026750809351512144, "loss": 5.2406, "step": 250200 }, { "epoch": 1.5089663655768744, "grad_norm": 1.9623651504516602, "learning_rate": 0.00026749544486264707, "loss": 5.1056, "step": 250250 }, { "epoch": 1.5092678573582083, "grad_norm": 1.2935644388198853, "learning_rate": 0.0002674827940478367, "loss": 5.4033, "step": 250300 }, { "epoch": 1.5095693491395425, "grad_norm": 1.4374420642852783, "learning_rate": 0.00026747014107092313, "loss": 5.2567, "step": 250350 }, { "epoch": 1.5098708409208765, "grad_norm": 1.5754681825637817, "learning_rate": 0.0002674574859321393, "loss": 5.0841, "step": 250400 }, { "epoch": 1.5101723327022105, "grad_norm": 1.5498350858688354, "learning_rate": 0.000267444828631718, "loss": 5.048, "step": 250450 }, { "epoch": 1.5104738244835447, "grad_norm": 1.7034388780593872, "learning_rate": 0.0002674321691698923, "loss": 4.8096, "step": 250500 }, { "epoch": 1.5107753162648785, "grad_norm": 2.025146007537842, "learning_rate": 0.00026741950754689505, "loss": 5.2437, "step": 250550 }, { "epoch": 1.5110768080462127, "grad_norm": 1.311733365058899, "learning_rate": 0.0002674068437629594, "loss": 4.7608, "step": 250600 }, { "epoch": 1.5113782998275467, "grad_norm": 1.610182523727417, "learning_rate": 0.00026739417781831835, "loss": 5.109, "step": 250650 }, { "epoch": 1.5116797916088807, "grad_norm": 1.9828333854675293, "learning_rate": 0.0002673815097132049, "loss": 5.0587, "step": 250700 }, { "epoch": 1.5119812833902149, "grad_norm": 1.1309963464736938, "learning_rate": 0.0002673688394478523, "loss": 5.2357, "step": 250750 }, { "epoch": 1.5122827751715489, "grad_norm": 1.6027759313583374, "learning_rate": 0.0002673561670224937, "loss": 5.1578, "step": 250800 }, { "epoch": 1.5125842669528828, "grad_norm": 1.1191171407699585, "learning_rate": 0.0002673434924373624, "loss": 5.3233, "step": 250850 }, { "epoch": 1.512885758734217, "grad_norm": 1.7582602500915527, "learning_rate": 0.0002673308156926915, "loss": 5.0259, "step": 250900 }, { "epoch": 1.5131872505155508, "grad_norm": 1.096724033355713, "learning_rate": 0.0002673181367887144, "loss": 5.1958, "step": 250950 }, { "epoch": 1.513488742296885, "grad_norm": 1.6479060649871826, "learning_rate": 0.0002673054557256644, "loss": 5.1126, "step": 251000 }, { "epoch": 1.513790234078219, "grad_norm": 1.1729646921157837, "learning_rate": 0.00026729277250377495, "loss": 4.7689, "step": 251050 }, { "epoch": 1.514091725859553, "grad_norm": 0.9956018924713135, "learning_rate": 0.00026728008712327935, "loss": 4.9974, "step": 251100 }, { "epoch": 1.5143932176408872, "grad_norm": 1.7046045064926147, "learning_rate": 0.00026726739958441115, "loss": 5.4627, "step": 251150 }, { "epoch": 1.5146947094222212, "grad_norm": 1.270582675933838, "learning_rate": 0.0002672547098874037, "loss": 5.1469, "step": 251200 }, { "epoch": 1.5149962012035552, "grad_norm": 0.6629209518432617, "learning_rate": 0.0002672420180324908, "loss": 4.8897, "step": 251250 }, { "epoch": 1.5152976929848894, "grad_norm": 1.2682278156280518, "learning_rate": 0.0002672293240199058, "loss": 4.7869, "step": 251300 }, { "epoch": 1.5155991847662231, "grad_norm": 1.490918517112732, "learning_rate": 0.0002672166278498824, "loss": 5.3017, "step": 251350 }, { "epoch": 1.5159006765475573, "grad_norm": 1.2034475803375244, "learning_rate": 0.00026720392952265425, "loss": 5.1118, "step": 251400 }, { "epoch": 1.5162021683288913, "grad_norm": 1.2092406749725342, "learning_rate": 0.000267191229038455, "loss": 5.1003, "step": 251450 }, { "epoch": 1.5165036601102253, "grad_norm": 1.7193734645843506, "learning_rate": 0.0002671785263975184, "loss": 4.82, "step": 251500 }, { "epoch": 1.5168051518915595, "grad_norm": 1.4312695264816284, "learning_rate": 0.0002671658216000783, "loss": 5.2986, "step": 251550 }, { "epoch": 1.5171066436728935, "grad_norm": 0.8741110563278198, "learning_rate": 0.00026715311464636847, "loss": 5.2159, "step": 251600 }, { "epoch": 1.5174081354542275, "grad_norm": 1.2728830575942993, "learning_rate": 0.00026714040553662275, "loss": 5.3868, "step": 251650 }, { "epoch": 1.5177096272355617, "grad_norm": 1.139791488647461, "learning_rate": 0.00026712769427107496, "loss": 5.4172, "step": 251700 }, { "epoch": 1.5180111190168954, "grad_norm": 1.1663726568222046, "learning_rate": 0.00026711498084995924, "loss": 5.3194, "step": 251750 }, { "epoch": 1.5183126107982297, "grad_norm": 1.6604955196380615, "learning_rate": 0.0002671022652735093, "loss": 5.1951, "step": 251800 }, { "epoch": 1.5186141025795636, "grad_norm": 1.8878905773162842, "learning_rate": 0.0002670895475419593, "loss": 4.9173, "step": 251850 }, { "epoch": 1.5189155943608976, "grad_norm": 1.484925627708435, "learning_rate": 0.0002670768276555434, "loss": 5.2148, "step": 251900 }, { "epoch": 1.5192170861422318, "grad_norm": 1.5364118814468384, "learning_rate": 0.00026706410561449536, "loss": 4.764, "step": 251950 }, { "epoch": 1.5195185779235658, "grad_norm": 1.3117597103118896, "learning_rate": 0.00026705138141904964, "loss": 4.9674, "step": 252000 }, { "epoch": 1.5195185779235658, "eval_loss": 5.571996688842773, "eval_runtime": 38.9457, "eval_samples_per_second": 13.147, "eval_steps_per_second": 6.573, "eval_tts_loss": 7.248013632114586, "step": 252000 }, { "epoch": 1.5198200697048998, "grad_norm": 1.3630378246307373, "learning_rate": 0.0002670386550694403, "loss": 5.1232, "step": 252050 }, { "epoch": 1.520121561486234, "grad_norm": 1.3302648067474365, "learning_rate": 0.0002670259265659015, "loss": 5.1885, "step": 252100 }, { "epoch": 1.5204230532675678, "grad_norm": 0.8135322332382202, "learning_rate": 0.0002670131959086675, "loss": 5.2585, "step": 252150 }, { "epoch": 1.520724545048902, "grad_norm": 2.1379220485687256, "learning_rate": 0.00026700046309797266, "loss": 5.0585, "step": 252200 }, { "epoch": 1.521026036830236, "grad_norm": 1.1598538160324097, "learning_rate": 0.00026698772813405126, "loss": 5.2075, "step": 252250 }, { "epoch": 1.52132752861157, "grad_norm": 1.7826976776123047, "learning_rate": 0.00026697499101713767, "loss": 5.3948, "step": 252300 }, { "epoch": 1.5216290203929042, "grad_norm": 0.8131643533706665, "learning_rate": 0.0002669622517474663, "loss": 4.7621, "step": 252350 }, { "epoch": 1.5219305121742381, "grad_norm": 0.7025651931762695, "learning_rate": 0.00026694951032527156, "loss": 5.1228, "step": 252400 }, { "epoch": 1.5222320039555721, "grad_norm": 0.35868650674819946, "learning_rate": 0.00026693676675078803, "loss": 5.0576, "step": 252450 }, { "epoch": 1.5225334957369063, "grad_norm": 1.247592568397522, "learning_rate": 0.0002669240210242502, "loss": 5.4491, "step": 252500 }, { "epoch": 1.52283498751824, "grad_norm": 1.5062057971954346, "learning_rate": 0.0002669112731458926, "loss": 4.8325, "step": 252550 }, { "epoch": 1.5231364792995743, "grad_norm": 2.9110031127929688, "learning_rate": 0.0002668985231159498, "loss": 5.1618, "step": 252600 }, { "epoch": 1.5234379710809083, "grad_norm": 1.9412074089050293, "learning_rate": 0.00026688577093465657, "loss": 5.2744, "step": 252650 }, { "epoch": 1.5237394628622423, "grad_norm": 1.2897250652313232, "learning_rate": 0.00026687301660224756, "loss": 4.9589, "step": 252700 }, { "epoch": 1.5240409546435765, "grad_norm": 1.2920644283294678, "learning_rate": 0.0002668602601189574, "loss": 5.1518, "step": 252750 }, { "epoch": 1.5243424464249105, "grad_norm": 1.3658106327056885, "learning_rate": 0.00026684750148502095, "loss": 5.1048, "step": 252800 }, { "epoch": 1.5246439382062444, "grad_norm": 1.513993501663208, "learning_rate": 0.00026683474070067297, "loss": 4.8088, "step": 252850 }, { "epoch": 1.5249454299875786, "grad_norm": 1.300506830215454, "learning_rate": 0.00026682197776614836, "loss": 4.7472, "step": 252900 }, { "epoch": 1.5252469217689126, "grad_norm": 1.4977729320526123, "learning_rate": 0.00026680921268168196, "loss": 5.2766, "step": 252950 }, { "epoch": 1.5255484135502466, "grad_norm": 2.272357940673828, "learning_rate": 0.00026679644544750866, "loss": 5.398, "step": 253000 }, { "epoch": 1.5258499053315808, "grad_norm": 1.2495571374893188, "learning_rate": 0.00026678367606386346, "loss": 5.3303, "step": 253050 }, { "epoch": 1.5261513971129146, "grad_norm": 2.394646167755127, "learning_rate": 0.00026677090453098136, "loss": 5.2218, "step": 253100 }, { "epoch": 1.5264528888942488, "grad_norm": 1.2461585998535156, "learning_rate": 0.0002667581308490975, "loss": 5.3644, "step": 253150 }, { "epoch": 1.5267543806755828, "grad_norm": 1.2954908609390259, "learning_rate": 0.0002667453550184467, "loss": 5.3508, "step": 253200 }, { "epoch": 1.5270558724569168, "grad_norm": 1.1366695165634155, "learning_rate": 0.00026673257703926435, "loss": 5.2805, "step": 253250 }, { "epoch": 1.527357364238251, "grad_norm": 1.351727843284607, "learning_rate": 0.0002667197969117855, "loss": 5.4279, "step": 253300 }, { "epoch": 1.527658856019585, "grad_norm": 1.2114629745483398, "learning_rate": 0.0002667070146362453, "loss": 5.2236, "step": 253350 }, { "epoch": 1.527960347800919, "grad_norm": 1.5139775276184082, "learning_rate": 0.0002666942302128791, "loss": 4.7542, "step": 253400 }, { "epoch": 1.5282618395822531, "grad_norm": 1.2480381727218628, "learning_rate": 0.00026668144364192214, "loss": 4.8764, "step": 253450 }, { "epoch": 1.528563331363587, "grad_norm": 1.8396543264389038, "learning_rate": 0.00026666865492360966, "loss": 5.3131, "step": 253500 }, { "epoch": 1.5288648231449211, "grad_norm": 0.9968388676643372, "learning_rate": 0.0002666558640581771, "loss": 5.388, "step": 253550 }, { "epoch": 1.529166314926255, "grad_norm": 1.4418420791625977, "learning_rate": 0.00026664307104585984, "loss": 5.2142, "step": 253600 }, { "epoch": 1.529467806707589, "grad_norm": 1.120851755142212, "learning_rate": 0.00026663027588689337, "loss": 4.7892, "step": 253650 }, { "epoch": 1.5297692984889233, "grad_norm": 1.3472179174423218, "learning_rate": 0.00026661747858151304, "loss": 5.2165, "step": 253700 }, { "epoch": 1.5300707902702573, "grad_norm": 2.096264362335205, "learning_rate": 0.0002666046791299545, "loss": 5.047, "step": 253750 }, { "epoch": 1.5303722820515913, "grad_norm": 1.1197575330734253, "learning_rate": 0.00026659187753245317, "loss": 5.4939, "step": 253800 }, { "epoch": 1.5306737738329255, "grad_norm": 1.9974554777145386, "learning_rate": 0.00026657907378924476, "loss": 5.1638, "step": 253850 }, { "epoch": 1.5309752656142592, "grad_norm": 1.3158202171325684, "learning_rate": 0.0002665662679005649, "loss": 5.2344, "step": 253900 }, { "epoch": 1.5312767573955934, "grad_norm": 1.2525326013565063, "learning_rate": 0.0002665534598666492, "loss": 5.4124, "step": 253950 }, { "epoch": 1.5315782491769274, "grad_norm": 1.3749587535858154, "learning_rate": 0.00026654064968773336, "loss": 4.7116, "step": 254000 }, { "epoch": 1.5318797409582614, "grad_norm": 1.421094298362732, "learning_rate": 0.0002665278373640532, "loss": 4.5175, "step": 254050 }, { "epoch": 1.5321812327395956, "grad_norm": 0.743510365486145, "learning_rate": 0.0002665150228958445, "loss": 4.8429, "step": 254100 }, { "epoch": 1.5324827245209296, "grad_norm": 1.1871696710586548, "learning_rate": 0.0002665022062833431, "loss": 5.0873, "step": 254150 }, { "epoch": 1.5327842163022636, "grad_norm": 1.4720474481582642, "learning_rate": 0.0002664893875267849, "loss": 5.1823, "step": 254200 }, { "epoch": 1.5330857080835978, "grad_norm": 1.1864901781082153, "learning_rate": 0.00026647656662640565, "loss": 4.9726, "step": 254250 }, { "epoch": 1.5333871998649315, "grad_norm": 0.23499995470046997, "learning_rate": 0.00026646374358244143, "loss": 4.9627, "step": 254300 }, { "epoch": 1.5336886916462658, "grad_norm": 0.3640781342983246, "learning_rate": 0.00026645091839512823, "loss": 4.9958, "step": 254350 }, { "epoch": 1.5339901834275997, "grad_norm": 2.1844992637634277, "learning_rate": 0.0002664380910647021, "loss": 5.0837, "step": 254400 }, { "epoch": 1.5342916752089337, "grad_norm": 0.4720163345336914, "learning_rate": 0.000266425261591399, "loss": 5.0755, "step": 254450 }, { "epoch": 1.534593166990268, "grad_norm": 1.7342991828918457, "learning_rate": 0.00026641242997545514, "loss": 5.1743, "step": 254500 }, { "epoch": 1.534894658771602, "grad_norm": 0.8493250012397766, "learning_rate": 0.00026639959621710667, "loss": 5.2389, "step": 254550 }, { "epoch": 1.535196150552936, "grad_norm": 1.4274389743804932, "learning_rate": 0.0002663867603165897, "loss": 5.2377, "step": 254600 }, { "epoch": 1.53549764233427, "grad_norm": 1.6466693878173828, "learning_rate": 0.00026637392227414055, "loss": 5.1146, "step": 254650 }, { "epoch": 1.5357991341156039, "grad_norm": 1.7893829345703125, "learning_rate": 0.00026636108208999534, "loss": 5.0073, "step": 254700 }, { "epoch": 1.536100625896938, "grad_norm": 1.2595546245574951, "learning_rate": 0.00026634823976439054, "loss": 4.9364, "step": 254750 }, { "epoch": 1.536402117678272, "grad_norm": 1.1733556985855103, "learning_rate": 0.0002663353952975624, "loss": 5.2556, "step": 254800 }, { "epoch": 1.536703609459606, "grad_norm": 1.045886754989624, "learning_rate": 0.0002663225486897474, "loss": 5.0464, "step": 254850 }, { "epoch": 1.5370051012409403, "grad_norm": 1.8842172622680664, "learning_rate": 0.00026630969994118184, "loss": 5.1667, "step": 254900 }, { "epoch": 1.5373065930222742, "grad_norm": 1.6844819784164429, "learning_rate": 0.00026629684905210224, "loss": 5.0746, "step": 254950 }, { "epoch": 1.5376080848036082, "grad_norm": 2.0428574085235596, "learning_rate": 0.00026628399602274505, "loss": 5.3039, "step": 255000 }, { "epoch": 1.5376080848036082, "eval_loss": 5.580300331115723, "eval_runtime": 39.1142, "eval_samples_per_second": 13.09, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.311069087447161, "step": 255000 }, { "epoch": 1.5379095765849424, "grad_norm": 2.0572152137756348, "learning_rate": 0.0002662711408533469, "loss": 5.2355, "step": 255050 }, { "epoch": 1.5382110683662762, "grad_norm": 1.8055461645126343, "learning_rate": 0.00026625828354414434, "loss": 5.2068, "step": 255100 }, { "epoch": 1.5385125601476104, "grad_norm": 1.608886957168579, "learning_rate": 0.00026624542409537397, "loss": 5.1275, "step": 255150 }, { "epoch": 1.5388140519289444, "grad_norm": 1.6354142427444458, "learning_rate": 0.00026623256250727246, "loss": 5.1161, "step": 255200 }, { "epoch": 1.5391155437102784, "grad_norm": 1.329006552696228, "learning_rate": 0.0002662196987800765, "loss": 5.2291, "step": 255250 }, { "epoch": 1.5394170354916126, "grad_norm": 2.1005091667175293, "learning_rate": 0.0002662068329140228, "loss": 5.1421, "step": 255300 }, { "epoch": 1.5397185272729466, "grad_norm": 1.3630272150039673, "learning_rate": 0.0002661939649093483, "loss": 5.2154, "step": 255350 }, { "epoch": 1.5400200190542805, "grad_norm": 0.6472405791282654, "learning_rate": 0.0002661810947662896, "loss": 5.179, "step": 255400 }, { "epoch": 1.5403215108356147, "grad_norm": 0.8772655129432678, "learning_rate": 0.00026616822248508364, "loss": 5.0563, "step": 255450 }, { "epoch": 1.5406230026169485, "grad_norm": 0.8431959748268127, "learning_rate": 0.00026615534806596734, "loss": 5.0327, "step": 255500 }, { "epoch": 1.5409244943982827, "grad_norm": 1.2158087491989136, "learning_rate": 0.00026614247150917765, "loss": 4.8523, "step": 255550 }, { "epoch": 1.5412259861796167, "grad_norm": 0.9987437129020691, "learning_rate": 0.0002661295928149515, "loss": 5.162, "step": 255600 }, { "epoch": 1.5415274779609507, "grad_norm": 1.7099109888076782, "learning_rate": 0.00026611671198352595, "loss": 5.243, "step": 255650 }, { "epoch": 1.541828969742285, "grad_norm": 1.5262398719787598, "learning_rate": 0.00026610382901513796, "loss": 5.1421, "step": 255700 }, { "epoch": 1.5421304615236189, "grad_norm": 1.2232542037963867, "learning_rate": 0.0002660909439100247, "loss": 5.2287, "step": 255750 }, { "epoch": 1.5424319533049529, "grad_norm": 1.946831226348877, "learning_rate": 0.0002660780566684233, "loss": 4.8246, "step": 255800 }, { "epoch": 1.542733445086287, "grad_norm": 0.613532304763794, "learning_rate": 0.00026606516729057094, "loss": 4.498, "step": 255850 }, { "epoch": 1.5430349368676208, "grad_norm": 1.1580721139907837, "learning_rate": 0.0002660522757767048, "loss": 5.5628, "step": 255900 }, { "epoch": 1.543336428648955, "grad_norm": 1.0008976459503174, "learning_rate": 0.0002660393821270622, "loss": 4.8772, "step": 255950 }, { "epoch": 1.543637920430289, "grad_norm": 0.7671777009963989, "learning_rate": 0.0002660264863418803, "loss": 4.013, "step": 256000 }, { "epoch": 1.543939412211623, "grad_norm": 1.9476598501205444, "learning_rate": 0.00026601358842139646, "loss": 4.8921, "step": 256050 }, { "epoch": 1.5442409039929572, "grad_norm": 2.1349873542785645, "learning_rate": 0.00026600068836584817, "loss": 5.6801, "step": 256100 }, { "epoch": 1.5445423957742912, "grad_norm": 1.1572211980819702, "learning_rate": 0.0002659877861754727, "loss": 4.7747, "step": 256150 }, { "epoch": 1.5448438875556252, "grad_norm": 1.2181235551834106, "learning_rate": 0.0002659748818505076, "loss": 4.6208, "step": 256200 }, { "epoch": 1.5451453793369594, "grad_norm": 1.2749308347702026, "learning_rate": 0.00026596197539119023, "loss": 5.1676, "step": 256250 }, { "epoch": 1.5454468711182932, "grad_norm": 1.0494893789291382, "learning_rate": 0.0002659490667977582, "loss": 4.9887, "step": 256300 }, { "epoch": 1.5457483628996274, "grad_norm": 1.171764850616455, "learning_rate": 0.00026593615607044914, "loss": 4.7464, "step": 256350 }, { "epoch": 1.5460498546809613, "grad_norm": 1.2489598989486694, "learning_rate": 0.0002659232432095005, "loss": 4.8889, "step": 256400 }, { "epoch": 1.5463513464622953, "grad_norm": 0.5655823945999146, "learning_rate": 0.00026591032821515005, "loss": 5.0873, "step": 256450 }, { "epoch": 1.5466528382436295, "grad_norm": 1.994828224182129, "learning_rate": 0.00026589741108763536, "loss": 5.0547, "step": 256500 }, { "epoch": 1.5469543300249635, "grad_norm": 1.672869086265564, "learning_rate": 0.0002658844918271943, "loss": 5.1876, "step": 256550 }, { "epoch": 1.5472558218062975, "grad_norm": 2.6212565898895264, "learning_rate": 0.00026587157043406445, "loss": 5.2111, "step": 256600 }, { "epoch": 1.5475573135876317, "grad_norm": 1.1755433082580566, "learning_rate": 0.00026585864690848376, "loss": 4.8916, "step": 256650 }, { "epoch": 1.5478588053689657, "grad_norm": 1.0887187719345093, "learning_rate": 0.00026584572125069, "loss": 4.9582, "step": 256700 }, { "epoch": 1.5481602971502997, "grad_norm": 1.3155168294906616, "learning_rate": 0.00026583279346092107, "loss": 5.1647, "step": 256750 }, { "epoch": 1.5484617889316339, "grad_norm": 3.225017786026001, "learning_rate": 0.00026581986353941487, "loss": 5.2116, "step": 256800 }, { "epoch": 1.5487632807129676, "grad_norm": 1.1369637250900269, "learning_rate": 0.0002658069314864094, "loss": 5.1705, "step": 256850 }, { "epoch": 1.5490647724943019, "grad_norm": 1.2194880247116089, "learning_rate": 0.00026579399730214257, "loss": 4.8499, "step": 256900 }, { "epoch": 1.5493662642756358, "grad_norm": 1.2620103359222412, "learning_rate": 0.00026578106098685243, "loss": 4.8729, "step": 256950 }, { "epoch": 1.5496677560569698, "grad_norm": 0.8470771312713623, "learning_rate": 0.0002657681225407772, "loss": 5.1371, "step": 257000 }, { "epoch": 1.549969247838304, "grad_norm": 2.3422887325286865, "learning_rate": 0.0002657551819641548, "loss": 5.0273, "step": 257050 }, { "epoch": 1.550270739619638, "grad_norm": 0.7045128345489502, "learning_rate": 0.00026574223925722353, "loss": 4.9752, "step": 257100 }, { "epoch": 1.550572231400972, "grad_norm": 1.8162422180175781, "learning_rate": 0.00026572929442022147, "loss": 5.424, "step": 257150 }, { "epoch": 1.5508737231823062, "grad_norm": 1.4193987846374512, "learning_rate": 0.0002657163474533869, "loss": 5.2022, "step": 257200 }, { "epoch": 1.55117521496364, "grad_norm": 2.226893663406372, "learning_rate": 0.00026570339835695815, "loss": 5.5806, "step": 257250 }, { "epoch": 1.5514767067449742, "grad_norm": 1.4941105842590332, "learning_rate": 0.0002656904471311734, "loss": 5.2569, "step": 257300 }, { "epoch": 1.5517781985263082, "grad_norm": 0.8624247312545776, "learning_rate": 0.0002656774937762711, "loss": 4.811, "step": 257350 }, { "epoch": 1.5520796903076421, "grad_norm": 1.884210467338562, "learning_rate": 0.0002656645382924896, "loss": 4.7576, "step": 257400 }, { "epoch": 1.5523811820889764, "grad_norm": 1.8452694416046143, "learning_rate": 0.0002656515806800674, "loss": 5.3083, "step": 257450 }, { "epoch": 1.5526826738703103, "grad_norm": 0.970173716545105, "learning_rate": 0.0002656386209392428, "loss": 5.176, "step": 257500 }, { "epoch": 1.5529841656516443, "grad_norm": 0.8442442417144775, "learning_rate": 0.00026562565907025445, "loss": 4.7379, "step": 257550 }, { "epoch": 1.5532856574329785, "grad_norm": 1.8012453317642212, "learning_rate": 0.0002656126950733409, "loss": 5.6269, "step": 257600 }, { "epoch": 1.5535871492143123, "grad_norm": 0.3077847361564636, "learning_rate": 0.0002655997289487406, "loss": 4.8084, "step": 257650 }, { "epoch": 1.5538886409956465, "grad_norm": 0.245221808552742, "learning_rate": 0.0002655867606966923, "loss": 5.0979, "step": 257700 }, { "epoch": 1.5541901327769805, "grad_norm": 0.8285914659500122, "learning_rate": 0.0002655737903174346, "loss": 5.1134, "step": 257750 }, { "epoch": 1.5544916245583145, "grad_norm": 1.317498803138733, "learning_rate": 0.00026556081781120626, "loss": 5.1172, "step": 257800 }, { "epoch": 1.5547931163396487, "grad_norm": 1.0553263425827026, "learning_rate": 0.00026554784317824596, "loss": 5.064, "step": 257850 }, { "epoch": 1.5550946081209827, "grad_norm": 0.5918529033660889, "learning_rate": 0.0002655348664187925, "loss": 4.6217, "step": 257900 }, { "epoch": 1.5553960999023166, "grad_norm": 0.8949860334396362, "learning_rate": 0.0002655218875330847, "loss": 4.8201, "step": 257950 }, { "epoch": 1.5556975916836508, "grad_norm": 0.9287566542625427, "learning_rate": 0.0002655089065213614, "loss": 5.5149, "step": 258000 }, { "epoch": 1.5556975916836508, "eval_loss": 5.567708492279053, "eval_runtime": 39.0611, "eval_samples_per_second": 13.108, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.236054623959762, "step": 258000 }, { "epoch": 1.5559990834649846, "grad_norm": 0.5429190397262573, "learning_rate": 0.0002654959233838616, "loss": 5.1612, "step": 258050 }, { "epoch": 1.5563005752463188, "grad_norm": 1.4499107599258423, "learning_rate": 0.00026548293812082407, "loss": 5.4947, "step": 258100 }, { "epoch": 1.5566020670276528, "grad_norm": 2.0373449325561523, "learning_rate": 0.00026546995073248786, "loss": 5.5376, "step": 258150 }, { "epoch": 1.5569035588089868, "grad_norm": 1.3210418224334717, "learning_rate": 0.000265456961219092, "loss": 5.1507, "step": 258200 }, { "epoch": 1.557205050590321, "grad_norm": 1.029836654663086, "learning_rate": 0.0002654439695808756, "loss": 4.7289, "step": 258250 }, { "epoch": 1.557506542371655, "grad_norm": 1.2661291360855103, "learning_rate": 0.0002654309758180776, "loss": 5.1045, "step": 258300 }, { "epoch": 1.557808034152989, "grad_norm": 1.6271905899047852, "learning_rate": 0.0002654179799309373, "loss": 4.7007, "step": 258350 }, { "epoch": 1.5581095259343232, "grad_norm": 1.684207797050476, "learning_rate": 0.00026540498191969373, "loss": 5.1096, "step": 258400 }, { "epoch": 1.558411017715657, "grad_norm": 1.1797358989715576, "learning_rate": 0.00026539198178458616, "loss": 5.3712, "step": 258450 }, { "epoch": 1.5587125094969911, "grad_norm": 1.4779611825942993, "learning_rate": 0.00026537897952585385, "loss": 5.0367, "step": 258500 }, { "epoch": 1.5590140012783251, "grad_norm": 1.3354727029800415, "learning_rate": 0.0002653659751437361, "loss": 5.2539, "step": 258550 }, { "epoch": 1.559315493059659, "grad_norm": 1.2245978116989136, "learning_rate": 0.0002653529686384722, "loss": 5.2746, "step": 258600 }, { "epoch": 1.5596169848409933, "grad_norm": 1.5121885538101196, "learning_rate": 0.0002653399600103015, "loss": 5.5309, "step": 258650 }, { "epoch": 1.5599184766223273, "grad_norm": 1.0795072317123413, "learning_rate": 0.0002653269492594635, "loss": 5.4447, "step": 258700 }, { "epoch": 1.5602199684036613, "grad_norm": 1.5437792539596558, "learning_rate": 0.0002653139363861975, "loss": 4.7134, "step": 258750 }, { "epoch": 1.5605214601849955, "grad_norm": 1.7244700193405151, "learning_rate": 0.00026530092139074316, "loss": 5.1661, "step": 258800 }, { "epoch": 1.5608229519663293, "grad_norm": 1.386286735534668, "learning_rate": 0.0002652879042733398, "loss": 4.8062, "step": 258850 }, { "epoch": 1.5611244437476635, "grad_norm": 1.4299813508987427, "learning_rate": 0.00026527488503422713, "loss": 5.2359, "step": 258900 }, { "epoch": 1.5614259355289974, "grad_norm": 1.1872079372406006, "learning_rate": 0.00026526186367364473, "loss": 4.8378, "step": 258950 }, { "epoch": 1.5617274273103314, "grad_norm": 2.6583433151245117, "learning_rate": 0.0002652488401918322, "loss": 5.0415, "step": 259000 }, { "epoch": 1.5620289190916656, "grad_norm": 2.67094349861145, "learning_rate": 0.00026523581458902926, "loss": 5.6367, "step": 259050 }, { "epoch": 1.5623304108729996, "grad_norm": 1.1412757635116577, "learning_rate": 0.0002652227868654755, "loss": 5.2976, "step": 259100 }, { "epoch": 1.5626319026543336, "grad_norm": 2.1302220821380615, "learning_rate": 0.0002652097570214109, "loss": 5.2064, "step": 259150 }, { "epoch": 1.5629333944356678, "grad_norm": 2.6227307319641113, "learning_rate": 0.0002651967250570751, "loss": 5.2522, "step": 259200 }, { "epoch": 1.5632348862170016, "grad_norm": 0.9532953500747681, "learning_rate": 0.000265183690972708, "loss": 4.3703, "step": 259250 }, { "epoch": 1.5635363779983358, "grad_norm": 1.5307865142822266, "learning_rate": 0.00026517065476854934, "loss": 5.1117, "step": 259300 }, { "epoch": 1.5638378697796698, "grad_norm": 1.6705530881881714, "learning_rate": 0.00026515761644483926, "loss": 4.8123, "step": 259350 }, { "epoch": 1.5641393615610037, "grad_norm": 2.3501040935516357, "learning_rate": 0.0002651445760018175, "loss": 5.3799, "step": 259400 }, { "epoch": 1.564440853342338, "grad_norm": 1.1247245073318481, "learning_rate": 0.00026513153343972414, "loss": 5.3258, "step": 259450 }, { "epoch": 1.564742345123672, "grad_norm": 1.6632877588272095, "learning_rate": 0.0002651184887587992, "loss": 4.9789, "step": 259500 }, { "epoch": 1.565043836905006, "grad_norm": 1.417670726776123, "learning_rate": 0.0002651054419592829, "loss": 4.8612, "step": 259550 }, { "epoch": 1.5653453286863401, "grad_norm": 1.3396788835525513, "learning_rate": 0.00026509239304141504, "loss": 5.3658, "step": 259600 }, { "epoch": 1.565646820467674, "grad_norm": 1.6158634424209595, "learning_rate": 0.000265079342005436, "loss": 5.1766, "step": 259650 }, { "epoch": 1.565948312249008, "grad_norm": 1.8065342903137207, "learning_rate": 0.0002650662888515859, "loss": 5.0379, "step": 259700 }, { "epoch": 1.566249804030342, "grad_norm": 1.8895435333251953, "learning_rate": 0.0002650532335801049, "loss": 4.9625, "step": 259750 }, { "epoch": 1.566551295811676, "grad_norm": 1.17500638961792, "learning_rate": 0.0002650401761912334, "loss": 5.294, "step": 259800 }, { "epoch": 1.5668527875930103, "grad_norm": 1.2574901580810547, "learning_rate": 0.00026502711668521166, "loss": 5.0442, "step": 259850 }, { "epoch": 1.5671542793743443, "grad_norm": 1.1663211584091187, "learning_rate": 0.00026501405506227987, "loss": 5.034, "step": 259900 }, { "epoch": 1.5674557711556782, "grad_norm": 2.212088108062744, "learning_rate": 0.0002650009913226786, "loss": 5.0543, "step": 259950 }, { "epoch": 1.5677572629370125, "grad_norm": 1.2318631410598755, "learning_rate": 0.0002649879254666482, "loss": 4.9667, "step": 260000 }, { "epoch": 1.5680587547183462, "grad_norm": 1.2771081924438477, "learning_rate": 0.0002649748574944291, "loss": 5.0293, "step": 260050 }, { "epoch": 1.5683602464996804, "grad_norm": 0.9471230506896973, "learning_rate": 0.00026496178740626186, "loss": 4.806, "step": 260100 }, { "epoch": 1.5686617382810144, "grad_norm": 1.4616409540176392, "learning_rate": 0.000264948715202387, "loss": 5.137, "step": 260150 }, { "epoch": 1.5689632300623484, "grad_norm": 2.8394157886505127, "learning_rate": 0.00026493564088304504, "loss": 4.5909, "step": 260200 }, { "epoch": 1.5692647218436826, "grad_norm": 2.0447936058044434, "learning_rate": 0.00026492256444847663, "loss": 4.6842, "step": 260250 }, { "epoch": 1.5695662136250166, "grad_norm": 0.4669080674648285, "learning_rate": 0.0002649094858989224, "loss": 5.1839, "step": 260300 }, { "epoch": 1.5698677054063506, "grad_norm": 1.1327120065689087, "learning_rate": 0.0002648964052346231, "loss": 5.0973, "step": 260350 }, { "epoch": 1.5701691971876848, "grad_norm": 1.2370020151138306, "learning_rate": 0.00026488332245581947, "loss": 5.1443, "step": 260400 }, { "epoch": 1.5704706889690188, "grad_norm": 0.5478556156158447, "learning_rate": 0.0002648702375627522, "loss": 4.9038, "step": 260450 }, { "epoch": 1.5707721807503527, "grad_norm": 1.4123655557632446, "learning_rate": 0.00026485715055566214, "loss": 5.2814, "step": 260500 }, { "epoch": 1.571073672531687, "grad_norm": 1.3851900100708008, "learning_rate": 0.0002648440614347901, "loss": 4.8696, "step": 260550 }, { "epoch": 1.5713751643130207, "grad_norm": 2.3860349655151367, "learning_rate": 0.00026483097020037704, "loss": 5.2403, "step": 260600 }, { "epoch": 1.571676656094355, "grad_norm": 1.4376654624938965, "learning_rate": 0.00026481787685266387, "loss": 4.9175, "step": 260650 }, { "epoch": 1.571978147875689, "grad_norm": 1.4607172012329102, "learning_rate": 0.00026480478139189145, "loss": 5.4996, "step": 260700 }, { "epoch": 1.5722796396570229, "grad_norm": 2.770872116088867, "learning_rate": 0.0002647916838183009, "loss": 4.9446, "step": 260750 }, { "epoch": 1.572581131438357, "grad_norm": 0.9771448373794556, "learning_rate": 0.00026477858413213325, "loss": 5.3992, "step": 260800 }, { "epoch": 1.572882623219691, "grad_norm": 1.6964033842086792, "learning_rate": 0.00026476548233362956, "loss": 4.9685, "step": 260850 }, { "epoch": 1.573184115001025, "grad_norm": 0.9613102674484253, "learning_rate": 0.0002647523784230309, "loss": 5.1225, "step": 260900 }, { "epoch": 1.5734856067823593, "grad_norm": 1.1227079629898071, "learning_rate": 0.0002647392724005785, "loss": 5.1532, "step": 260950 }, { "epoch": 1.573787098563693, "grad_norm": 2.354346513748169, "learning_rate": 0.00026472616426651354, "loss": 4.8094, "step": 261000 }, { "epoch": 1.573787098563693, "eval_loss": 5.569216728210449, "eval_runtime": 38.8757, "eval_samples_per_second": 13.17, "eval_steps_per_second": 6.585, "eval_tts_loss": 7.284273498144919, "step": 261000 }, { "epoch": 1.5740885903450272, "grad_norm": 1.8445298671722412, "learning_rate": 0.00026471305402107724, "loss": 5.5235, "step": 261050 }, { "epoch": 1.5743900821263612, "grad_norm": 1.2897545099258423, "learning_rate": 0.0002646999416645109, "loss": 5.0234, "step": 261100 }, { "epoch": 1.5746915739076952, "grad_norm": 1.2313668727874756, "learning_rate": 0.0002646868271970558, "loss": 5.0172, "step": 261150 }, { "epoch": 1.5749930656890294, "grad_norm": 1.3795243501663208, "learning_rate": 0.0002646737106189533, "loss": 4.923, "step": 261200 }, { "epoch": 1.5752945574703634, "grad_norm": 1.2228889465332031, "learning_rate": 0.0002646605919304449, "loss": 5.2789, "step": 261250 }, { "epoch": 1.5755960492516974, "grad_norm": 0.6764228940010071, "learning_rate": 0.0002646474711317718, "loss": 5.0264, "step": 261300 }, { "epoch": 1.5758975410330316, "grad_norm": 1.2703052759170532, "learning_rate": 0.00026463434822317564, "loss": 4.9543, "step": 261350 }, { "epoch": 1.5761990328143654, "grad_norm": 1.716489553451538, "learning_rate": 0.00026462122320489796, "loss": 5.2195, "step": 261400 }, { "epoch": 1.5765005245956996, "grad_norm": 2.057558298110962, "learning_rate": 0.0002646080960771802, "loss": 5.0921, "step": 261450 }, { "epoch": 1.5768020163770335, "grad_norm": 1.220347285270691, "learning_rate": 0.00026459496684026393, "loss": 5.1724, "step": 261500 }, { "epoch": 1.5771035081583675, "grad_norm": 1.6234582662582397, "learning_rate": 0.0002645818354943909, "loss": 5.0044, "step": 261550 }, { "epoch": 1.5774049999397017, "grad_norm": 1.589263916015625, "learning_rate": 0.00026456870203980267, "loss": 4.9811, "step": 261600 }, { "epoch": 1.5777064917210357, "grad_norm": 1.409324288368225, "learning_rate": 0.00026455556647674103, "loss": 5.5191, "step": 261650 }, { "epoch": 1.5780079835023697, "grad_norm": 1.594559669494629, "learning_rate": 0.00026454242880544763, "loss": 4.7965, "step": 261700 }, { "epoch": 1.578309475283704, "grad_norm": 1.3594106435775757, "learning_rate": 0.00026452928902616433, "loss": 5.0856, "step": 261750 }, { "epoch": 1.5786109670650377, "grad_norm": 3.239436149597168, "learning_rate": 0.00026451614713913287, "loss": 4.6814, "step": 261800 }, { "epoch": 1.5789124588463719, "grad_norm": 2.759812355041504, "learning_rate": 0.0002645030031445951, "loss": 4.8362, "step": 261850 }, { "epoch": 1.5792139506277059, "grad_norm": 1.3221324682235718, "learning_rate": 0.00026448985704279304, "loss": 5.046, "step": 261900 }, { "epoch": 1.5795154424090398, "grad_norm": 1.1139609813690186, "learning_rate": 0.0002644767088339685, "loss": 4.8839, "step": 261950 }, { "epoch": 1.579816934190374, "grad_norm": 1.086215853691101, "learning_rate": 0.00026446355851836355, "loss": 4.8844, "step": 262000 }, { "epoch": 1.580118425971708, "grad_norm": 1.443249225616455, "learning_rate": 0.0002644504060962201, "loss": 5.1635, "step": 262050 }, { "epoch": 1.580419917753042, "grad_norm": 0.5397782325744629, "learning_rate": 0.0002644372515677803, "loss": 5.2521, "step": 262100 }, { "epoch": 1.5807214095343762, "grad_norm": 1.3245214223861694, "learning_rate": 0.00026442409493328614, "loss": 5.3072, "step": 262150 }, { "epoch": 1.58102290131571, "grad_norm": 1.1984272003173828, "learning_rate": 0.0002644109361929798, "loss": 5.8014, "step": 262200 }, { "epoch": 1.5813243930970442, "grad_norm": 0.8003365993499756, "learning_rate": 0.00026439777534710356, "loss": 4.3954, "step": 262250 }, { "epoch": 1.5816258848783782, "grad_norm": 1.682599663734436, "learning_rate": 0.0002643846123958994, "loss": 4.943, "step": 262300 }, { "epoch": 1.5819273766597122, "grad_norm": 1.2315877676010132, "learning_rate": 0.00026437144733960975, "loss": 5.3015, "step": 262350 }, { "epoch": 1.5822288684410464, "grad_norm": 1.6016038656234741, "learning_rate": 0.0002643582801784768, "loss": 5.0729, "step": 262400 }, { "epoch": 1.5825303602223804, "grad_norm": 1.7511265277862549, "learning_rate": 0.00026434511091274294, "loss": 5.3194, "step": 262450 }, { "epoch": 1.5828318520037143, "grad_norm": 1.8641128540039062, "learning_rate": 0.00026433193954265047, "loss": 5.3734, "step": 262500 }, { "epoch": 1.5831333437850486, "grad_norm": 2.0171077251434326, "learning_rate": 0.0002643187660684418, "loss": 5.2072, "step": 262550 }, { "epoch": 1.5834348355663823, "grad_norm": 1.1739977598190308, "learning_rate": 0.0002643055904903594, "loss": 5.2242, "step": 262600 }, { "epoch": 1.5837363273477165, "grad_norm": 1.5557090044021606, "learning_rate": 0.00026429241280864573, "loss": 4.862, "step": 262650 }, { "epoch": 1.5840378191290505, "grad_norm": 1.0894218683242798, "learning_rate": 0.00026427923302354325, "loss": 5.2891, "step": 262700 }, { "epoch": 1.5843393109103845, "grad_norm": 1.9163111448287964, "learning_rate": 0.0002642660511352946, "loss": 5.385, "step": 262750 }, { "epoch": 1.5846408026917187, "grad_norm": 1.7702001333236694, "learning_rate": 0.00026425286714414244, "loss": 5.2428, "step": 262800 }, { "epoch": 1.5849422944730527, "grad_norm": 2.133155345916748, "learning_rate": 0.0002642396810503292, "loss": 5.0534, "step": 262850 }, { "epoch": 1.5852437862543867, "grad_norm": 1.2932690382003784, "learning_rate": 0.0002642264928540977, "loss": 5.068, "step": 262900 }, { "epoch": 1.5855452780357209, "grad_norm": 1.331229567527771, "learning_rate": 0.0002642133025556906, "loss": 5.1426, "step": 262950 }, { "epoch": 1.5858467698170546, "grad_norm": 1.683717131614685, "learning_rate": 0.00026420011015535063, "loss": 5.4427, "step": 263000 }, { "epoch": 1.5861482615983888, "grad_norm": 1.3031259775161743, "learning_rate": 0.00026418691565332066, "loss": 5.0106, "step": 263050 }, { "epoch": 1.5864497533797228, "grad_norm": 1.0116435289382935, "learning_rate": 0.0002641737190498434, "loss": 4.8043, "step": 263100 }, { "epoch": 1.5867512451610568, "grad_norm": 1.1979273557662964, "learning_rate": 0.0002641605203451618, "loss": 4.7441, "step": 263150 }, { "epoch": 1.587052736942391, "grad_norm": 1.8953503370285034, "learning_rate": 0.00026414731953951876, "loss": 5.0707, "step": 263200 }, { "epoch": 1.587354228723725, "grad_norm": 1.0933196544647217, "learning_rate": 0.0002641341166331572, "loss": 5.2859, "step": 263250 }, { "epoch": 1.587655720505059, "grad_norm": 0.8667054772377014, "learning_rate": 0.00026412091162632007, "loss": 5.3904, "step": 263300 }, { "epoch": 1.5879572122863932, "grad_norm": 1.2681547403335571, "learning_rate": 0.00026410770451925047, "loss": 5.4919, "step": 263350 }, { "epoch": 1.588258704067727, "grad_norm": 1.1696900129318237, "learning_rate": 0.00026409449531219136, "loss": 4.9459, "step": 263400 }, { "epoch": 1.5885601958490612, "grad_norm": 0.8956561088562012, "learning_rate": 0.0002640812840053859, "loss": 5.1347, "step": 263450 }, { "epoch": 1.5888616876303951, "grad_norm": 1.598671793937683, "learning_rate": 0.00026406807059907726, "loss": 4.9468, "step": 263500 }, { "epoch": 1.5891631794117291, "grad_norm": 2.128286600112915, "learning_rate": 0.0002640548550935085, "loss": 5.0947, "step": 263550 }, { "epoch": 1.5894646711930633, "grad_norm": 1.8817132711410522, "learning_rate": 0.0002640416374889229, "loss": 5.2298, "step": 263600 }, { "epoch": 1.5897661629743973, "grad_norm": 1.1040279865264893, "learning_rate": 0.0002640284177855637, "loss": 5.0769, "step": 263650 }, { "epoch": 1.5900676547557313, "grad_norm": 1.5901402235031128, "learning_rate": 0.00026401519598367426, "loss": 5.0149, "step": 263700 }, { "epoch": 1.5903691465370655, "grad_norm": 1.3388121128082275, "learning_rate": 0.00026400197208349777, "loss": 5.0655, "step": 263750 }, { "epoch": 1.5906706383183993, "grad_norm": 1.2653738260269165, "learning_rate": 0.00026398874608527775, "loss": 5.2234, "step": 263800 }, { "epoch": 1.5909721300997335, "grad_norm": 1.1694892644882202, "learning_rate": 0.00026397551798925745, "loss": 5.1538, "step": 263850 }, { "epoch": 1.5912736218810675, "grad_norm": 1.2370305061340332, "learning_rate": 0.0002639622877956804, "loss": 4.9862, "step": 263900 }, { "epoch": 1.5915751136624015, "grad_norm": 1.3243701457977295, "learning_rate": 0.0002639490555047902, "loss": 4.6384, "step": 263950 }, { "epoch": 1.5918766054437357, "grad_norm": 1.5401606559753418, "learning_rate": 0.0002639358211168301, "loss": 4.8445, "step": 264000 }, { "epoch": 1.5918766054437357, "eval_loss": 5.551597595214844, "eval_runtime": 39.0907, "eval_samples_per_second": 13.098, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.325900098413393, "step": 264000 }, { "epoch": 1.5921780972250696, "grad_norm": 2.049849510192871, "learning_rate": 0.0002639225846320439, "loss": 5.3995, "step": 264050 }, { "epoch": 1.5924795890064036, "grad_norm": 1.692319393157959, "learning_rate": 0.00026390934605067504, "loss": 4.9169, "step": 264100 }, { "epoch": 1.5927810807877378, "grad_norm": 1.1998205184936523, "learning_rate": 0.0002638961053729673, "loss": 5.1349, "step": 264150 }, { "epoch": 1.5930825725690718, "grad_norm": 1.6680009365081787, "learning_rate": 0.0002638828625991642, "loss": 5.2683, "step": 264200 }, { "epoch": 1.5933840643504058, "grad_norm": 1.939265251159668, "learning_rate": 0.00026386961772950954, "loss": 5.4705, "step": 264250 }, { "epoch": 1.59368555613174, "grad_norm": 1.5170655250549316, "learning_rate": 0.00026385637076424713, "loss": 5.4806, "step": 264300 }, { "epoch": 1.5939870479130738, "grad_norm": 0.5568525791168213, "learning_rate": 0.00026384312170362064, "loss": 4.862, "step": 264350 }, { "epoch": 1.594288539694408, "grad_norm": 2.1243860721588135, "learning_rate": 0.000263829870547874, "loss": 5.4579, "step": 264400 }, { "epoch": 1.594590031475742, "grad_norm": 1.8553266525268555, "learning_rate": 0.000263816617297251, "loss": 4.812, "step": 264450 }, { "epoch": 1.594891523257076, "grad_norm": 1.1450613737106323, "learning_rate": 0.0002638033619519956, "loss": 4.9347, "step": 264500 }, { "epoch": 1.5951930150384102, "grad_norm": 1.0170390605926514, "learning_rate": 0.00026379010451235176, "loss": 5.1763, "step": 264550 }, { "epoch": 1.5954945068197441, "grad_norm": 1.2894370555877686, "learning_rate": 0.0002637768449785634, "loss": 4.9032, "step": 264600 }, { "epoch": 1.5957959986010781, "grad_norm": 1.8067182302474976, "learning_rate": 0.0002637635833508746, "loss": 4.9421, "step": 264650 }, { "epoch": 1.5960974903824123, "grad_norm": 0.7090394496917725, "learning_rate": 0.0002637503196295294, "loss": 5.1936, "step": 264700 }, { "epoch": 1.596398982163746, "grad_norm": 1.280145525932312, "learning_rate": 0.0002637370538147718, "loss": 5.3599, "step": 264750 }, { "epoch": 1.5967004739450803, "grad_norm": 0.9792385697364807, "learning_rate": 0.0002637237859068461, "loss": 4.9692, "step": 264800 }, { "epoch": 1.5970019657264143, "grad_norm": 1.6520475149154663, "learning_rate": 0.0002637105159059965, "loss": 5.3447, "step": 264850 }, { "epoch": 1.5973034575077483, "grad_norm": 1.7902089357376099, "learning_rate": 0.000263697243812467, "loss": 5.3351, "step": 264900 }, { "epoch": 1.5976049492890825, "grad_norm": 1.2247096300125122, "learning_rate": 0.00026368396962650204, "loss": 4.9311, "step": 264950 }, { "epoch": 1.5979064410704165, "grad_norm": 1.172075629234314, "learning_rate": 0.00026367069334834583, "loss": 4.9757, "step": 265000 }, { "epoch": 1.5982079328517504, "grad_norm": 0.5325053334236145, "learning_rate": 0.0002636574149782428, "loss": 4.9449, "step": 265050 }, { "epoch": 1.5985094246330847, "grad_norm": 0.5123977661132812, "learning_rate": 0.00026364413451643716, "loss": 5.5928, "step": 265100 }, { "epoch": 1.5988109164144184, "grad_norm": 1.5965527296066284, "learning_rate": 0.00026363085196317345, "loss": 5.2237, "step": 265150 }, { "epoch": 1.5991124081957526, "grad_norm": 1.2511743307113647, "learning_rate": 0.00026361756731869604, "loss": 4.8564, "step": 265200 }, { "epoch": 1.5994138999770866, "grad_norm": 1.4621014595031738, "learning_rate": 0.00026360428058324947, "loss": 5.1944, "step": 265250 }, { "epoch": 1.5997153917584206, "grad_norm": 1.7092828750610352, "learning_rate": 0.0002635909917570782, "loss": 5.267, "step": 265300 }, { "epoch": 1.6000168835397548, "grad_norm": 1.4792275428771973, "learning_rate": 0.00026357770084042684, "loss": 4.9136, "step": 265350 }, { "epoch": 1.6003183753210888, "grad_norm": 1.682801365852356, "learning_rate": 0.00026356440783354004, "loss": 4.903, "step": 265400 }, { "epoch": 1.6006198671024228, "grad_norm": 1.697981595993042, "learning_rate": 0.0002635511127366623, "loss": 4.9976, "step": 265450 }, { "epoch": 1.600921358883757, "grad_norm": 1.2019509077072144, "learning_rate": 0.0002635378155500384, "loss": 4.8556, "step": 265500 }, { "epoch": 1.6012228506650907, "grad_norm": 1.5083388090133667, "learning_rate": 0.00026352451627391306, "loss": 4.7982, "step": 265550 }, { "epoch": 1.601524342446425, "grad_norm": 1.308143973350525, "learning_rate": 0.000263511214908531, "loss": 5.236, "step": 265600 }, { "epoch": 1.601825834227759, "grad_norm": 1.2611157894134521, "learning_rate": 0.000263497911454137, "loss": 5.397, "step": 265650 }, { "epoch": 1.602127326009093, "grad_norm": 1.847076654434204, "learning_rate": 0.0002634846059109759, "loss": 5.5032, "step": 265700 }, { "epoch": 1.6024288177904271, "grad_norm": 2.208761692047119, "learning_rate": 0.00026347129827929264, "loss": 4.9505, "step": 265750 }, { "epoch": 1.602730309571761, "grad_norm": 1.2366176843643188, "learning_rate": 0.000263457988559332, "loss": 4.7906, "step": 265800 }, { "epoch": 1.603031801353095, "grad_norm": 1.192304015159607, "learning_rate": 0.000263444676751339, "loss": 5.1709, "step": 265850 }, { "epoch": 1.6033332931344293, "grad_norm": 1.7864917516708374, "learning_rate": 0.00026343136285555865, "loss": 4.9747, "step": 265900 }, { "epoch": 1.603634784915763, "grad_norm": 2.128044605255127, "learning_rate": 0.00026341804687223593, "loss": 5.0428, "step": 265950 }, { "epoch": 1.6039362766970973, "grad_norm": 1.7174742221832275, "learning_rate": 0.0002634047288016159, "loss": 4.9716, "step": 266000 }, { "epoch": 1.6042377684784312, "grad_norm": 1.1261513233184814, "learning_rate": 0.0002633914086439437, "loss": 5.0942, "step": 266050 }, { "epoch": 1.6045392602597652, "grad_norm": 1.1829582452774048, "learning_rate": 0.0002633780863994644, "loss": 5.2496, "step": 266100 }, { "epoch": 1.6048407520410994, "grad_norm": 1.361994743347168, "learning_rate": 0.00026336476206842327, "loss": 4.8019, "step": 266150 }, { "epoch": 1.6051422438224334, "grad_norm": 1.7324262857437134, "learning_rate": 0.00026335143565106545, "loss": 5.4574, "step": 266200 }, { "epoch": 1.6054437356037674, "grad_norm": 1.3718762397766113, "learning_rate": 0.00026333810714763625, "loss": 4.9364, "step": 266250 }, { "epoch": 1.6057452273851016, "grad_norm": 1.3583314418792725, "learning_rate": 0.00026332477655838086, "loss": 4.9828, "step": 266300 }, { "epoch": 1.6060467191664354, "grad_norm": 1.936055302619934, "learning_rate": 0.00026331144388354473, "loss": 4.8481, "step": 266350 }, { "epoch": 1.6063482109477696, "grad_norm": 1.2942053079605103, "learning_rate": 0.0002632981091233731, "loss": 5.3731, "step": 266400 }, { "epoch": 1.6066497027291036, "grad_norm": 1.5853357315063477, "learning_rate": 0.0002632847722781116, "loss": 5.1708, "step": 266450 }, { "epoch": 1.6069511945104376, "grad_norm": 1.5230449438095093, "learning_rate": 0.00026327143334800546, "loss": 4.9495, "step": 266500 }, { "epoch": 1.6072526862917718, "grad_norm": 1.8183894157409668, "learning_rate": 0.00026325809233330023, "loss": 4.9087, "step": 266550 }, { "epoch": 1.6075541780731057, "grad_norm": 0.728563129901886, "learning_rate": 0.0002632447492342414, "loss": 5.3608, "step": 266600 }, { "epoch": 1.6078556698544397, "grad_norm": 1.7745957374572754, "learning_rate": 0.0002632314040510746, "loss": 4.9828, "step": 266650 }, { "epoch": 1.608157161635774, "grad_norm": 1.4239439964294434, "learning_rate": 0.00026321805678404547, "loss": 5.2836, "step": 266700 }, { "epoch": 1.6084586534171077, "grad_norm": 0.7648910880088806, "learning_rate": 0.00026320470743339947, "loss": 5.0084, "step": 266750 }, { "epoch": 1.608760145198442, "grad_norm": 1.5229053497314453, "learning_rate": 0.00026319135599938245, "loss": 4.7394, "step": 266800 }, { "epoch": 1.6090616369797759, "grad_norm": 0.3071628510951996, "learning_rate": 0.00026317800248224003, "loss": 4.8881, "step": 266850 }, { "epoch": 1.6093631287611099, "grad_norm": 1.3051646947860718, "learning_rate": 0.000263164646882218, "loss": 5.3085, "step": 266900 }, { "epoch": 1.609664620542444, "grad_norm": 0.910548985004425, "learning_rate": 0.0002631512891995621, "loss": 5.2615, "step": 266950 }, { "epoch": 1.609966112323778, "grad_norm": 0.6366989612579346, "learning_rate": 0.0002631379294345182, "loss": 5.4781, "step": 267000 }, { "epoch": 1.609966112323778, "eval_loss": 5.553127288818359, "eval_runtime": 39.0914, "eval_samples_per_second": 13.098, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.355060176352622, "step": 267000 }, { "epoch": 1.610267604105112, "grad_norm": 0.6342905163764954, "learning_rate": 0.0002631245675873322, "loss": 4.8889, "step": 267050 }, { "epoch": 1.6105690958864463, "grad_norm": 1.7637962102890015, "learning_rate": 0.00026311120365825, "loss": 5.0678, "step": 267100 }, { "epoch": 1.61087058766778, "grad_norm": 1.5574876070022583, "learning_rate": 0.0002630978376475175, "loss": 5.2646, "step": 267150 }, { "epoch": 1.6111720794491142, "grad_norm": 0.9155411720275879, "learning_rate": 0.00026308446955538065, "loss": 4.9692, "step": 267200 }, { "epoch": 1.6114735712304482, "grad_norm": 0.40666598081588745, "learning_rate": 0.00026307109938208554, "loss": 5.2149, "step": 267250 }, { "epoch": 1.6117750630117822, "grad_norm": 1.5947716236114502, "learning_rate": 0.0002630577271278782, "loss": 5.1389, "step": 267300 }, { "epoch": 1.6120765547931164, "grad_norm": 2.7709009647369385, "learning_rate": 0.00026304435279300475, "loss": 4.9559, "step": 267350 }, { "epoch": 1.6123780465744504, "grad_norm": 1.6373817920684814, "learning_rate": 0.00026303097637771135, "loss": 5.2509, "step": 267400 }, { "epoch": 1.6126795383557844, "grad_norm": 2.6618475914001465, "learning_rate": 0.0002630175978822441, "loss": 4.9571, "step": 267450 }, { "epoch": 1.6129810301371186, "grad_norm": 1.3009377717971802, "learning_rate": 0.0002630042173068492, "loss": 5.04, "step": 267500 }, { "epoch": 1.6132825219184523, "grad_norm": 1.770609974861145, "learning_rate": 0.00026299083465177294, "loss": 5.4194, "step": 267550 }, { "epoch": 1.6135840136997865, "grad_norm": 0.9888728857040405, "learning_rate": 0.00026297744991726166, "loss": 5.2155, "step": 267600 }, { "epoch": 1.6138855054811205, "grad_norm": 1.1340783834457397, "learning_rate": 0.0002629640631035617, "loss": 5.5959, "step": 267650 }, { "epoch": 1.6141869972624545, "grad_norm": 1.3277373313903809, "learning_rate": 0.00026295067421091927, "loss": 5.1794, "step": 267700 }, { "epoch": 1.6144884890437887, "grad_norm": 1.2490954399108887, "learning_rate": 0.0002629372832395809, "loss": 4.9493, "step": 267750 }, { "epoch": 1.6147899808251227, "grad_norm": 1.3579858541488647, "learning_rate": 0.000262923890189793, "loss": 5.0196, "step": 267800 }, { "epoch": 1.6150914726064567, "grad_norm": 2.1423180103302, "learning_rate": 0.0002629104950618021, "loss": 5.2673, "step": 267850 }, { "epoch": 1.615392964387791, "grad_norm": 1.4651334285736084, "learning_rate": 0.0002628970978558546, "loss": 4.7833, "step": 267900 }, { "epoch": 1.6156944561691249, "grad_norm": 0.7193354964256287, "learning_rate": 0.00026288369857219716, "loss": 5.0834, "step": 267950 }, { "epoch": 1.6159959479504589, "grad_norm": 0.8478320837020874, "learning_rate": 0.00026287029721107634, "loss": 5.1912, "step": 268000 }, { "epoch": 1.616297439731793, "grad_norm": 0.9454058408737183, "learning_rate": 0.00026285689377273877, "loss": 5.1291, "step": 268050 }, { "epoch": 1.6165989315131268, "grad_norm": 1.6312193870544434, "learning_rate": 0.0002628434882574311, "loss": 4.6502, "step": 268100 }, { "epoch": 1.616900423294461, "grad_norm": 1.65073823928833, "learning_rate": 0.0002628300806654001, "loss": 4.306, "step": 268150 }, { "epoch": 1.617201915075795, "grad_norm": 1.7829962968826294, "learning_rate": 0.0002628166709968925, "loss": 5.4502, "step": 268200 }, { "epoch": 1.617503406857129, "grad_norm": 1.7925853729248047, "learning_rate": 0.00026280325925215505, "loss": 5.1497, "step": 268250 }, { "epoch": 1.6178048986384632, "grad_norm": 1.96367347240448, "learning_rate": 0.0002627898454314346, "loss": 5.0479, "step": 268300 }, { "epoch": 1.6181063904197972, "grad_norm": 1.0954649448394775, "learning_rate": 0.000262776429534978, "loss": 5.5957, "step": 268350 }, { "epoch": 1.6184078822011312, "grad_norm": 1.3654422760009766, "learning_rate": 0.0002627630115630321, "loss": 5.2272, "step": 268400 }, { "epoch": 1.6187093739824654, "grad_norm": 1.223415732383728, "learning_rate": 0.00026274959151584396, "loss": 5.3419, "step": 268450 }, { "epoch": 1.6190108657637992, "grad_norm": 2.183393716812134, "learning_rate": 0.00026273616939366046, "loss": 4.8823, "step": 268500 }, { "epoch": 1.6193123575451334, "grad_norm": 1.2522863149642944, "learning_rate": 0.0002627227451967286, "loss": 4.9804, "step": 268550 }, { "epoch": 1.6196138493264673, "grad_norm": 1.3090565204620361, "learning_rate": 0.0002627093189252956, "loss": 5.3193, "step": 268600 }, { "epoch": 1.6199153411078013, "grad_norm": 1.6377346515655518, "learning_rate": 0.00026269589057960834, "loss": 4.6705, "step": 268650 }, { "epoch": 1.6202168328891355, "grad_norm": 2.4062435626983643, "learning_rate": 0.00026268246015991406, "loss": 5.3269, "step": 268700 }, { "epoch": 1.6205183246704695, "grad_norm": 1.4967395067214966, "learning_rate": 0.00026266902766645993, "loss": 5.3799, "step": 268750 }, { "epoch": 1.6208198164518035, "grad_norm": 1.6349194049835205, "learning_rate": 0.0002626555930994931, "loss": 5.4208, "step": 268800 }, { "epoch": 1.6211213082331377, "grad_norm": 1.1629507541656494, "learning_rate": 0.00026264215645926085, "loss": 5.4864, "step": 268850 }, { "epoch": 1.6214228000144715, "grad_norm": 1.102025032043457, "learning_rate": 0.00026262871774601046, "loss": 5.2149, "step": 268900 }, { "epoch": 1.6217242917958057, "grad_norm": 1.1515977382659912, "learning_rate": 0.0002626152769599893, "loss": 4.9293, "step": 268950 }, { "epoch": 1.6220257835771397, "grad_norm": 1.5637191534042358, "learning_rate": 0.0002626018341014446, "loss": 4.7416, "step": 269000 }, { "epoch": 1.6223272753584737, "grad_norm": 1.22495698928833, "learning_rate": 0.00026258838917062386, "loss": 5.0127, "step": 269050 }, { "epoch": 1.6226287671398079, "grad_norm": 1.5952725410461426, "learning_rate": 0.0002625749421677745, "loss": 5.0668, "step": 269100 }, { "epoch": 1.6229302589211418, "grad_norm": 1.622452735900879, "learning_rate": 0.00026256149309314403, "loss": 5.2646, "step": 269150 }, { "epoch": 1.6232317507024758, "grad_norm": 2.003434419631958, "learning_rate": 0.00026254804194697983, "loss": 4.9326, "step": 269200 }, { "epoch": 1.62353324248381, "grad_norm": 1.3021539449691772, "learning_rate": 0.00026253458872952957, "loss": 5.3619, "step": 269250 }, { "epoch": 1.6238347342651438, "grad_norm": 1.3803168535232544, "learning_rate": 0.0002625211334410408, "loss": 5.1322, "step": 269300 }, { "epoch": 1.624136226046478, "grad_norm": 1.7959426641464233, "learning_rate": 0.00026250767608176114, "loss": 4.9349, "step": 269350 }, { "epoch": 1.624437717827812, "grad_norm": 1.357149600982666, "learning_rate": 0.0002624942166519383, "loss": 4.6262, "step": 269400 }, { "epoch": 1.624739209609146, "grad_norm": 2.7843854427337646, "learning_rate": 0.00026248075515181987, "loss": 4.8633, "step": 269450 }, { "epoch": 1.6250407013904802, "grad_norm": 2.475437879562378, "learning_rate": 0.00026246729158165374, "loss": 5.1806, "step": 269500 }, { "epoch": 1.6253421931718142, "grad_norm": 1.2992674112319946, "learning_rate": 0.0002624538259416875, "loss": 5.1096, "step": 269550 }, { "epoch": 1.6256436849531481, "grad_norm": 1.3064650297164917, "learning_rate": 0.0002624403582321692, "loss": 4.2673, "step": 269600 }, { "epoch": 1.6259451767344824, "grad_norm": 1.1356524229049683, "learning_rate": 0.0002624268884533465, "loss": 5.1905, "step": 269650 }, { "epoch": 1.6262466685158161, "grad_norm": 1.1168631315231323, "learning_rate": 0.0002624134166054673, "loss": 5.014, "step": 269700 }, { "epoch": 1.6265481602971503, "grad_norm": 1.4250792264938354, "learning_rate": 0.0002623999426887797, "loss": 5.0118, "step": 269750 }, { "epoch": 1.6268496520784843, "grad_norm": 1.0657737255096436, "learning_rate": 0.00026238646670353154, "loss": 4.8875, "step": 269800 }, { "epoch": 1.6271511438598183, "grad_norm": 1.2241055965423584, "learning_rate": 0.0002623729886499708, "loss": 4.8449, "step": 269850 }, { "epoch": 1.6274526356411525, "grad_norm": 2.03121018409729, "learning_rate": 0.00026235950852834556, "loss": 5.381, "step": 269900 }, { "epoch": 1.6277541274224865, "grad_norm": 1.3756617307662964, "learning_rate": 0.000262346026338904, "loss": 5.2723, "step": 269950 }, { "epoch": 1.6280556192038205, "grad_norm": 0.32211053371429443, "learning_rate": 0.00026233254208189407, "loss": 5.0906, "step": 270000 }, { "epoch": 1.6280556192038205, "eval_loss": 5.54581356048584, "eval_runtime": 39.2266, "eval_samples_per_second": 13.052, "eval_steps_per_second": 6.526, "eval_tts_loss": 7.3170474228030695, "step": 270000 }, { "epoch": 1.6283571109851547, "grad_norm": 1.6198832988739014, "learning_rate": 0.00026231905575756405, "loss": 5.3898, "step": 270050 }, { "epoch": 1.6286586027664884, "grad_norm": 1.4621562957763672, "learning_rate": 0.00026230556736616203, "loss": 5.0759, "step": 270100 }, { "epoch": 1.6289600945478226, "grad_norm": 1.6185623407363892, "learning_rate": 0.00026229207690793636, "loss": 5.0315, "step": 270150 }, { "epoch": 1.6292615863291566, "grad_norm": 1.1055922508239746, "learning_rate": 0.00026227858438313526, "loss": 5.1768, "step": 270200 }, { "epoch": 1.6295630781104906, "grad_norm": 1.7805907726287842, "learning_rate": 0.000262265089792007, "loss": 5.0411, "step": 270250 }, { "epoch": 1.6298645698918248, "grad_norm": 1.8605239391326904, "learning_rate": 0.00026225159313480003, "loss": 4.9806, "step": 270300 }, { "epoch": 1.6301660616731588, "grad_norm": 1.3079200983047485, "learning_rate": 0.0002622380944117627, "loss": 4.8485, "step": 270350 }, { "epoch": 1.6304675534544928, "grad_norm": 1.2657041549682617, "learning_rate": 0.0002622245936231434, "loss": 4.9843, "step": 270400 }, { "epoch": 1.630769045235827, "grad_norm": 1.1300009489059448, "learning_rate": 0.0002622110907691906, "loss": 5.0562, "step": 270450 }, { "epoch": 1.6310705370171608, "grad_norm": 1.2762913703918457, "learning_rate": 0.00026219758585015277, "loss": 4.9004, "step": 270500 }, { "epoch": 1.631372028798495, "grad_norm": 1.99836003780365, "learning_rate": 0.0002621840788662785, "loss": 5.2991, "step": 270550 }, { "epoch": 1.631673520579829, "grad_norm": 0.7744551301002502, "learning_rate": 0.0002621705698178164, "loss": 5.1353, "step": 270600 }, { "epoch": 1.631975012361163, "grad_norm": 2.339663505554199, "learning_rate": 0.000262157058705015, "loss": 4.9791, "step": 270650 }, { "epoch": 1.6322765041424971, "grad_norm": 1.7025494575500488, "learning_rate": 0.000262143545528123, "loss": 5.0139, "step": 270700 }, { "epoch": 1.6325779959238311, "grad_norm": 1.2923370599746704, "learning_rate": 0.0002621300302873891, "loss": 5.1626, "step": 270750 }, { "epoch": 1.6328794877051651, "grad_norm": 1.37605619430542, "learning_rate": 0.00026211651298306196, "loss": 4.9238, "step": 270800 }, { "epoch": 1.6331809794864993, "grad_norm": 1.7355436086654663, "learning_rate": 0.00026210299361539044, "loss": 5.0719, "step": 270850 }, { "epoch": 1.633482471267833, "grad_norm": 1.1588364839553833, "learning_rate": 0.0002620894721846233, "loss": 5.001, "step": 270900 }, { "epoch": 1.6337839630491673, "grad_norm": 1.2212491035461426, "learning_rate": 0.00026207594869100934, "loss": 5.0267, "step": 270950 }, { "epoch": 1.6340854548305013, "grad_norm": 1.8845715522766113, "learning_rate": 0.0002620624231347975, "loss": 5.0167, "step": 271000 }, { "epoch": 1.6343869466118353, "grad_norm": 1.2853533029556274, "learning_rate": 0.0002620488955162367, "loss": 5.189, "step": 271050 }, { "epoch": 1.6346884383931695, "grad_norm": 1.7456482648849487, "learning_rate": 0.00026203536583557595, "loss": 5.2829, "step": 271100 }, { "epoch": 1.6349899301745034, "grad_norm": 2.553572654724121, "learning_rate": 0.00026202183409306407, "loss": 5.4304, "step": 271150 }, { "epoch": 1.6352914219558374, "grad_norm": 0.5853917598724365, "learning_rate": 0.00026200830028895027, "loss": 5.1656, "step": 271200 }, { "epoch": 1.6355929137371716, "grad_norm": 1.7125186920166016, "learning_rate": 0.0002619947644234835, "loss": 5.4788, "step": 271250 }, { "epoch": 1.6358944055185054, "grad_norm": 1.9920778274536133, "learning_rate": 0.0002619812264969129, "loss": 5.3033, "step": 271300 }, { "epoch": 1.6361958972998396, "grad_norm": 1.954554557800293, "learning_rate": 0.00026196768650948765, "loss": 5.102, "step": 271350 }, { "epoch": 1.6364973890811736, "grad_norm": 2.020752191543579, "learning_rate": 0.000261954144461457, "loss": 5.0812, "step": 271400 }, { "epoch": 1.6367988808625076, "grad_norm": 2.6488869190216064, "learning_rate": 0.00026194060035306994, "loss": 5.0632, "step": 271450 }, { "epoch": 1.6371003726438418, "grad_norm": 1.8308409452438354, "learning_rate": 0.000261927054184576, "loss": 5.1046, "step": 271500 }, { "epoch": 1.6374018644251758, "grad_norm": 1.2752960920333862, "learning_rate": 0.0002619135059562243, "loss": 4.9533, "step": 271550 }, { "epoch": 1.6377033562065098, "grad_norm": 0.45859235525131226, "learning_rate": 0.00026189995566826426, "loss": 5.2037, "step": 271600 }, { "epoch": 1.638004847987844, "grad_norm": 1.5288605690002441, "learning_rate": 0.0002618864033209452, "loss": 4.8471, "step": 271650 }, { "epoch": 1.638306339769178, "grad_norm": 1.1921534538269043, "learning_rate": 0.00026187284891451655, "loss": 4.8679, "step": 271700 }, { "epoch": 1.638607831550512, "grad_norm": 1.1672621965408325, "learning_rate": 0.0002618592924492278, "loss": 5.4032, "step": 271750 }, { "epoch": 1.6389093233318461, "grad_norm": 1.3699874877929688, "learning_rate": 0.0002618457339253284, "loss": 4.9923, "step": 271800 }, { "epoch": 1.63921081511318, "grad_norm": 1.4482542276382446, "learning_rate": 0.00026183217334306786, "loss": 5.2298, "step": 271850 }, { "epoch": 1.639512306894514, "grad_norm": 0.7239922285079956, "learning_rate": 0.0002618186107026958, "loss": 5.1929, "step": 271900 }, { "epoch": 1.639813798675848, "grad_norm": 1.8826806545257568, "learning_rate": 0.00026180504600446173, "loss": 4.8112, "step": 271950 }, { "epoch": 1.640115290457182, "grad_norm": 1.739383339881897, "learning_rate": 0.0002617914792486154, "loss": 5.2071, "step": 272000 }, { "epoch": 1.6404167822385163, "grad_norm": 1.7315938472747803, "learning_rate": 0.00026177791043540636, "loss": 5.3941, "step": 272050 }, { "epoch": 1.6407182740198503, "grad_norm": 1.5272570848464966, "learning_rate": 0.00026176433956508443, "loss": 5.2445, "step": 272100 }, { "epoch": 1.6410197658011842, "grad_norm": 1.9619585275650024, "learning_rate": 0.0002617507666378993, "loss": 5.1246, "step": 272150 }, { "epoch": 1.6413212575825185, "grad_norm": 1.4578447341918945, "learning_rate": 0.00026173719165410085, "loss": 4.9223, "step": 272200 }, { "epoch": 1.6416227493638522, "grad_norm": 1.653969645500183, "learning_rate": 0.0002617236146139388, "loss": 4.8777, "step": 272250 }, { "epoch": 1.6419242411451864, "grad_norm": 1.2773289680480957, "learning_rate": 0.0002617100355176631, "loss": 4.7637, "step": 272300 }, { "epoch": 1.6422257329265204, "grad_norm": 1.0954365730285645, "learning_rate": 0.00026169645436552357, "loss": 5.1012, "step": 272350 }, { "epoch": 1.6425272247078544, "grad_norm": 1.3608616590499878, "learning_rate": 0.0002616828711577702, "loss": 4.9673, "step": 272400 }, { "epoch": 1.6428287164891886, "grad_norm": 1.6543397903442383, "learning_rate": 0.00026166928589465297, "loss": 5.1752, "step": 272450 }, { "epoch": 1.6431302082705226, "grad_norm": 0.9894275069236755, "learning_rate": 0.00026165569857642193, "loss": 5.1839, "step": 272500 }, { "epoch": 1.6434317000518566, "grad_norm": 1.4724472761154175, "learning_rate": 0.00026164210920332707, "loss": 5.2778, "step": 272550 }, { "epoch": 1.6437331918331908, "grad_norm": 1.302442193031311, "learning_rate": 0.0002616285177756185, "loss": 5.0756, "step": 272600 }, { "epoch": 1.6440346836145245, "grad_norm": 1.6610968112945557, "learning_rate": 0.00026161492429354634, "loss": 5.2471, "step": 272650 }, { "epoch": 1.6443361753958587, "grad_norm": 1.3389320373535156, "learning_rate": 0.00026160132875736083, "loss": 5.0856, "step": 272700 }, { "epoch": 1.6446376671771927, "grad_norm": 2.0148026943206787, "learning_rate": 0.0002615877311673121, "loss": 4.9377, "step": 272750 }, { "epoch": 1.6449391589585267, "grad_norm": 1.066085934638977, "learning_rate": 0.00026157413152365046, "loss": 4.7848, "step": 272800 }, { "epoch": 1.645240650739861, "grad_norm": 0.9822547435760498, "learning_rate": 0.00026156052982662613, "loss": 5.2249, "step": 272850 }, { "epoch": 1.645542142521195, "grad_norm": 1.348382592201233, "learning_rate": 0.0002615469260764895, "loss": 5.0166, "step": 272900 }, { "epoch": 1.645843634302529, "grad_norm": 1.7830188274383545, "learning_rate": 0.0002615333202734908, "loss": 5.4176, "step": 272950 }, { "epoch": 1.646145126083863, "grad_norm": 1.7995027303695679, "learning_rate": 0.0002615197124178806, "loss": 4.9112, "step": 273000 }, { "epoch": 1.646145126083863, "eval_loss": 5.563337326049805, "eval_runtime": 39.0388, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.292001952463633, "step": 273000 }, { "epoch": 1.6464466178651969, "grad_norm": 0.8138687014579773, "learning_rate": 0.00026150610250990914, "loss": 5.1643, "step": 273050 }, { "epoch": 1.646748109646531, "grad_norm": 1.5239882469177246, "learning_rate": 0.00026149249054982705, "loss": 5.3473, "step": 273100 }, { "epoch": 1.647049601427865, "grad_norm": 1.6901788711547852, "learning_rate": 0.00026147887653788475, "loss": 5.1845, "step": 273150 }, { "epoch": 1.647351093209199, "grad_norm": 1.383639931678772, "learning_rate": 0.0002614652604743328, "loss": 4.9651, "step": 273200 }, { "epoch": 1.6476525849905332, "grad_norm": 2.133045196533203, "learning_rate": 0.00026145164235942184, "loss": 5.5264, "step": 273250 }, { "epoch": 1.6479540767718672, "grad_norm": 1.711511492729187, "learning_rate": 0.00026143802219340246, "loss": 5.2569, "step": 273300 }, { "epoch": 1.6482555685532012, "grad_norm": 2.007734775543213, "learning_rate": 0.00026142439997652527, "loss": 5.159, "step": 273350 }, { "epoch": 1.6485570603345354, "grad_norm": 1.6312079429626465, "learning_rate": 0.00026141077570904106, "loss": 4.9377, "step": 273400 }, { "epoch": 1.6488585521158692, "grad_norm": 1.3502122163772583, "learning_rate": 0.0002613971493912005, "loss": 5.2702, "step": 273450 }, { "epoch": 1.6491600438972034, "grad_norm": 1.097564458847046, "learning_rate": 0.00026138352102325434, "loss": 4.6426, "step": 273500 }, { "epoch": 1.6494615356785374, "grad_norm": 1.2825857400894165, "learning_rate": 0.0002613698906054535, "loss": 5.0931, "step": 273550 }, { "epoch": 1.6497630274598714, "grad_norm": 1.2328730821609497, "learning_rate": 0.00026135625813804867, "loss": 5.5239, "step": 273600 }, { "epoch": 1.6500645192412056, "grad_norm": 0.6439483761787415, "learning_rate": 0.00026134262362129086, "loss": 5.0276, "step": 273650 }, { "epoch": 1.6503660110225395, "grad_norm": 1.874759316444397, "learning_rate": 0.000261328987055431, "loss": 5.24, "step": 273700 }, { "epoch": 1.6506675028038735, "grad_norm": 2.3153510093688965, "learning_rate": 0.00026131534844071994, "loss": 4.8998, "step": 273750 }, { "epoch": 1.6509689945852077, "grad_norm": 1.8148342370986938, "learning_rate": 0.0002613017077774088, "loss": 4.784, "step": 273800 }, { "epoch": 1.6512704863665415, "grad_norm": 1.7760632038116455, "learning_rate": 0.00026128806506574856, "loss": 4.8599, "step": 273850 }, { "epoch": 1.6515719781478757, "grad_norm": 1.17805016040802, "learning_rate": 0.0002612744203059903, "loss": 5.3907, "step": 273900 }, { "epoch": 1.6518734699292097, "grad_norm": 1.758459448814392, "learning_rate": 0.00026126077349838506, "loss": 4.6432, "step": 273950 }, { "epoch": 1.6521749617105437, "grad_norm": 1.2426873445510864, "learning_rate": 0.00026124712464318416, "loss": 5.1885, "step": 274000 }, { "epoch": 1.6524764534918779, "grad_norm": 1.2137812376022339, "learning_rate": 0.0002612334737406386, "loss": 5.0815, "step": 274050 }, { "epoch": 1.6527779452732119, "grad_norm": 1.5434801578521729, "learning_rate": 0.00026121982079099976, "loss": 5.1793, "step": 274100 }, { "epoch": 1.6530794370545459, "grad_norm": 1.2797999382019043, "learning_rate": 0.0002612061657945188, "loss": 5.3694, "step": 274150 }, { "epoch": 1.65338092883588, "grad_norm": 2.1273369789123535, "learning_rate": 0.0002611925087514471, "loss": 5.0914, "step": 274200 }, { "epoch": 1.6536824206172138, "grad_norm": 1.0759724378585815, "learning_rate": 0.0002611788496620359, "loss": 5.0353, "step": 274250 }, { "epoch": 1.653983912398548, "grad_norm": 1.346104621887207, "learning_rate": 0.00026116518852653666, "loss": 4.9299, "step": 274300 }, { "epoch": 1.654285404179882, "grad_norm": 0.9528908133506775, "learning_rate": 0.0002611515253452008, "loss": 5.4584, "step": 274350 }, { "epoch": 1.654586895961216, "grad_norm": 1.5700863599777222, "learning_rate": 0.0002611378601182797, "loss": 4.9547, "step": 274400 }, { "epoch": 1.6548883877425502, "grad_norm": 1.3066293001174927, "learning_rate": 0.0002611241928460249, "loss": 5.311, "step": 274450 }, { "epoch": 1.6551898795238842, "grad_norm": 1.567366123199463, "learning_rate": 0.0002611105235286879, "loss": 5.1085, "step": 274500 }, { "epoch": 1.6554913713052182, "grad_norm": 1.7031062841415405, "learning_rate": 0.0002610968521665203, "loss": 5.089, "step": 274550 }, { "epoch": 1.6557928630865524, "grad_norm": 1.5892168283462524, "learning_rate": 0.00026108317875977365, "loss": 5.0747, "step": 274600 }, { "epoch": 1.6560943548678861, "grad_norm": 1.4057714939117432, "learning_rate": 0.0002610695033086996, "loss": 4.7596, "step": 274650 }, { "epoch": 1.6563958466492203, "grad_norm": 1.3500401973724365, "learning_rate": 0.0002610558258135499, "loss": 5.3464, "step": 274700 }, { "epoch": 1.6566973384305543, "grad_norm": 0.549066960811615, "learning_rate": 0.0002610421462745762, "loss": 5.0735, "step": 274750 }, { "epoch": 1.6569988302118883, "grad_norm": 2.1510045528411865, "learning_rate": 0.00026102846469203025, "loss": 5.3495, "step": 274800 }, { "epoch": 1.6573003219932225, "grad_norm": 1.0732851028442383, "learning_rate": 0.00026101478106616385, "loss": 5.1355, "step": 274850 }, { "epoch": 1.6576018137745565, "grad_norm": 1.6088740825653076, "learning_rate": 0.00026100109539722887, "loss": 5.4185, "step": 274900 }, { "epoch": 1.6579033055558905, "grad_norm": 1.79628324508667, "learning_rate": 0.0002609874076854771, "loss": 4.8962, "step": 274950 }, { "epoch": 1.6582047973372247, "grad_norm": 1.5015602111816406, "learning_rate": 0.0002609737179311605, "loss": 5.404, "step": 275000 }, { "epoch": 1.6585062891185585, "grad_norm": 0.5707817077636719, "learning_rate": 0.000260960026134531, "loss": 5.4071, "step": 275050 }, { "epoch": 1.6588077808998927, "grad_norm": 1.010384440422058, "learning_rate": 0.0002609463322958405, "loss": 5.0478, "step": 275100 }, { "epoch": 1.6591092726812267, "grad_norm": 1.3842849731445312, "learning_rate": 0.0002609326364153411, "loss": 5.1496, "step": 275150 }, { "epoch": 1.6594107644625606, "grad_norm": 1.2737655639648438, "learning_rate": 0.0002609189384932849, "loss": 4.8616, "step": 275200 }, { "epoch": 1.6597122562438948, "grad_norm": 1.153509259223938, "learning_rate": 0.00026090523852992386, "loss": 5.1697, "step": 275250 }, { "epoch": 1.6600137480252288, "grad_norm": 0.41424861550331116, "learning_rate": 0.00026089153652551024, "loss": 5.1949, "step": 275300 }, { "epoch": 1.6603152398065628, "grad_norm": 1.6537811756134033, "learning_rate": 0.0002608778324802961, "loss": 4.7884, "step": 275350 }, { "epoch": 1.660616731587897, "grad_norm": 1.0961982011795044, "learning_rate": 0.0002608641263945337, "loss": 4.8677, "step": 275400 }, { "epoch": 1.660918223369231, "grad_norm": 1.651275396347046, "learning_rate": 0.0002608504182684752, "loss": 4.9193, "step": 275450 }, { "epoch": 1.661219715150565, "grad_norm": 0.8220316767692566, "learning_rate": 0.000260836708102373, "loss": 4.9311, "step": 275500 }, { "epoch": 1.6615212069318992, "grad_norm": 1.2259647846221924, "learning_rate": 0.0002608229958964794, "loss": 5.6961, "step": 275550 }, { "epoch": 1.661822698713233, "grad_norm": 1.8761454820632935, "learning_rate": 0.00026080928165104664, "loss": 4.9536, "step": 275600 }, { "epoch": 1.6621241904945672, "grad_norm": 1.0111699104309082, "learning_rate": 0.00026079556536632724, "loss": 5.1536, "step": 275650 }, { "epoch": 1.6624256822759012, "grad_norm": 1.1334127187728882, "learning_rate": 0.00026078184704257353, "loss": 4.9494, "step": 275700 }, { "epoch": 1.6627271740572351, "grad_norm": 1.5974477529525757, "learning_rate": 0.000260768126680038, "loss": 5.0671, "step": 275750 }, { "epoch": 1.6630286658385693, "grad_norm": 1.0309475660324097, "learning_rate": 0.0002607544042789733, "loss": 4.9209, "step": 275800 }, { "epoch": 1.6633301576199033, "grad_norm": 1.3072034120559692, "learning_rate": 0.0002607406798396317, "loss": 4.8652, "step": 275850 }, { "epoch": 1.6636316494012373, "grad_norm": 1.1907099485397339, "learning_rate": 0.000260726953362266, "loss": 4.8184, "step": 275900 }, { "epoch": 1.6639331411825715, "grad_norm": 1.4980902671813965, "learning_rate": 0.0002607132248471287, "loss": 4.9783, "step": 275950 }, { "epoch": 1.6642346329639053, "grad_norm": 1.5130270719528198, "learning_rate": 0.0002606994942944725, "loss": 4.6352, "step": 276000 }, { "epoch": 1.6642346329639053, "eval_loss": 5.541139125823975, "eval_runtime": 39.0882, "eval_samples_per_second": 13.099, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.283028633073029, "step": 276000 }, { "epoch": 1.6645361247452395, "grad_norm": 1.1898685693740845, "learning_rate": 0.0002606857617045501, "loss": 5.3517, "step": 276050 }, { "epoch": 1.6648376165265735, "grad_norm": 1.267276644706726, "learning_rate": 0.0002606720270776142, "loss": 5.373, "step": 276100 }, { "epoch": 1.6651391083079075, "grad_norm": 1.9760866165161133, "learning_rate": 0.0002606582904139176, "loss": 5.3657, "step": 276150 }, { "epoch": 1.6654406000892417, "grad_norm": 1.5992763042449951, "learning_rate": 0.00026064455171371305, "loss": 5.3435, "step": 276200 }, { "epoch": 1.6657420918705756, "grad_norm": 1.247166395187378, "learning_rate": 0.0002606308109772535, "loss": 5.0603, "step": 276250 }, { "epoch": 1.6660435836519096, "grad_norm": 1.6010019779205322, "learning_rate": 0.00026061706820479163, "loss": 5.1647, "step": 276300 }, { "epoch": 1.6663450754332438, "grad_norm": 1.9927425384521484, "learning_rate": 0.0002606033233965806, "loss": 4.896, "step": 276350 }, { "epoch": 1.6666465672145776, "grad_norm": 0.7813222408294678, "learning_rate": 0.00026058957655287313, "loss": 4.7792, "step": 276400 }, { "epoch": 1.6669480589959118, "grad_norm": 1.6124346256256104, "learning_rate": 0.00026057582767392243, "loss": 4.913, "step": 276450 }, { "epoch": 1.6672495507772458, "grad_norm": 1.7731976509094238, "learning_rate": 0.0002605620767599814, "loss": 5.4909, "step": 276500 }, { "epoch": 1.6675510425585798, "grad_norm": 0.6638645529747009, "learning_rate": 0.00026054832381130306, "loss": 5.2443, "step": 276550 }, { "epoch": 1.667852534339914, "grad_norm": 1.2170408964157104, "learning_rate": 0.0002605345688281406, "loss": 4.9507, "step": 276600 }, { "epoch": 1.668154026121248, "grad_norm": 1.8862988948822021, "learning_rate": 0.00026052081181074713, "loss": 5.2735, "step": 276650 }, { "epoch": 1.668455517902582, "grad_norm": 1.1515498161315918, "learning_rate": 0.0002605070527593759, "loss": 5.7528, "step": 276700 }, { "epoch": 1.6687570096839162, "grad_norm": 2.2199270725250244, "learning_rate": 0.0002604932916742801, "loss": 5.2479, "step": 276750 }, { "epoch": 1.66905850146525, "grad_norm": 1.7999428510665894, "learning_rate": 0.00026047952855571285, "loss": 5.0211, "step": 276800 }, { "epoch": 1.6693599932465841, "grad_norm": 2.274672269821167, "learning_rate": 0.00026046576340392756, "loss": 5.318, "step": 276850 }, { "epoch": 1.6696614850279181, "grad_norm": 1.273626446723938, "learning_rate": 0.00026045199621917755, "loss": 5.3007, "step": 276900 }, { "epoch": 1.669962976809252, "grad_norm": 1.3722866773605347, "learning_rate": 0.00026043822700171624, "loss": 5.2277, "step": 276950 }, { "epoch": 1.6702644685905863, "grad_norm": 0.8799328804016113, "learning_rate": 0.00026042445575179686, "loss": 4.5086, "step": 277000 }, { "epoch": 1.6705659603719203, "grad_norm": 1.4077339172363281, "learning_rate": 0.00026041068246967304, "loss": 5.3161, "step": 277050 }, { "epoch": 1.6708674521532543, "grad_norm": 1.5922958850860596, "learning_rate": 0.00026039690715559807, "loss": 5.1177, "step": 277100 }, { "epoch": 1.6711689439345885, "grad_norm": 1.2228963375091553, "learning_rate": 0.00026038312980982564, "loss": 5.1308, "step": 277150 }, { "epoch": 1.6714704357159222, "grad_norm": 2.005831480026245, "learning_rate": 0.0002603693504326093, "loss": 4.9358, "step": 277200 }, { "epoch": 1.6717719274972564, "grad_norm": 1.4184941053390503, "learning_rate": 0.00026035556902420247, "loss": 5.1072, "step": 277250 }, { "epoch": 1.6720734192785904, "grad_norm": 1.4292465448379517, "learning_rate": 0.00026034178558485886, "loss": 4.9752, "step": 277300 }, { "epoch": 1.6723749110599244, "grad_norm": 1.848199486732483, "learning_rate": 0.00026032800011483223, "loss": 5.2376, "step": 277350 }, { "epoch": 1.6726764028412586, "grad_norm": 1.2854666709899902, "learning_rate": 0.0002603142126143761, "loss": 4.8873, "step": 277400 }, { "epoch": 1.6729778946225926, "grad_norm": 2.3564341068267822, "learning_rate": 0.00026030042308374446, "loss": 5.2808, "step": 277450 }, { "epoch": 1.6732793864039266, "grad_norm": 1.5158227682113647, "learning_rate": 0.0002602866315231908, "loss": 5.4752, "step": 277500 }, { "epoch": 1.6735808781852608, "grad_norm": 1.8001657724380493, "learning_rate": 0.0002602728379329692, "loss": 5.238, "step": 277550 }, { "epoch": 1.6738823699665946, "grad_norm": 1.8016825914382935, "learning_rate": 0.00026025904231333333, "loss": 5.0988, "step": 277600 }, { "epoch": 1.6741838617479288, "grad_norm": 1.3177006244659424, "learning_rate": 0.0002602452446645371, "loss": 5.3795, "step": 277650 }, { "epoch": 1.6744853535292628, "grad_norm": 1.3529717922210693, "learning_rate": 0.0002602314449868345, "loss": 5.0689, "step": 277700 }, { "epoch": 1.6747868453105967, "grad_norm": 1.7591527700424194, "learning_rate": 0.0002602176432804795, "loss": 5.1619, "step": 277750 }, { "epoch": 1.675088337091931, "grad_norm": 1.2397090196609497, "learning_rate": 0.000260203839545726, "loss": 5.0131, "step": 277800 }, { "epoch": 1.675389828873265, "grad_norm": 2.1298108100891113, "learning_rate": 0.0002601900337828281, "loss": 5.285, "step": 277850 }, { "epoch": 1.675691320654599, "grad_norm": 1.4115670919418335, "learning_rate": 0.00026017622599203995, "loss": 4.8105, "step": 277900 }, { "epoch": 1.6759928124359331, "grad_norm": 1.5686997175216675, "learning_rate": 0.0002601624161736155, "loss": 4.9061, "step": 277950 }, { "epoch": 1.6762943042172669, "grad_norm": 0.6961222290992737, "learning_rate": 0.00026014860432780907, "loss": 5.1568, "step": 278000 }, { "epoch": 1.676595795998601, "grad_norm": 2.338724136352539, "learning_rate": 0.0002601347904548747, "loss": 4.9363, "step": 278050 }, { "epoch": 1.676897287779935, "grad_norm": 2.027804374694824, "learning_rate": 0.00026012097455506677, "loss": 4.9919, "step": 278100 }, { "epoch": 1.677198779561269, "grad_norm": 1.0287662744522095, "learning_rate": 0.0002601071566286394, "loss": 4.4383, "step": 278150 }, { "epoch": 1.6775002713426033, "grad_norm": 0.936530590057373, "learning_rate": 0.00026009333667584696, "loss": 5.1151, "step": 278200 }, { "epoch": 1.6778017631239373, "grad_norm": 1.0611987113952637, "learning_rate": 0.0002600795146969437, "loss": 5.1246, "step": 278250 }, { "epoch": 1.6781032549052712, "grad_norm": 0.762914776802063, "learning_rate": 0.00026006569069218414, "loss": 5.3956, "step": 278300 }, { "epoch": 1.6784047466866054, "grad_norm": 1.7235112190246582, "learning_rate": 0.0002600518646618226, "loss": 5.4477, "step": 278350 }, { "epoch": 1.6787062384679392, "grad_norm": 2.586287021636963, "learning_rate": 0.0002600380366061135, "loss": 5.428, "step": 278400 }, { "epoch": 1.6790077302492734, "grad_norm": 1.2409497499465942, "learning_rate": 0.0002600242065253115, "loss": 5.1325, "step": 278450 }, { "epoch": 1.6793092220306074, "grad_norm": 2.324235200881958, "learning_rate": 0.00026001037441967086, "loss": 5.1722, "step": 278500 }, { "epoch": 1.6796107138119414, "grad_norm": 3.093552589416504, "learning_rate": 0.0002599965402894463, "loss": 4.8485, "step": 278550 }, { "epoch": 1.6799122055932756, "grad_norm": 1.198153018951416, "learning_rate": 0.0002599827041348924, "loss": 5.0579, "step": 278600 }, { "epoch": 1.6802136973746096, "grad_norm": 1.3477503061294556, "learning_rate": 0.0002599688659562638, "loss": 5.2351, "step": 278650 }, { "epoch": 1.6805151891559436, "grad_norm": 1.4503929615020752, "learning_rate": 0.0002599550257538151, "loss": 4.8935, "step": 278700 }, { "epoch": 1.6808166809372778, "grad_norm": 1.3940457105636597, "learning_rate": 0.00025994118352780116, "loss": 4.8329, "step": 278750 }, { "epoch": 1.6811181727186115, "grad_norm": 1.603865623474121, "learning_rate": 0.00025992733927847654, "loss": 4.9933, "step": 278800 }, { "epoch": 1.6814196644999457, "grad_norm": 1.0420031547546387, "learning_rate": 0.00025991349300609614, "loss": 4.8617, "step": 278850 }, { "epoch": 1.68172115628128, "grad_norm": 1.35036039352417, "learning_rate": 0.00025989964471091477, "loss": 4.8363, "step": 278900 }, { "epoch": 1.6820226480626137, "grad_norm": 2.268207550048828, "learning_rate": 0.0002598857943931873, "loss": 5.0205, "step": 278950 }, { "epoch": 1.682324139843948, "grad_norm": 0.6544342637062073, "learning_rate": 0.0002598719420531686, "loss": 4.6557, "step": 279000 }, { "epoch": 1.682324139843948, "eval_loss": 5.556447505950928, "eval_runtime": 39.1524, "eval_samples_per_second": 13.077, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.239839862140906, "step": 279000 }, { "epoch": 1.682625631625282, "grad_norm": 1.0836925506591797, "learning_rate": 0.00025985808769111356, "loss": 5.0672, "step": 279050 }, { "epoch": 1.6829271234066159, "grad_norm": 1.0583115816116333, "learning_rate": 0.0002598442313072772, "loss": 5.1968, "step": 279100 }, { "epoch": 1.68322861518795, "grad_norm": 1.8575009107589722, "learning_rate": 0.00025983037290191454, "loss": 5.1188, "step": 279150 }, { "epoch": 1.683530106969284, "grad_norm": 1.7813873291015625, "learning_rate": 0.0002598165124752806, "loss": 4.9924, "step": 279200 }, { "epoch": 1.683831598750618, "grad_norm": 1.414002537727356, "learning_rate": 0.0002598026500276304, "loss": 4.8029, "step": 279250 }, { "epoch": 1.6841330905319523, "grad_norm": 2.382591485977173, "learning_rate": 0.0002597887855592192, "loss": 5.0149, "step": 279300 }, { "epoch": 1.684434582313286, "grad_norm": 1.7565715312957764, "learning_rate": 0.000259774919070302, "loss": 5.1296, "step": 279350 }, { "epoch": 1.6847360740946202, "grad_norm": 1.3626254796981812, "learning_rate": 0.00025976105056113417, "loss": 4.8508, "step": 279400 }, { "epoch": 1.6850375658759542, "grad_norm": 1.5549862384796143, "learning_rate": 0.0002597471800319708, "loss": 5.0147, "step": 279450 }, { "epoch": 1.6853390576572882, "grad_norm": 1.4160821437835693, "learning_rate": 0.0002597333074830672, "loss": 5.4985, "step": 279500 }, { "epoch": 1.6856405494386224, "grad_norm": 2.1349964141845703, "learning_rate": 0.00025971943291467866, "loss": 5.5509, "step": 279550 }, { "epoch": 1.6859420412199564, "grad_norm": 1.8962960243225098, "learning_rate": 0.0002597055563270606, "loss": 4.9104, "step": 279600 }, { "epoch": 1.6862435330012904, "grad_norm": 0.4542712867259979, "learning_rate": 0.0002596916777204682, "loss": 4.9879, "step": 279650 }, { "epoch": 1.6865450247826246, "grad_norm": 0.8043432831764221, "learning_rate": 0.0002596777970951572, "loss": 5.6353, "step": 279700 }, { "epoch": 1.6868465165639583, "grad_norm": 2.1892666816711426, "learning_rate": 0.0002596639144513827, "loss": 5.1902, "step": 279750 }, { "epoch": 1.6871480083452925, "grad_norm": 1.6840274333953857, "learning_rate": 0.0002596500297894004, "loss": 5.6232, "step": 279800 }, { "epoch": 1.6874495001266265, "grad_norm": 2.19391131401062, "learning_rate": 0.0002596361431094658, "loss": 5.0082, "step": 279850 }, { "epoch": 1.6877509919079605, "grad_norm": 1.5058139562606812, "learning_rate": 0.0002596222544118344, "loss": 5.3206, "step": 279900 }, { "epoch": 1.6880524836892947, "grad_norm": 1.212516188621521, "learning_rate": 0.00025960836369676193, "loss": 5.2632, "step": 279950 }, { "epoch": 1.6883539754706287, "grad_norm": 1.3770065307617188, "learning_rate": 0.00025959447096450387, "loss": 5.0228, "step": 280000 }, { "epoch": 1.6886554672519627, "grad_norm": 1.8514422178268433, "learning_rate": 0.00025958057621531605, "loss": 5.2408, "step": 280050 }, { "epoch": 1.688956959033297, "grad_norm": 1.392763614654541, "learning_rate": 0.00025956667944945403, "loss": 5.2427, "step": 280100 }, { "epoch": 1.6892584508146307, "grad_norm": 1.229843258857727, "learning_rate": 0.0002595527806671736, "loss": 5.1481, "step": 280150 }, { "epoch": 1.6895599425959649, "grad_norm": 1.7800136804580688, "learning_rate": 0.0002595388798687307, "loss": 5.1202, "step": 280200 }, { "epoch": 1.6898614343772989, "grad_norm": 1.6125874519348145, "learning_rate": 0.00025952497705438094, "loss": 4.5155, "step": 280250 }, { "epoch": 1.6901629261586328, "grad_norm": 0.593605637550354, "learning_rate": 0.0002595110722243803, "loss": 4.4512, "step": 280300 }, { "epoch": 1.690464417939967, "grad_norm": 1.3783130645751953, "learning_rate": 0.0002594971653789847, "loss": 5.2134, "step": 280350 }, { "epoch": 1.690765909721301, "grad_norm": 1.4204673767089844, "learning_rate": 0.00025948325651845, "loss": 5.4769, "step": 280400 }, { "epoch": 1.691067401502635, "grad_norm": 1.77642822265625, "learning_rate": 0.0002594693456430322, "loss": 5.4076, "step": 280450 }, { "epoch": 1.6913688932839692, "grad_norm": 1.2568401098251343, "learning_rate": 0.0002594554327529873, "loss": 5.395, "step": 280500 }, { "epoch": 1.691670385065303, "grad_norm": 1.683128833770752, "learning_rate": 0.00025944151784857144, "loss": 5.1092, "step": 280550 }, { "epoch": 1.6919718768466372, "grad_norm": 1.8050639629364014, "learning_rate": 0.00025942760093004055, "loss": 5.2843, "step": 280600 }, { "epoch": 1.6922733686279712, "grad_norm": 2.0161526203155518, "learning_rate": 0.00025941368199765085, "loss": 4.7013, "step": 280650 }, { "epoch": 1.6925748604093052, "grad_norm": 1.8581879138946533, "learning_rate": 0.0002593997610516585, "loss": 4.8368, "step": 280700 }, { "epoch": 1.6928763521906394, "grad_norm": 2.182685613632202, "learning_rate": 0.0002593858380923196, "loss": 5.0417, "step": 280750 }, { "epoch": 1.6931778439719734, "grad_norm": 1.2200309038162231, "learning_rate": 0.0002593719131198905, "loss": 5.3451, "step": 280800 }, { "epoch": 1.6934793357533073, "grad_norm": 0.7511424422264099, "learning_rate": 0.0002593579861346274, "loss": 5.4282, "step": 280850 }, { "epoch": 1.6937808275346415, "grad_norm": 2.066769599914551, "learning_rate": 0.0002593440571367867, "loss": 5.1823, "step": 280900 }, { "epoch": 1.6940823193159753, "grad_norm": 2.3349602222442627, "learning_rate": 0.00025933012612662455, "loss": 5.2192, "step": 280950 }, { "epoch": 1.6943838110973095, "grad_norm": 1.4083194732666016, "learning_rate": 0.0002593161931043975, "loss": 5.1092, "step": 281000 }, { "epoch": 1.6946853028786435, "grad_norm": 1.5970038175582886, "learning_rate": 0.00025930225807036195, "loss": 5.0187, "step": 281050 }, { "epoch": 1.6949867946599775, "grad_norm": 1.5500974655151367, "learning_rate": 0.0002592883210247743, "loss": 5.1956, "step": 281100 }, { "epoch": 1.6952882864413117, "grad_norm": 1.1994131803512573, "learning_rate": 0.00025927438196789106, "loss": 5.0653, "step": 281150 }, { "epoch": 1.6955897782226457, "grad_norm": 1.3608638048171997, "learning_rate": 0.00025926044089996875, "loss": 4.7567, "step": 281200 }, { "epoch": 1.6958912700039797, "grad_norm": 1.8240000009536743, "learning_rate": 0.000259246497821264, "loss": 5.5286, "step": 281250 }, { "epoch": 1.6961927617853139, "grad_norm": 1.0570324659347534, "learning_rate": 0.0002592325527320333, "loss": 5.151, "step": 281300 }, { "epoch": 1.6964942535666476, "grad_norm": 1.3613985776901245, "learning_rate": 0.0002592186056325334, "loss": 4.9622, "step": 281350 }, { "epoch": 1.6967957453479818, "grad_norm": 1.3215086460113525, "learning_rate": 0.0002592046565230209, "loss": 4.9234, "step": 281400 }, { "epoch": 1.6970972371293158, "grad_norm": 1.3361918926239014, "learning_rate": 0.00025919070540375257, "loss": 5.5025, "step": 281450 }, { "epoch": 1.6973987289106498, "grad_norm": 1.0575685501098633, "learning_rate": 0.0002591767522749851, "loss": 5.165, "step": 281500 }, { "epoch": 1.697700220691984, "grad_norm": 1.2545268535614014, "learning_rate": 0.0002591627971369753, "loss": 4.9917, "step": 281550 }, { "epoch": 1.698001712473318, "grad_norm": 1.3622411489486694, "learning_rate": 0.00025914883998998003, "loss": 4.942, "step": 281600 }, { "epoch": 1.698303204254652, "grad_norm": 1.8767982721328735, "learning_rate": 0.0002591348808342561, "loss": 5.1018, "step": 281650 }, { "epoch": 1.6986046960359862, "grad_norm": 1.6576100587844849, "learning_rate": 0.0002591209196700605, "loss": 5.0353, "step": 281700 }, { "epoch": 1.69890618781732, "grad_norm": 1.518423318862915, "learning_rate": 0.00025910695649765, "loss": 4.7096, "step": 281750 }, { "epoch": 1.6992076795986542, "grad_norm": 1.7906404733657837, "learning_rate": 0.00025909299131728167, "loss": 5.2639, "step": 281800 }, { "epoch": 1.6995091713799881, "grad_norm": 1.1243200302124023, "learning_rate": 0.0002590790241292126, "loss": 5.3434, "step": 281850 }, { "epoch": 1.6998106631613221, "grad_norm": 1.0898723602294922, "learning_rate": 0.00025906505493369973, "loss": 5.1191, "step": 281900 }, { "epoch": 1.7001121549426563, "grad_norm": 2.5134165287017822, "learning_rate": 0.0002590510837310001, "loss": 5.0258, "step": 281950 }, { "epoch": 1.7004136467239903, "grad_norm": 1.6095051765441895, "learning_rate": 0.00025903711052137097, "loss": 5.1569, "step": 282000 }, { "epoch": 1.7004136467239903, "eval_loss": 5.545632362365723, "eval_runtime": 38.9203, "eval_samples_per_second": 13.155, "eval_steps_per_second": 6.578, "eval_tts_loss": 7.341371939863952, "step": 282000 }, { "epoch": 1.7007151385053243, "grad_norm": 1.5583101511001587, "learning_rate": 0.0002590231353050694, "loss": 4.8434, "step": 282050 }, { "epoch": 1.7010166302866585, "grad_norm": 1.2926301956176758, "learning_rate": 0.00025900915808235264, "loss": 4.8729, "step": 282100 }, { "epoch": 1.7013181220679923, "grad_norm": 1.415293574333191, "learning_rate": 0.0002589951788534778, "loss": 4.8569, "step": 282150 }, { "epoch": 1.7016196138493265, "grad_norm": 1.603332757949829, "learning_rate": 0.00025898119761870236, "loss": 5.3655, "step": 282200 }, { "epoch": 1.7019211056306605, "grad_norm": 1.1839059591293335, "learning_rate": 0.0002589672143782834, "loss": 5.166, "step": 282250 }, { "epoch": 1.7022225974119944, "grad_norm": 1.4859639406204224, "learning_rate": 0.0002589532291324784, "loss": 5.0585, "step": 282300 }, { "epoch": 1.7025240891933286, "grad_norm": 1.8597729206085205, "learning_rate": 0.00025893924188154476, "loss": 5.1124, "step": 282350 }, { "epoch": 1.7028255809746626, "grad_norm": 0.9347102642059326, "learning_rate": 0.0002589252526257398, "loss": 5.0096, "step": 282400 }, { "epoch": 1.7031270727559966, "grad_norm": 2.358609199523926, "learning_rate": 0.00025891126136532095, "loss": 5.3228, "step": 282450 }, { "epoch": 1.7034285645373308, "grad_norm": 1.1809818744659424, "learning_rate": 0.00025889726810054584, "loss": 5.0224, "step": 282500 }, { "epoch": 1.7037300563186646, "grad_norm": 1.9314231872558594, "learning_rate": 0.00025888327283167187, "loss": 5.1107, "step": 282550 }, { "epoch": 1.7040315480999988, "grad_norm": 0.7937692999839783, "learning_rate": 0.0002588692755589567, "loss": 5.0071, "step": 282600 }, { "epoch": 1.704333039881333, "grad_norm": 1.616914987564087, "learning_rate": 0.00025885527628265783, "loss": 5.0255, "step": 282650 }, { "epoch": 1.7046345316626668, "grad_norm": 1.652809739112854, "learning_rate": 0.00025884127500303293, "loss": 4.8745, "step": 282700 }, { "epoch": 1.704936023444001, "grad_norm": 1.7238608598709106, "learning_rate": 0.00025882727172033974, "loss": 4.9274, "step": 282750 }, { "epoch": 1.705237515225335, "grad_norm": 1.2413177490234375, "learning_rate": 0.0002588132664348359, "loss": 5.0799, "step": 282800 }, { "epoch": 1.705539007006669, "grad_norm": 1.1909294128417969, "learning_rate": 0.00025879925914677916, "loss": 4.9754, "step": 282850 }, { "epoch": 1.7058404987880031, "grad_norm": 1.543746829032898, "learning_rate": 0.0002587852498564273, "loss": 5.0383, "step": 282900 }, { "epoch": 1.7061419905693371, "grad_norm": 1.4944171905517578, "learning_rate": 0.0002587712385640382, "loss": 5.1169, "step": 282950 }, { "epoch": 1.7064434823506711, "grad_norm": 1.505275011062622, "learning_rate": 0.0002587572252698697, "loss": 5.0329, "step": 283000 }, { "epoch": 1.7067449741320053, "grad_norm": 2.194803476333618, "learning_rate": 0.00025874320997417956, "loss": 5.0489, "step": 283050 }, { "epoch": 1.707046465913339, "grad_norm": 1.7933683395385742, "learning_rate": 0.000258729192677226, "loss": 5.3651, "step": 283100 }, { "epoch": 1.7073479576946733, "grad_norm": 1.5374006032943726, "learning_rate": 0.0002587151733792667, "loss": 4.7873, "step": 283150 }, { "epoch": 1.7076494494760073, "grad_norm": 1.2134385108947754, "learning_rate": 0.00025870115208055975, "loss": 4.8726, "step": 283200 }, { "epoch": 1.7079509412573413, "grad_norm": 1.2430787086486816, "learning_rate": 0.0002586871287813633, "loss": 5.0843, "step": 283250 }, { "epoch": 1.7082524330386755, "grad_norm": 1.2976603507995605, "learning_rate": 0.0002586731034819353, "loss": 5.2423, "step": 283300 }, { "epoch": 1.7085539248200095, "grad_norm": 2.0123543739318848, "learning_rate": 0.0002586590761825339, "loss": 5.2927, "step": 283350 }, { "epoch": 1.7088554166013434, "grad_norm": 1.8337883949279785, "learning_rate": 0.00025864504688341734, "loss": 4.9638, "step": 283400 }, { "epoch": 1.7091569083826776, "grad_norm": 1.4949687719345093, "learning_rate": 0.0002586310155848437, "loss": 4.9735, "step": 283450 }, { "epoch": 1.7094584001640114, "grad_norm": 1.2705386877059937, "learning_rate": 0.00025861698228707127, "loss": 5.254, "step": 283500 }, { "epoch": 1.7097598919453456, "grad_norm": 1.509268045425415, "learning_rate": 0.00025860294699035823, "loss": 5.354, "step": 283550 }, { "epoch": 1.7100613837266796, "grad_norm": 2.2583091259002686, "learning_rate": 0.00025858890969496294, "loss": 5.3007, "step": 283600 }, { "epoch": 1.7103628755080136, "grad_norm": 1.4537214040756226, "learning_rate": 0.0002585748704011438, "loss": 5.4727, "step": 283650 }, { "epoch": 1.7106643672893478, "grad_norm": 1.6156331300735474, "learning_rate": 0.00025856082910915906, "loss": 4.5015, "step": 283700 }, { "epoch": 1.7109658590706818, "grad_norm": 0.8568907976150513, "learning_rate": 0.0002585467858192672, "loss": 5.4995, "step": 283750 }, { "epoch": 1.7112673508520158, "grad_norm": 1.3125072717666626, "learning_rate": 0.0002585327405317266, "loss": 5.3505, "step": 283800 }, { "epoch": 1.71156884263335, "grad_norm": 1.5893810987472534, "learning_rate": 0.0002585186932467959, "loss": 5.3581, "step": 283850 }, { "epoch": 1.7118703344146837, "grad_norm": 0.5906180739402771, "learning_rate": 0.00025850464396473343, "loss": 5.1186, "step": 283900 }, { "epoch": 1.712171826196018, "grad_norm": 1.0607578754425049, "learning_rate": 0.00025849059268579785, "loss": 5.0046, "step": 283950 }, { "epoch": 1.712473317977352, "grad_norm": 1.8403308391571045, "learning_rate": 0.00025847653941024775, "loss": 4.8319, "step": 284000 }, { "epoch": 1.712774809758686, "grad_norm": 1.367615818977356, "learning_rate": 0.00025846248413834177, "loss": 4.7602, "step": 284050 }, { "epoch": 1.71307630154002, "grad_norm": 1.936359167098999, "learning_rate": 0.00025844842687033855, "loss": 4.8585, "step": 284100 }, { "epoch": 1.713377793321354, "grad_norm": 1.635105848312378, "learning_rate": 0.00025843436760649677, "loss": 5.1836, "step": 284150 }, { "epoch": 1.713679285102688, "grad_norm": 1.2862423658370972, "learning_rate": 0.0002584203063470752, "loss": 4.7499, "step": 284200 }, { "epoch": 1.7139807768840223, "grad_norm": 1.3785057067871094, "learning_rate": 0.0002584062430923326, "loss": 4.8573, "step": 284250 }, { "epoch": 1.714282268665356, "grad_norm": 1.2453644275665283, "learning_rate": 0.0002583921778425278, "loss": 5.1427, "step": 284300 }, { "epoch": 1.7145837604466903, "grad_norm": 1.440780520439148, "learning_rate": 0.0002583781105979197, "loss": 5.1573, "step": 284350 }, { "epoch": 1.7148852522280242, "grad_norm": 1.3295401334762573, "learning_rate": 0.0002583640413587671, "loss": 4.7933, "step": 284400 }, { "epoch": 1.7151867440093582, "grad_norm": 1.333013653755188, "learning_rate": 0.000258349970125329, "loss": 5.1635, "step": 284450 }, { "epoch": 1.7154882357906924, "grad_norm": 3.0289430618286133, "learning_rate": 0.00025833589689786424, "loss": 5.3068, "step": 284500 }, { "epoch": 1.7157897275720264, "grad_norm": 1.6869094371795654, "learning_rate": 0.000258321821676632, "loss": 4.885, "step": 284550 }, { "epoch": 1.7160912193533604, "grad_norm": 1.500382423400879, "learning_rate": 0.00025830774446189116, "loss": 4.8884, "step": 284600 }, { "epoch": 1.7163927111346946, "grad_norm": 1.6167839765548706, "learning_rate": 0.00025829366525390084, "loss": 5.2401, "step": 284650 }, { "epoch": 1.7166942029160284, "grad_norm": 2.079836845397949, "learning_rate": 0.0002582795840529201, "loss": 4.9009, "step": 284700 }, { "epoch": 1.7169956946973626, "grad_norm": 1.1099867820739746, "learning_rate": 0.00025826550085920826, "loss": 5.4522, "step": 284750 }, { "epoch": 1.7172971864786966, "grad_norm": 1.087445616722107, "learning_rate": 0.0002582514156730243, "loss": 5.093, "step": 284800 }, { "epoch": 1.7175986782600305, "grad_norm": 0.9574087858200073, "learning_rate": 0.0002582373284946276, "loss": 5.1904, "step": 284850 }, { "epoch": 1.7179001700413647, "grad_norm": 1.921156883239746, "learning_rate": 0.00025822323932427727, "loss": 5.1175, "step": 284900 }, { "epoch": 1.7182016618226987, "grad_norm": 1.2436124086380005, "learning_rate": 0.00025820914816223267, "loss": 4.942, "step": 284950 }, { "epoch": 1.7185031536040327, "grad_norm": 2.922903537750244, "learning_rate": 0.0002581950550087531, "loss": 5.3895, "step": 285000 }, { "epoch": 1.7185031536040327, "eval_loss": 5.533604621887207, "eval_runtime": 39.0572, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.312281431013742, "step": 285000 }, { "epoch": 1.718804645385367, "grad_norm": 1.7662001848220825, "learning_rate": 0.000258180959864098, "loss": 5.3056, "step": 285050 }, { "epoch": 1.7191061371667007, "grad_norm": 0.7075507640838623, "learning_rate": 0.0002581668627285267, "loss": 4.8489, "step": 285100 }, { "epoch": 1.719407628948035, "grad_norm": 1.3056856393814087, "learning_rate": 0.00025815276360229857, "loss": 5.0079, "step": 285150 }, { "epoch": 1.7197091207293689, "grad_norm": 0.8183543682098389, "learning_rate": 0.0002581386624856733, "loss": 5.0448, "step": 285200 }, { "epoch": 1.7200106125107029, "grad_norm": 1.749692440032959, "learning_rate": 0.0002581245593789102, "loss": 4.8423, "step": 285250 }, { "epoch": 1.720312104292037, "grad_norm": 1.9343551397323608, "learning_rate": 0.00025811045428226885, "loss": 5.3091, "step": 285300 }, { "epoch": 1.720613596073371, "grad_norm": 1.349454402923584, "learning_rate": 0.00025809634719600895, "loss": 5.0513, "step": 285350 }, { "epoch": 1.720915087854705, "grad_norm": 2.1073057651519775, "learning_rate": 0.00025808223812039, "loss": 5.2265, "step": 285400 }, { "epoch": 1.7212165796360392, "grad_norm": 1.647732138633728, "learning_rate": 0.0002580681270556717, "loss": 4.8773, "step": 285450 }, { "epoch": 1.721518071417373, "grad_norm": 0.7411003112792969, "learning_rate": 0.0002580540140021138, "loss": 4.7106, "step": 285500 }, { "epoch": 1.7218195631987072, "grad_norm": 2.305394411087036, "learning_rate": 0.00025803989895997585, "loss": 5.4467, "step": 285550 }, { "epoch": 1.7221210549800412, "grad_norm": 1.3450051546096802, "learning_rate": 0.0002580257819295178, "loss": 5.053, "step": 285600 }, { "epoch": 1.7224225467613752, "grad_norm": 1.9922411441802979, "learning_rate": 0.0002580116629109994, "loss": 4.8542, "step": 285650 }, { "epoch": 1.7227240385427094, "grad_norm": 1.4573838710784912, "learning_rate": 0.0002579975419046805, "loss": 5.0198, "step": 285700 }, { "epoch": 1.7230255303240434, "grad_norm": 1.2809432744979858, "learning_rate": 0.000257983418910821, "loss": 5.3328, "step": 285750 }, { "epoch": 1.7233270221053774, "grad_norm": 1.1290701627731323, "learning_rate": 0.0002579692939296807, "loss": 5.2831, "step": 285800 }, { "epoch": 1.7236285138867116, "grad_norm": 1.3584692478179932, "learning_rate": 0.0002579551669615197, "loss": 5.2079, "step": 285850 }, { "epoch": 1.7239300056680453, "grad_norm": 0.5665931701660156, "learning_rate": 0.00025794103800659784, "loss": 4.9237, "step": 285900 }, { "epoch": 1.7242314974493795, "grad_norm": 1.808962106704712, "learning_rate": 0.0002579269070651752, "loss": 5.0615, "step": 285950 }, { "epoch": 1.7245329892307135, "grad_norm": 1.3397554159164429, "learning_rate": 0.0002579127741375119, "loss": 5.26, "step": 286000 }, { "epoch": 1.7248344810120475, "grad_norm": 1.994986891746521, "learning_rate": 0.00025789863922386807, "loss": 5.0466, "step": 286050 }, { "epoch": 1.7251359727933817, "grad_norm": 1.0712196826934814, "learning_rate": 0.00025788450232450366, "loss": 4.7332, "step": 286100 }, { "epoch": 1.7254374645747157, "grad_norm": 1.5952677726745605, "learning_rate": 0.00025787036343967896, "loss": 4.8821, "step": 286150 }, { "epoch": 1.7257389563560497, "grad_norm": 2.0927224159240723, "learning_rate": 0.00025785622256965416, "loss": 5.1576, "step": 286200 }, { "epoch": 1.7260404481373839, "grad_norm": 0.9906880259513855, "learning_rate": 0.0002578420797146895, "loss": 4.8522, "step": 286250 }, { "epoch": 1.7263419399187176, "grad_norm": 1.7688453197479248, "learning_rate": 0.00025782793487504524, "loss": 5.3077, "step": 286300 }, { "epoch": 1.7266434317000519, "grad_norm": 1.9721075296401978, "learning_rate": 0.0002578137880509818, "loss": 4.857, "step": 286350 }, { "epoch": 1.726944923481386, "grad_norm": 0.8498314023017883, "learning_rate": 0.0002577996392427594, "loss": 4.9886, "step": 286400 }, { "epoch": 1.7272464152627198, "grad_norm": 1.6250865459442139, "learning_rate": 0.00025778548845063846, "loss": 5.5112, "step": 286450 }, { "epoch": 1.727547907044054, "grad_norm": 1.981697916984558, "learning_rate": 0.00025777133567487945, "loss": 5.3107, "step": 286500 }, { "epoch": 1.727849398825388, "grad_norm": 1.2338392734527588, "learning_rate": 0.0002577571809157428, "loss": 4.7319, "step": 286550 }, { "epoch": 1.728150890606722, "grad_norm": 1.428571105003357, "learning_rate": 0.000257743024173489, "loss": 4.9456, "step": 286600 }, { "epoch": 1.7284523823880562, "grad_norm": 1.790143370628357, "learning_rate": 0.00025772886544837863, "loss": 5.0888, "step": 286650 }, { "epoch": 1.7287538741693902, "grad_norm": 1.4519721269607544, "learning_rate": 0.0002577147047406722, "loss": 4.8091, "step": 286700 }, { "epoch": 1.7290553659507242, "grad_norm": 1.1822022199630737, "learning_rate": 0.0002577005420506304, "loss": 4.895, "step": 286750 }, { "epoch": 1.7293568577320584, "grad_norm": 1.0642255544662476, "learning_rate": 0.0002576863773785137, "loss": 5.2458, "step": 286800 }, { "epoch": 1.7296583495133921, "grad_norm": 2.19081711769104, "learning_rate": 0.000257672210724583, "loss": 4.8692, "step": 286850 }, { "epoch": 1.7299598412947264, "grad_norm": 2.1817920207977295, "learning_rate": 0.0002576580420890989, "loss": 5.2574, "step": 286900 }, { "epoch": 1.7302613330760603, "grad_norm": 1.076996922492981, "learning_rate": 0.00025764387147232215, "loss": 5.019, "step": 286950 }, { "epoch": 1.7305628248573943, "grad_norm": 1.1705827713012695, "learning_rate": 0.0002576296988745136, "loss": 5.1319, "step": 287000 }, { "epoch": 1.7308643166387285, "grad_norm": 1.2502938508987427, "learning_rate": 0.00025761552429593397, "loss": 5.1216, "step": 287050 }, { "epoch": 1.7311658084200625, "grad_norm": 2.3214874267578125, "learning_rate": 0.00025760134773684425, "loss": 4.9482, "step": 287100 }, { "epoch": 1.7314673002013965, "grad_norm": 1.4794073104858398, "learning_rate": 0.0002575871691975052, "loss": 5.1875, "step": 287150 }, { "epoch": 1.7317687919827307, "grad_norm": 1.305126667022705, "learning_rate": 0.0002575729886781779, "loss": 5.1714, "step": 287200 }, { "epoch": 1.7320702837640645, "grad_norm": 0.7493317127227783, "learning_rate": 0.0002575588061791233, "loss": 5.4529, "step": 287250 }, { "epoch": 1.7323717755453987, "grad_norm": 3.1893045902252197, "learning_rate": 0.00025754462170060236, "loss": 5.4148, "step": 287300 }, { "epoch": 1.7326732673267327, "grad_norm": 2.1748509407043457, "learning_rate": 0.0002575304352428761, "loss": 4.8461, "step": 287350 }, { "epoch": 1.7329747591080666, "grad_norm": 1.460519552230835, "learning_rate": 0.0002575162468062056, "loss": 5.2699, "step": 287400 }, { "epoch": 1.7332762508894008, "grad_norm": 1.76861572265625, "learning_rate": 0.00025750205639085206, "loss": 5.3038, "step": 287450 }, { "epoch": 1.7335777426707348, "grad_norm": 1.3930680751800537, "learning_rate": 0.0002574878639970765, "loss": 5.3782, "step": 287500 }, { "epoch": 1.7338792344520688, "grad_norm": 2.1344847679138184, "learning_rate": 0.0002574736696251403, "loss": 5.2819, "step": 287550 }, { "epoch": 1.734180726233403, "grad_norm": 1.4091404676437378, "learning_rate": 0.00025745947327530456, "loss": 5.041, "step": 287600 }, { "epoch": 1.7344822180147368, "grad_norm": 0.517598032951355, "learning_rate": 0.00025744527494783053, "loss": 4.7916, "step": 287650 }, { "epoch": 1.734783709796071, "grad_norm": 1.2441086769104004, "learning_rate": 0.0002574310746429796, "loss": 5.3212, "step": 287700 }, { "epoch": 1.735085201577405, "grad_norm": 1.1996467113494873, "learning_rate": 0.000257416872361013, "loss": 4.995, "step": 287750 }, { "epoch": 1.735386693358739, "grad_norm": 1.9420945644378662, "learning_rate": 0.00025740266810219217, "loss": 4.6359, "step": 287800 }, { "epoch": 1.7356881851400732, "grad_norm": 1.6674786806106567, "learning_rate": 0.00025738846186677854, "loss": 4.9041, "step": 287850 }, { "epoch": 1.7359896769214072, "grad_norm": 1.6945830583572388, "learning_rate": 0.0002573742536550335, "loss": 5.2293, "step": 287900 }, { "epoch": 1.7362911687027411, "grad_norm": 0.8114044666290283, "learning_rate": 0.0002573600434672186, "loss": 5.0992, "step": 287950 }, { "epoch": 1.7365926604840753, "grad_norm": 1.2082996368408203, "learning_rate": 0.0002573458313035952, "loss": 4.9944, "step": 288000 }, { "epoch": 1.7365926604840753, "eval_loss": 5.549341201782227, "eval_runtime": 38.8383, "eval_samples_per_second": 13.183, "eval_steps_per_second": 6.591, "eval_tts_loss": 7.324844133886893, "step": 288000 }, { "epoch": 1.736894152265409, "grad_norm": 1.288675308227539, "learning_rate": 0.00025733161716442505, "loss": 4.9772, "step": 288050 }, { "epoch": 1.7371956440467433, "grad_norm": 1.807084321975708, "learning_rate": 0.0002573174010499696, "loss": 5.522, "step": 288100 }, { "epoch": 1.7374971358280773, "grad_norm": 1.750871181488037, "learning_rate": 0.0002573031829604906, "loss": 5.1304, "step": 288150 }, { "epoch": 1.7377986276094113, "grad_norm": 1.8540598154067993, "learning_rate": 0.0002572889628962496, "loss": 5.3564, "step": 288200 }, { "epoch": 1.7381001193907455, "grad_norm": 1.3645986318588257, "learning_rate": 0.0002572747408575084, "loss": 5.5113, "step": 288250 }, { "epoch": 1.7384016111720795, "grad_norm": 1.1879253387451172, "learning_rate": 0.00025726051684452864, "loss": 4.9295, "step": 288300 }, { "epoch": 1.7387031029534135, "grad_norm": 2.127316474914551, "learning_rate": 0.00025724629085757217, "loss": 5.3579, "step": 288350 }, { "epoch": 1.7390045947347477, "grad_norm": 2.0292258262634277, "learning_rate": 0.0002572320628969007, "loss": 4.9643, "step": 288400 }, { "epoch": 1.7393060865160814, "grad_norm": 0.6165685057640076, "learning_rate": 0.0002572178329627762, "loss": 5.1595, "step": 288450 }, { "epoch": 1.7396075782974156, "grad_norm": 1.2010053396224976, "learning_rate": 0.00025720360105546046, "loss": 5.2447, "step": 288500 }, { "epoch": 1.7399090700787496, "grad_norm": 1.4280214309692383, "learning_rate": 0.0002571893671752155, "loss": 5.2807, "step": 288550 }, { "epoch": 1.7402105618600836, "grad_norm": 0.5983676314353943, "learning_rate": 0.0002571751313223031, "loss": 5.1206, "step": 288600 }, { "epoch": 1.7405120536414178, "grad_norm": 1.6075204610824585, "learning_rate": 0.0002571608934969854, "loss": 4.9251, "step": 288650 }, { "epoch": 1.7408135454227518, "grad_norm": 1.8794037103652954, "learning_rate": 0.0002571466536995244, "loss": 5.4359, "step": 288700 }, { "epoch": 1.7411150372040858, "grad_norm": 0.7782021760940552, "learning_rate": 0.0002571324119301821, "loss": 5.2269, "step": 288750 }, { "epoch": 1.74141652898542, "grad_norm": 1.1087394952774048, "learning_rate": 0.00025711816818922066, "loss": 5.0673, "step": 288800 }, { "epoch": 1.7417180207667537, "grad_norm": 1.5785061120986938, "learning_rate": 0.0002571039224769023, "loss": 5.4192, "step": 288850 }, { "epoch": 1.742019512548088, "grad_norm": 0.8351266980171204, "learning_rate": 0.00025708967479348894, "loss": 5.0052, "step": 288900 }, { "epoch": 1.742321004329422, "grad_norm": 1.2717211246490479, "learning_rate": 0.00025707542513924304, "loss": 5.2848, "step": 288950 }, { "epoch": 1.742622496110756, "grad_norm": 1.2042393684387207, "learning_rate": 0.0002570611735144267, "loss": 4.7033, "step": 289000 }, { "epoch": 1.7429239878920901, "grad_norm": 1.98430335521698, "learning_rate": 0.0002570469199193023, "loss": 4.8134, "step": 289050 }, { "epoch": 1.7432254796734241, "grad_norm": 1.98003089427948, "learning_rate": 0.000257032664354132, "loss": 4.8776, "step": 289100 }, { "epoch": 1.743526971454758, "grad_norm": 1.488688349723816, "learning_rate": 0.0002570184068191783, "loss": 4.9932, "step": 289150 }, { "epoch": 1.7438284632360923, "grad_norm": 2.3813822269439697, "learning_rate": 0.0002570041473147036, "loss": 4.9284, "step": 289200 }, { "epoch": 1.744129955017426, "grad_norm": 1.8234659433364868, "learning_rate": 0.00025698988584097024, "loss": 5.0279, "step": 289250 }, { "epoch": 1.7444314467987603, "grad_norm": 1.5042943954467773, "learning_rate": 0.0002569756223982407, "loss": 5.0601, "step": 289300 }, { "epoch": 1.7447329385800943, "grad_norm": 1.4371713399887085, "learning_rate": 0.0002569613569867775, "loss": 5.1341, "step": 289350 }, { "epoch": 1.7450344303614282, "grad_norm": 1.6402641534805298, "learning_rate": 0.0002569470896068432, "loss": 4.8362, "step": 289400 }, { "epoch": 1.7453359221427625, "grad_norm": 0.7516206502914429, "learning_rate": 0.0002569328202587003, "loss": 5.0272, "step": 289450 }, { "epoch": 1.7456374139240964, "grad_norm": 1.4598087072372437, "learning_rate": 0.0002569185489426114, "loss": 5.1416, "step": 289500 }, { "epoch": 1.7459389057054304, "grad_norm": 1.6147011518478394, "learning_rate": 0.00025690427565883925, "loss": 5.1588, "step": 289550 }, { "epoch": 1.7462403974867646, "grad_norm": 1.1116734743118286, "learning_rate": 0.0002568900004076464, "loss": 4.936, "step": 289600 }, { "epoch": 1.7465418892680984, "grad_norm": 1.9873515367507935, "learning_rate": 0.00025687572318929577, "loss": 5.4878, "step": 289650 }, { "epoch": 1.7468433810494326, "grad_norm": 1.3345580101013184, "learning_rate": 0.0002568614440040499, "loss": 4.8792, "step": 289700 }, { "epoch": 1.7471448728307666, "grad_norm": 1.1108676195144653, "learning_rate": 0.00025684716285217157, "loss": 5.0708, "step": 289750 }, { "epoch": 1.7474463646121006, "grad_norm": 1.5521506071090698, "learning_rate": 0.0002568328797339238, "loss": 5.287, "step": 289800 }, { "epoch": 1.7477478563934348, "grad_norm": 1.5151070356369019, "learning_rate": 0.0002568185946495693, "loss": 4.983, "step": 289850 }, { "epoch": 1.7480493481747688, "grad_norm": 1.5616565942764282, "learning_rate": 0.000256804307599371, "loss": 4.8483, "step": 289900 }, { "epoch": 1.7483508399561027, "grad_norm": 1.3577616214752197, "learning_rate": 0.0002567900185835918, "loss": 4.8958, "step": 289950 }, { "epoch": 1.748652331737437, "grad_norm": 2.350933790206909, "learning_rate": 0.0002567757276024948, "loss": 4.7918, "step": 290000 }, { "epoch": 1.748953823518771, "grad_norm": 0.5174716114997864, "learning_rate": 0.00025676143465634286, "loss": 4.8968, "step": 290050 }, { "epoch": 1.749255315300105, "grad_norm": 1.635595440864563, "learning_rate": 0.00025674713974539905, "loss": 5.2064, "step": 290100 }, { "epoch": 1.7495568070814391, "grad_norm": 1.1445696353912354, "learning_rate": 0.0002567328428699265, "loss": 5.0963, "step": 290150 }, { "epoch": 1.7498582988627729, "grad_norm": 0.9328393936157227, "learning_rate": 0.00025671854403018827, "loss": 4.781, "step": 290200 }, { "epoch": 1.750159790644107, "grad_norm": 2.1882171630859375, "learning_rate": 0.00025670424322644757, "loss": 5.1103, "step": 290250 }, { "epoch": 1.750461282425441, "grad_norm": 2.1560893058776855, "learning_rate": 0.00025668994045896753, "loss": 4.9206, "step": 290300 }, { "epoch": 1.750762774206775, "grad_norm": 1.2438950538635254, "learning_rate": 0.00025667563572801135, "loss": 5.3746, "step": 290350 }, { "epoch": 1.7510642659881093, "grad_norm": 1.0187861919403076, "learning_rate": 0.0002566613290338423, "loss": 5.5084, "step": 290400 }, { "epoch": 1.7513657577694433, "grad_norm": 1.2348217964172363, "learning_rate": 0.0002566470203767238, "loss": 5.2491, "step": 290450 }, { "epoch": 1.7516672495507772, "grad_norm": 2.2008352279663086, "learning_rate": 0.00025663270975691905, "loss": 4.9442, "step": 290500 }, { "epoch": 1.7519687413321114, "grad_norm": 1.7657009363174438, "learning_rate": 0.0002566183971746915, "loss": 5.557, "step": 290550 }, { "epoch": 1.7522702331134452, "grad_norm": 0.8872597813606262, "learning_rate": 0.0002566040826303044, "loss": 5.0302, "step": 290600 }, { "epoch": 1.7525717248947794, "grad_norm": 1.4577187299728394, "learning_rate": 0.00025658976612402137, "loss": 5.0032, "step": 290650 }, { "epoch": 1.7528732166761134, "grad_norm": 1.1110814809799194, "learning_rate": 0.00025657544765610575, "loss": 5.2361, "step": 290700 }, { "epoch": 1.7531747084574474, "grad_norm": 2.309576988220215, "learning_rate": 0.00025656112722682113, "loss": 5.1919, "step": 290750 }, { "epoch": 1.7534762002387816, "grad_norm": 1.4136079549789429, "learning_rate": 0.00025654680483643103, "loss": 5.1292, "step": 290800 }, { "epoch": 1.7537776920201156, "grad_norm": 1.3171770572662354, "learning_rate": 0.00025653248048519906, "loss": 4.9977, "step": 290850 }, { "epoch": 1.7540791838014496, "grad_norm": 1.4042543172836304, "learning_rate": 0.0002565181541733888, "loss": 5.0798, "step": 290900 }, { "epoch": 1.7543806755827838, "grad_norm": 1.3419686555862427, "learning_rate": 0.00025650382590126393, "loss": 5.2407, "step": 290950 }, { "epoch": 1.7546821673641175, "grad_norm": 2.1127231121063232, "learning_rate": 0.0002564894956690881, "loss": 5.2211, "step": 291000 }, { "epoch": 1.7546821673641175, "eval_loss": 5.539180755615234, "eval_runtime": 39.158, "eval_samples_per_second": 13.075, "eval_steps_per_second": 6.538, "eval_tts_loss": 7.2547622360811905, "step": 291000 }, { "epoch": 1.7549836591454517, "grad_norm": 1.9204357862472534, "learning_rate": 0.0002564751634771251, "loss": 5.0606, "step": 291050 }, { "epoch": 1.7552851509267857, "grad_norm": 1.4294734001159668, "learning_rate": 0.0002564608293256387, "loss": 4.9924, "step": 291100 }, { "epoch": 1.7555866427081197, "grad_norm": 0.936290442943573, "learning_rate": 0.0002564464932148926, "loss": 5.0428, "step": 291150 }, { "epoch": 1.755888134489454, "grad_norm": 1.8725082874298096, "learning_rate": 0.0002564321551451507, "loss": 5.3611, "step": 291200 }, { "epoch": 1.756189626270788, "grad_norm": 0.25161468982696533, "learning_rate": 0.00025641781511667695, "loss": 5.0095, "step": 291250 }, { "epoch": 1.7564911180521219, "grad_norm": 1.8153867721557617, "learning_rate": 0.00025640347312973514, "loss": 5.32, "step": 291300 }, { "epoch": 1.756792609833456, "grad_norm": 1.3528834581375122, "learning_rate": 0.00025638912918458923, "loss": 4.943, "step": 291350 }, { "epoch": 1.7570941016147898, "grad_norm": 1.6349341869354248, "learning_rate": 0.0002563747832815032, "loss": 5.0846, "step": 291400 }, { "epoch": 1.757395593396124, "grad_norm": 1.1890966892242432, "learning_rate": 0.00025636043542074115, "loss": 5.6564, "step": 291450 }, { "epoch": 1.757697085177458, "grad_norm": 1.1319109201431274, "learning_rate": 0.00025634608560256704, "loss": 5.1949, "step": 291500 }, { "epoch": 1.757998576958792, "grad_norm": 1.6815693378448486, "learning_rate": 0.00025633173382724495, "loss": 5.3838, "step": 291550 }, { "epoch": 1.7583000687401262, "grad_norm": 1.3531800508499146, "learning_rate": 0.0002563173800950391, "loss": 5.3745, "step": 291600 }, { "epoch": 1.7586015605214602, "grad_norm": 1.4113394021987915, "learning_rate": 0.0002563030244062136, "loss": 4.6433, "step": 291650 }, { "epoch": 1.7589030523027942, "grad_norm": 1.8147670030593872, "learning_rate": 0.0002562886667610326, "loss": 4.9345, "step": 291700 }, { "epoch": 1.7592045440841284, "grad_norm": 1.6569451093673706, "learning_rate": 0.0002562743071597604, "loss": 5.5045, "step": 291750 }, { "epoch": 1.7595060358654622, "grad_norm": 1.107985258102417, "learning_rate": 0.0002562599456026612, "loss": 5.3337, "step": 291800 }, { "epoch": 1.7598075276467964, "grad_norm": 1.6269336938858032, "learning_rate": 0.00025624558208999935, "loss": 4.7583, "step": 291850 }, { "epoch": 1.7601090194281304, "grad_norm": 1.2394888401031494, "learning_rate": 0.00025623121662203926, "loss": 5.4726, "step": 291900 }, { "epoch": 1.7604105112094643, "grad_norm": 0.9519831538200378, "learning_rate": 0.00025621684919904515, "loss": 5.0867, "step": 291950 }, { "epoch": 1.7607120029907986, "grad_norm": 1.8748273849487305, "learning_rate": 0.00025620247982128153, "loss": 5.4374, "step": 292000 }, { "epoch": 1.7610134947721325, "grad_norm": 2.0728797912597656, "learning_rate": 0.0002561881084890128, "loss": 4.5121, "step": 292050 }, { "epoch": 1.7613149865534665, "grad_norm": 1.331089735031128, "learning_rate": 0.0002561737352025035, "loss": 5.4421, "step": 292100 }, { "epoch": 1.7616164783348007, "grad_norm": 1.670708179473877, "learning_rate": 0.00025615935996201814, "loss": 4.4965, "step": 292150 }, { "epoch": 1.7619179701161345, "grad_norm": 1.336700677871704, "learning_rate": 0.0002561449827678212, "loss": 4.8109, "step": 292200 }, { "epoch": 1.7622194618974687, "grad_norm": 1.3850988149642944, "learning_rate": 0.00025613060362017736, "loss": 5.6135, "step": 292250 }, { "epoch": 1.7625209536788027, "grad_norm": 1.948289394378662, "learning_rate": 0.0002561162225193512, "loss": 5.0188, "step": 292300 }, { "epoch": 1.7628224454601367, "grad_norm": 2.096419095993042, "learning_rate": 0.00025610183946560744, "loss": 4.9464, "step": 292350 }, { "epoch": 1.7631239372414709, "grad_norm": 0.9257486462593079, "learning_rate": 0.0002560874544592107, "loss": 4.8806, "step": 292400 }, { "epoch": 1.7634254290228049, "grad_norm": 1.3979673385620117, "learning_rate": 0.0002560730675004258, "loss": 5.1542, "step": 292450 }, { "epoch": 1.7637269208041388, "grad_norm": 2.0444657802581787, "learning_rate": 0.00025605867858951746, "loss": 5.3145, "step": 292500 }, { "epoch": 1.764028412585473, "grad_norm": 1.5602535009384155, "learning_rate": 0.0002560442877267505, "loss": 5.0088, "step": 292550 }, { "epoch": 1.7643299043668068, "grad_norm": 1.354636311531067, "learning_rate": 0.0002560298949123897, "loss": 5.2134, "step": 292600 }, { "epoch": 1.764631396148141, "grad_norm": 1.7044451236724854, "learning_rate": 0.00025601550014670005, "loss": 5.1354, "step": 292650 }, { "epoch": 1.764932887929475, "grad_norm": 1.5892082452774048, "learning_rate": 0.00025600110342994636, "loss": 4.9531, "step": 292700 }, { "epoch": 1.765234379710809, "grad_norm": 1.7706494331359863, "learning_rate": 0.00025598670476239367, "loss": 4.811, "step": 292750 }, { "epoch": 1.7655358714921432, "grad_norm": 3.3899378776550293, "learning_rate": 0.00025597230414430694, "loss": 4.6925, "step": 292800 }, { "epoch": 1.7658373632734772, "grad_norm": 0.9999024271965027, "learning_rate": 0.00025595790157595117, "loss": 5.2298, "step": 292850 }, { "epoch": 1.7661388550548112, "grad_norm": 1.9693193435668945, "learning_rate": 0.00025594349705759144, "loss": 5.1618, "step": 292900 }, { "epoch": 1.7664403468361454, "grad_norm": 0.9553192257881165, "learning_rate": 0.00025592909058949284, "loss": 4.827, "step": 292950 }, { "epoch": 1.7667418386174791, "grad_norm": 2.0715417861938477, "learning_rate": 0.0002559146821719205, "loss": 5.4347, "step": 293000 }, { "epoch": 1.7670433303988133, "grad_norm": 0.4799889028072357, "learning_rate": 0.0002559002718051395, "loss": 4.6804, "step": 293050 }, { "epoch": 1.7673448221801473, "grad_norm": 1.824138879776001, "learning_rate": 0.00025588585948941526, "loss": 5.3697, "step": 293100 }, { "epoch": 1.7676463139614813, "grad_norm": 1.060971975326538, "learning_rate": 0.0002558714452250128, "loss": 5.3624, "step": 293150 }, { "epoch": 1.7679478057428155, "grad_norm": 1.7641654014587402, "learning_rate": 0.00025585702901219744, "loss": 4.942, "step": 293200 }, { "epoch": 1.7682492975241495, "grad_norm": 1.9271626472473145, "learning_rate": 0.00025584261085123466, "loss": 5.0231, "step": 293250 }, { "epoch": 1.7685507893054835, "grad_norm": 1.161774754524231, "learning_rate": 0.0002558281907423896, "loss": 5.0477, "step": 293300 }, { "epoch": 1.7688522810868177, "grad_norm": 1.8378363847732544, "learning_rate": 0.00025581376868592765, "loss": 5.3049, "step": 293350 }, { "epoch": 1.7691537728681515, "grad_norm": 1.7620183229446411, "learning_rate": 0.00025579934468211436, "loss": 4.914, "step": 293400 }, { "epoch": 1.7694552646494857, "grad_norm": 1.835019826889038, "learning_rate": 0.00025578491873121513, "loss": 5.1785, "step": 293450 }, { "epoch": 1.7697567564308196, "grad_norm": 1.5970535278320312, "learning_rate": 0.0002557704908334954, "loss": 5.2649, "step": 293500 }, { "epoch": 1.7700582482121536, "grad_norm": 1.5889419317245483, "learning_rate": 0.0002557560609892207, "loss": 5.0976, "step": 293550 }, { "epoch": 1.7703597399934878, "grad_norm": 2.2295095920562744, "learning_rate": 0.0002557416291986567, "loss": 5.2602, "step": 293600 }, { "epoch": 1.7706612317748218, "grad_norm": 1.0924291610717773, "learning_rate": 0.0002557271954620688, "loss": 4.8792, "step": 293650 }, { "epoch": 1.7709627235561558, "grad_norm": 1.5674395561218262, "learning_rate": 0.00025571275977972287, "loss": 5.2258, "step": 293700 }, { "epoch": 1.77126421533749, "grad_norm": 2.042534828186035, "learning_rate": 0.00025569832215188444, "loss": 5.0672, "step": 293750 }, { "epoch": 1.771565707118824, "grad_norm": 1.3464261293411255, "learning_rate": 0.0002556838825788192, "loss": 5.1375, "step": 293800 }, { "epoch": 1.771867198900158, "grad_norm": 2.1689062118530273, "learning_rate": 0.0002556694410607929, "loss": 5.2273, "step": 293850 }, { "epoch": 1.7721686906814922, "grad_norm": 2.743601083755493, "learning_rate": 0.0002556549975980714, "loss": 5.1604, "step": 293900 }, { "epoch": 1.772470182462826, "grad_norm": 4.832919597625732, "learning_rate": 0.00025564055219092033, "loss": 5.0718, "step": 293950 }, { "epoch": 1.7727716742441602, "grad_norm": 0.9972565770149231, "learning_rate": 0.00025562610483960576, "loss": 4.4125, "step": 294000 }, { "epoch": 1.7727716742441602, "eval_loss": 5.5302839279174805, "eval_runtime": 39.2024, "eval_samples_per_second": 13.06, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.285705649536427, "step": 294000 }, { "epoch": 1.7730731660254941, "grad_norm": 1.7865039110183716, "learning_rate": 0.00025561165554439344, "loss": 4.9886, "step": 294050 }, { "epoch": 1.7733746578068281, "grad_norm": 1.299277663230896, "learning_rate": 0.00025559720430554924, "loss": 5.1017, "step": 294100 }, { "epoch": 1.7736761495881623, "grad_norm": 2.0895330905914307, "learning_rate": 0.00025558275112333933, "loss": 5.1537, "step": 294150 }, { "epoch": 1.7739776413694963, "grad_norm": 1.205875039100647, "learning_rate": 0.0002555682959980295, "loss": 5.0067, "step": 294200 }, { "epoch": 1.7742791331508303, "grad_norm": 1.5945661067962646, "learning_rate": 0.0002555538389298858, "loss": 4.8948, "step": 294250 }, { "epoch": 1.7745806249321645, "grad_norm": 0.7842682600021362, "learning_rate": 0.00025553937991917435, "loss": 4.5545, "step": 294300 }, { "epoch": 1.7748821167134983, "grad_norm": 1.8484278917312622, "learning_rate": 0.0002555249189661612, "loss": 4.8587, "step": 294350 }, { "epoch": 1.7751836084948325, "grad_norm": 2.9426004886627197, "learning_rate": 0.00025551045607111254, "loss": 5.2012, "step": 294400 }, { "epoch": 1.7754851002761665, "grad_norm": 1.1057796478271484, "learning_rate": 0.0002554959912342945, "loss": 4.6161, "step": 294450 }, { "epoch": 1.7757865920575004, "grad_norm": 1.4831757545471191, "learning_rate": 0.0002554815244559733, "loss": 5.1326, "step": 294500 }, { "epoch": 1.7760880838388347, "grad_norm": 1.3191787004470825, "learning_rate": 0.00025546705573641516, "loss": 5.062, "step": 294550 }, { "epoch": 1.7763895756201686, "grad_norm": 1.6733365058898926, "learning_rate": 0.0002554525850758864, "loss": 4.87, "step": 294600 }, { "epoch": 1.7766910674015026, "grad_norm": 3.2909603118896484, "learning_rate": 0.00025543811247465326, "loss": 5.2, "step": 294650 }, { "epoch": 1.7769925591828368, "grad_norm": 1.3901344537734985, "learning_rate": 0.00025542363793298217, "loss": 5.0071, "step": 294700 }, { "epoch": 1.7772940509641706, "grad_norm": 1.4862306118011475, "learning_rate": 0.00025540916145113947, "loss": 4.9156, "step": 294750 }, { "epoch": 1.7775955427455048, "grad_norm": 2.9925146102905273, "learning_rate": 0.0002553946830293915, "loss": 4.4276, "step": 294800 }, { "epoch": 1.7778970345268388, "grad_norm": 0.6811074614524841, "learning_rate": 0.00025538020266800487, "loss": 5.0994, "step": 294850 }, { "epoch": 1.7781985263081728, "grad_norm": 2.0463945865631104, "learning_rate": 0.00025536572036724596, "loss": 5.488, "step": 294900 }, { "epoch": 1.778500018089507, "grad_norm": 1.4868230819702148, "learning_rate": 0.00025535123612738134, "loss": 5.1902, "step": 294950 }, { "epoch": 1.778801509870841, "grad_norm": 1.21467125415802, "learning_rate": 0.0002553367499486776, "loss": 5.2735, "step": 295000 }, { "epoch": 1.779103001652175, "grad_norm": 1.3754189014434814, "learning_rate": 0.0002553222618314013, "loss": 4.999, "step": 295050 }, { "epoch": 1.7794044934335091, "grad_norm": 11.631924629211426, "learning_rate": 0.000255307771775819, "loss": 5.4242, "step": 295100 }, { "epoch": 1.779705985214843, "grad_norm": 1.8711192607879639, "learning_rate": 0.00025529327978219754, "loss": 4.984, "step": 295150 }, { "epoch": 1.7800074769961771, "grad_norm": 1.1589736938476562, "learning_rate": 0.00025527878585080343, "loss": 5.5069, "step": 295200 }, { "epoch": 1.780308968777511, "grad_norm": 1.5476336479187012, "learning_rate": 0.00025526428998190356, "loss": 5.0619, "step": 295250 }, { "epoch": 1.780610460558845, "grad_norm": 1.961417555809021, "learning_rate": 0.00025524979217576464, "loss": 4.7153, "step": 295300 }, { "epoch": 1.7809119523401793, "grad_norm": 1.3008366823196411, "learning_rate": 0.00025523529243265354, "loss": 5.2186, "step": 295350 }, { "epoch": 1.7812134441215133, "grad_norm": 1.445340633392334, "learning_rate": 0.000255220790752837, "loss": 5.2182, "step": 295400 }, { "epoch": 1.7815149359028473, "grad_norm": 0.41158875823020935, "learning_rate": 0.000255206287136582, "loss": 5.1897, "step": 295450 }, { "epoch": 1.7818164276841815, "grad_norm": 1.4770948886871338, "learning_rate": 0.0002551917815841554, "loss": 4.9793, "step": 295500 }, { "epoch": 1.7821179194655152, "grad_norm": 2.287393569946289, "learning_rate": 0.0002551772740958242, "loss": 5.318, "step": 295550 }, { "epoch": 1.7824194112468494, "grad_norm": 1.6954338550567627, "learning_rate": 0.00025516276467185537, "loss": 5.3615, "step": 295600 }, { "epoch": 1.7827209030281834, "grad_norm": 1.3776477575302124, "learning_rate": 0.0002551482533125159, "loss": 5.1601, "step": 295650 }, { "epoch": 1.7830223948095174, "grad_norm": 1.3488483428955078, "learning_rate": 0.0002551337400180729, "loss": 4.6716, "step": 295700 }, { "epoch": 1.7833238865908516, "grad_norm": 1.224694013595581, "learning_rate": 0.0002551192247887934, "loss": 4.8847, "step": 295750 }, { "epoch": 1.7836253783721856, "grad_norm": 1.4939262866973877, "learning_rate": 0.00025510470762494467, "loss": 5.246, "step": 295800 }, { "epoch": 1.7839268701535196, "grad_norm": 0.9691727161407471, "learning_rate": 0.00025509018852679376, "loss": 5.0054, "step": 295850 }, { "epoch": 1.7842283619348538, "grad_norm": 2.0056700706481934, "learning_rate": 0.0002550756674946079, "loss": 5.0324, "step": 295900 }, { "epoch": 1.7845298537161876, "grad_norm": 1.314873456954956, "learning_rate": 0.0002550611445286543, "loss": 4.9831, "step": 295950 }, { "epoch": 1.7848313454975218, "grad_norm": 1.792862892150879, "learning_rate": 0.0002550466196292003, "loss": 5.0361, "step": 296000 }, { "epoch": 1.7851328372788557, "grad_norm": 1.811560869216919, "learning_rate": 0.00025503209279651313, "loss": 4.9704, "step": 296050 }, { "epoch": 1.7854343290601897, "grad_norm": 1.0152555704116821, "learning_rate": 0.0002550175640308602, "loss": 5.1328, "step": 296100 }, { "epoch": 1.785735820841524, "grad_norm": 1.283992886543274, "learning_rate": 0.0002550030333325089, "loss": 5.186, "step": 296150 }, { "epoch": 1.786037312622858, "grad_norm": 1.8306527137756348, "learning_rate": 0.0002549885007017266, "loss": 4.6839, "step": 296200 }, { "epoch": 1.786338804404192, "grad_norm": 1.1678756475448608, "learning_rate": 0.00025497396613878077, "loss": 4.5083, "step": 296250 }, { "epoch": 1.7866402961855261, "grad_norm": 1.2936204671859741, "learning_rate": 0.0002549594296439389, "loss": 5.142, "step": 296300 }, { "epoch": 1.7869417879668599, "grad_norm": 1.1792701482772827, "learning_rate": 0.00025494489121746853, "loss": 4.9778, "step": 296350 }, { "epoch": 1.787243279748194, "grad_norm": 1.3544262647628784, "learning_rate": 0.00025493035085963713, "loss": 5.3287, "step": 296400 }, { "epoch": 1.787544771529528, "grad_norm": 1.3649744987487793, "learning_rate": 0.0002549158085707125, "loss": 5.0499, "step": 296450 }, { "epoch": 1.787846263310862, "grad_norm": 0.4105837047100067, "learning_rate": 0.000254901264350962, "loss": 4.5887, "step": 296500 }, { "epoch": 1.7881477550921963, "grad_norm": 1.8457356691360474, "learning_rate": 0.00025488671820065353, "loss": 5.0693, "step": 296550 }, { "epoch": 1.7884492468735302, "grad_norm": 1.4676077365875244, "learning_rate": 0.00025487217012005466, "loss": 5.1561, "step": 296600 }, { "epoch": 1.7887507386548642, "grad_norm": 2.330418109893799, "learning_rate": 0.00025485762010943314, "loss": 5.1223, "step": 296650 }, { "epoch": 1.7890522304361984, "grad_norm": 1.2249845266342163, "learning_rate": 0.00025484306816905676, "loss": 5.0856, "step": 296700 }, { "epoch": 1.7893537222175322, "grad_norm": 1.3989886045455933, "learning_rate": 0.00025482851429919337, "loss": 5.1926, "step": 296750 }, { "epoch": 1.7896552139988664, "grad_norm": 1.3549292087554932, "learning_rate": 0.0002548139585001107, "loss": 4.8966, "step": 296800 }, { "epoch": 1.7899567057802004, "grad_norm": 1.9585729837417603, "learning_rate": 0.00025479940077207676, "loss": 5.3571, "step": 296850 }, { "epoch": 1.7902581975615344, "grad_norm": 0.7895092368125916, "learning_rate": 0.0002547848411153594, "loss": 5.0842, "step": 296900 }, { "epoch": 1.7905596893428686, "grad_norm": 1.473893642425537, "learning_rate": 0.0002547702795302266, "loss": 5.1846, "step": 296950 }, { "epoch": 1.7908611811242026, "grad_norm": 1.3873460292816162, "learning_rate": 0.00025475571601694626, "loss": 4.8496, "step": 297000 }, { "epoch": 1.7908611811242026, "eval_loss": 5.538431167602539, "eval_runtime": 39.1106, "eval_samples_per_second": 13.091, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.314097662197406, "step": 297000 }, { "epoch": 1.7911626729055365, "grad_norm": 1.3941352367401123, "learning_rate": 0.0002547411505757865, "loss": 5.5195, "step": 297050 }, { "epoch": 1.7914641646868708, "grad_norm": 1.2045384645462036, "learning_rate": 0.0002547265832070154, "loss": 4.9932, "step": 297100 }, { "epoch": 1.7917656564682045, "grad_norm": 1.3989869356155396, "learning_rate": 0.00025471201391090084, "loss": 4.9204, "step": 297150 }, { "epoch": 1.7920671482495387, "grad_norm": 2.0926222801208496, "learning_rate": 0.00025469744268771116, "loss": 5.2288, "step": 297200 }, { "epoch": 1.7923686400308727, "grad_norm": 1.5490082502365112, "learning_rate": 0.0002546828695377145, "loss": 5.026, "step": 297250 }, { "epoch": 1.7926701318122067, "grad_norm": 1.5660889148712158, "learning_rate": 0.000254668294461179, "loss": 5.1622, "step": 297300 }, { "epoch": 1.792971623593541, "grad_norm": 1.5624371767044067, "learning_rate": 0.00025465371745837285, "loss": 4.9448, "step": 297350 }, { "epoch": 1.7932731153748749, "grad_norm": 1.0881998538970947, "learning_rate": 0.0002546391385295645, "loss": 4.8071, "step": 297400 }, { "epoch": 1.7935746071562089, "grad_norm": 0.7959088683128357, "learning_rate": 0.000254624557675022, "loss": 4.8021, "step": 297450 }, { "epoch": 1.793876098937543, "grad_norm": 1.5063650608062744, "learning_rate": 0.0002546099748950139, "loss": 4.866, "step": 297500 }, { "epoch": 1.794177590718877, "grad_norm": 1.5417810678482056, "learning_rate": 0.00025459539018980845, "loss": 4.463, "step": 297550 }, { "epoch": 1.794479082500211, "grad_norm": 1.2132543325424194, "learning_rate": 0.00025458080355967417, "loss": 4.9005, "step": 297600 }, { "epoch": 1.7947805742815452, "grad_norm": 1.3540395498275757, "learning_rate": 0.0002545662150048794, "loss": 5.2446, "step": 297650 }, { "epoch": 1.795082066062879, "grad_norm": 1.5452055931091309, "learning_rate": 0.0002545516245256927, "loss": 5.2907, "step": 297700 }, { "epoch": 1.7953835578442132, "grad_norm": 1.677780032157898, "learning_rate": 0.00025453703212238255, "loss": 5.3337, "step": 297750 }, { "epoch": 1.7956850496255472, "grad_norm": 0.42064064741134644, "learning_rate": 0.0002545224377952175, "loss": 5.199, "step": 297800 }, { "epoch": 1.7959865414068812, "grad_norm": 1.9668972492218018, "learning_rate": 0.00025450784154446606, "loss": 5.1896, "step": 297850 }, { "epoch": 1.7962880331882154, "grad_norm": 0.7440846562385559, "learning_rate": 0.0002544932433703971, "loss": 5.3533, "step": 297900 }, { "epoch": 1.7965895249695494, "grad_norm": 1.922393560409546, "learning_rate": 0.00025447864327327903, "loss": 5.1955, "step": 297950 }, { "epoch": 1.7968910167508834, "grad_norm": 1.4018950462341309, "learning_rate": 0.0002544640412533806, "loss": 4.9752, "step": 298000 }, { "epoch": 1.7971925085322176, "grad_norm": 1.3149127960205078, "learning_rate": 0.0002544494373109706, "loss": 4.788, "step": 298050 }, { "epoch": 1.7974940003135513, "grad_norm": 1.2192645072937012, "learning_rate": 0.00025443483144631777, "loss": 4.8547, "step": 298100 }, { "epoch": 1.7977954920948855, "grad_norm": 1.265761375427246, "learning_rate": 0.0002544202236596909, "loss": 5.3073, "step": 298150 }, { "epoch": 1.7980969838762195, "grad_norm": 1.4467929601669312, "learning_rate": 0.0002544056139513588, "loss": 5.1745, "step": 298200 }, { "epoch": 1.7983984756575535, "grad_norm": 1.3531838655471802, "learning_rate": 0.0002543910023215904, "loss": 5.1853, "step": 298250 }, { "epoch": 1.7986999674388877, "grad_norm": 1.833417534828186, "learning_rate": 0.00025437638877065456, "loss": 5.0861, "step": 298300 }, { "epoch": 1.7990014592202217, "grad_norm": 1.3405888080596924, "learning_rate": 0.00025436177329882025, "loss": 5.1805, "step": 298350 }, { "epoch": 1.7993029510015557, "grad_norm": 1.0424466133117676, "learning_rate": 0.00025434715590635644, "loss": 4.9536, "step": 298400 }, { "epoch": 1.79960444278289, "grad_norm": 2.8740177154541016, "learning_rate": 0.00025433253659353216, "loss": 5.0797, "step": 298450 }, { "epoch": 1.7999059345642237, "grad_norm": 1.7851006984710693, "learning_rate": 0.0002543179153606164, "loss": 4.534, "step": 298500 }, { "epoch": 1.8002074263455579, "grad_norm": 1.9919253587722778, "learning_rate": 0.0002543032922078783, "loss": 5.0668, "step": 298550 }, { "epoch": 1.8005089181268918, "grad_norm": 1.401992917060852, "learning_rate": 0.00025428866713558693, "loss": 5.3538, "step": 298600 }, { "epoch": 1.8008104099082258, "grad_norm": 1.3752084970474243, "learning_rate": 0.00025427404014401146, "loss": 5.3067, "step": 298650 }, { "epoch": 1.80111190168956, "grad_norm": 0.9428870677947998, "learning_rate": 0.00025425941123342113, "loss": 5.195, "step": 298700 }, { "epoch": 1.801413393470894, "grad_norm": 2.1369457244873047, "learning_rate": 0.0002542447804040851, "loss": 4.83, "step": 298750 }, { "epoch": 1.801714885252228, "grad_norm": 1.4425822496414185, "learning_rate": 0.0002542301476562726, "loss": 4.9576, "step": 298800 }, { "epoch": 1.8020163770335622, "grad_norm": 1.2060034275054932, "learning_rate": 0.00025421551299025305, "loss": 5.1745, "step": 298850 }, { "epoch": 1.802317868814896, "grad_norm": 1.2641077041625977, "learning_rate": 0.0002542008764062957, "loss": 4.8981, "step": 298900 }, { "epoch": 1.8026193605962302, "grad_norm": 1.295078158378601, "learning_rate": 0.0002541862379046699, "loss": 5.1639, "step": 298950 }, { "epoch": 1.8029208523775642, "grad_norm": 1.7690569162368774, "learning_rate": 0.0002541715974856451, "loss": 5.3196, "step": 299000 }, { "epoch": 1.8032223441588981, "grad_norm": 0.9520177245140076, "learning_rate": 0.00025415695514949065, "loss": 4.9174, "step": 299050 }, { "epoch": 1.8035238359402324, "grad_norm": 2.755366563796997, "learning_rate": 0.00025414231089647616, "loss": 5.0806, "step": 299100 }, { "epoch": 1.8038253277215663, "grad_norm": 1.221346378326416, "learning_rate": 0.000254127664726871, "loss": 5.375, "step": 299150 }, { "epoch": 1.8041268195029003, "grad_norm": 1.4629987478256226, "learning_rate": 0.00025411301664094474, "loss": 5.0066, "step": 299200 }, { "epoch": 1.8044283112842345, "grad_norm": 1.2691092491149902, "learning_rate": 0.0002540983666389671, "loss": 5.2536, "step": 299250 }, { "epoch": 1.8047298030655683, "grad_norm": 1.9106460809707642, "learning_rate": 0.00025408371472120746, "loss": 4.8593, "step": 299300 }, { "epoch": 1.8050312948469025, "grad_norm": 1.03182053565979, "learning_rate": 0.00025406906088793564, "loss": 5.0497, "step": 299350 }, { "epoch": 1.8053327866282365, "grad_norm": 0.6063446402549744, "learning_rate": 0.0002540544051394213, "loss": 5.1745, "step": 299400 }, { "epoch": 1.8056342784095705, "grad_norm": 1.5768970251083374, "learning_rate": 0.0002540397474759341, "loss": 4.9583, "step": 299450 }, { "epoch": 1.8059357701909047, "grad_norm": 1.8390758037567139, "learning_rate": 0.0002540250878977438, "loss": 4.7536, "step": 299500 }, { "epoch": 1.8062372619722387, "grad_norm": 1.5862765312194824, "learning_rate": 0.0002540104264051202, "loss": 5.1386, "step": 299550 }, { "epoch": 1.8065387537535726, "grad_norm": 0.7131795287132263, "learning_rate": 0.0002539957629983331, "loss": 4.9699, "step": 299600 }, { "epoch": 1.8068402455349069, "grad_norm": 1.2496669292449951, "learning_rate": 0.0002539810976776525, "loss": 5.0481, "step": 299650 }, { "epoch": 1.8071417373162406, "grad_norm": 1.8556360006332397, "learning_rate": 0.0002539664304433481, "loss": 4.8069, "step": 299700 }, { "epoch": 1.8074432290975748, "grad_norm": 1.7822988033294678, "learning_rate": 0.00025395176129568994, "loss": 5.3296, "step": 299750 }, { "epoch": 1.8077447208789088, "grad_norm": 0.9227627515792847, "learning_rate": 0.000253937090234948, "loss": 5.2746, "step": 299800 }, { "epoch": 1.8080462126602428, "grad_norm": 1.1587529182434082, "learning_rate": 0.0002539224172613922, "loss": 5.065, "step": 299850 }, { "epoch": 1.808347704441577, "grad_norm": 0.9879549145698547, "learning_rate": 0.00025390774237529263, "loss": 5.0707, "step": 299900 }, { "epoch": 1.808649196222911, "grad_norm": 1.372429370880127, "learning_rate": 0.0002538930655769193, "loss": 4.9507, "step": 299950 }, { "epoch": 1.808950688004245, "grad_norm": 1.3229057788848877, "learning_rate": 0.00025387838686654246, "loss": 5.2379, "step": 300000 }, { "epoch": 1.808950688004245, "eval_loss": 5.540033340454102, "eval_runtime": 39.2876, "eval_samples_per_second": 13.032, "eval_steps_per_second": 6.516, "eval_tts_loss": 7.215152503746963, "step": 300000 }, { "epoch": 1.8092521797855792, "grad_norm": 2.2696800231933594, "learning_rate": 0.0002538637062444321, "loss": 4.5877, "step": 300050 }, { "epoch": 1.809553671566913, "grad_norm": 1.9000815153121948, "learning_rate": 0.0002538490237108584, "loss": 4.8393, "step": 300100 }, { "epoch": 1.8098551633482471, "grad_norm": 1.43900465965271, "learning_rate": 0.00025383433926609166, "loss": 4.9844, "step": 300150 }, { "epoch": 1.8101566551295811, "grad_norm": 2.5236504077911377, "learning_rate": 0.0002538196529104021, "loss": 5.2524, "step": 300200 }, { "epoch": 1.8104581469109151, "grad_norm": 1.317436695098877, "learning_rate": 0.00025380496464406, "loss": 4.8835, "step": 300250 }, { "epoch": 1.8107596386922493, "grad_norm": 1.094543218612671, "learning_rate": 0.00025379027446733557, "loss": 5.0231, "step": 300300 }, { "epoch": 1.8110611304735833, "grad_norm": 1.8807806968688965, "learning_rate": 0.0002537755823804993, "loss": 5.3526, "step": 300350 }, { "epoch": 1.8113626222549173, "grad_norm": 2.0700931549072266, "learning_rate": 0.0002537608883838216, "loss": 5.5671, "step": 300400 }, { "epoch": 1.8116641140362515, "grad_norm": 1.0530403852462769, "learning_rate": 0.0002537461924775728, "loss": 5.1864, "step": 300450 }, { "epoch": 1.8119656058175853, "grad_norm": 1.5361460447311401, "learning_rate": 0.00025373149466202335, "loss": 4.9586, "step": 300500 }, { "epoch": 1.8122670975989195, "grad_norm": 1.6567363739013672, "learning_rate": 0.0002537167949374438, "loss": 5.2081, "step": 300550 }, { "epoch": 1.8125685893802534, "grad_norm": 2.231956720352173, "learning_rate": 0.0002537020933041046, "loss": 4.6117, "step": 300600 }, { "epoch": 1.8128700811615874, "grad_norm": 1.5761771202087402, "learning_rate": 0.0002536873897622764, "loss": 4.949, "step": 300650 }, { "epoch": 1.8131715729429216, "grad_norm": 1.9028334617614746, "learning_rate": 0.0002536726843122298, "loss": 5.3721, "step": 300700 }, { "epoch": 1.8134730647242556, "grad_norm": 1.3932063579559326, "learning_rate": 0.0002536579769542353, "loss": 4.9269, "step": 300750 }, { "epoch": 1.8137745565055896, "grad_norm": 1.7783784866333008, "learning_rate": 0.0002536432676885637, "loss": 4.6457, "step": 300800 }, { "epoch": 1.8140760482869238, "grad_norm": 1.3486148118972778, "learning_rate": 0.0002536285565154857, "loss": 5.1466, "step": 300850 }, { "epoch": 1.8143775400682576, "grad_norm": 1.2839196920394897, "learning_rate": 0.000253613843435272, "loss": 5.3653, "step": 300900 }, { "epoch": 1.8146790318495918, "grad_norm": 0.4238676428794861, "learning_rate": 0.00025359912844819335, "loss": 5.0416, "step": 300950 }, { "epoch": 1.8149805236309258, "grad_norm": 1.2929370403289795, "learning_rate": 0.0002535844115545206, "loss": 4.7779, "step": 301000 }, { "epoch": 1.8152820154122598, "grad_norm": 1.780387043952942, "learning_rate": 0.0002535696927545246, "loss": 4.9794, "step": 301050 }, { "epoch": 1.815583507193594, "grad_norm": 1.3449969291687012, "learning_rate": 0.0002535549720484761, "loss": 5.2136, "step": 301100 }, { "epoch": 1.815884998974928, "grad_norm": 1.334861397743225, "learning_rate": 0.0002535402494366462, "loss": 4.8778, "step": 301150 }, { "epoch": 1.816186490756262, "grad_norm": 1.2097327709197998, "learning_rate": 0.0002535255249193058, "loss": 4.7676, "step": 301200 }, { "epoch": 1.8164879825375961, "grad_norm": 1.8173567056655884, "learning_rate": 0.00025351079849672586, "loss": 4.9309, "step": 301250 }, { "epoch": 1.8167894743189301, "grad_norm": 1.678159236907959, "learning_rate": 0.00025349607016917735, "loss": 4.634, "step": 301300 }, { "epoch": 1.817090966100264, "grad_norm": 1.2696317434310913, "learning_rate": 0.0002534813399369313, "loss": 5.0331, "step": 301350 }, { "epoch": 1.8173924578815983, "grad_norm": 1.5176719427108765, "learning_rate": 0.00025346660780025895, "loss": 4.863, "step": 301400 }, { "epoch": 1.817693949662932, "grad_norm": 2.391817569732666, "learning_rate": 0.0002534518737594313, "loss": 4.9162, "step": 301450 }, { "epoch": 1.8179954414442663, "grad_norm": 1.952167272567749, "learning_rate": 0.0002534371378147196, "loss": 4.8677, "step": 301500 }, { "epoch": 1.8182969332256003, "grad_norm": 1.4075065851211548, "learning_rate": 0.00025342239996639497, "loss": 5.0865, "step": 301550 }, { "epoch": 1.8185984250069342, "grad_norm": 0.6303558945655823, "learning_rate": 0.0002534076602147286, "loss": 5.308, "step": 301600 }, { "epoch": 1.8188999167882685, "grad_norm": 2.0393028259277344, "learning_rate": 0.0002533929185599919, "loss": 5.426, "step": 301650 }, { "epoch": 1.8192014085696024, "grad_norm": 1.442688226699829, "learning_rate": 0.000253378175002456, "loss": 4.9622, "step": 301700 }, { "epoch": 1.8195029003509364, "grad_norm": 1.8697270154953003, "learning_rate": 0.0002533634295423924, "loss": 5.0448, "step": 301750 }, { "epoch": 1.8198043921322706, "grad_norm": 1.1509990692138672, "learning_rate": 0.00025334868218007233, "loss": 4.7001, "step": 301800 }, { "epoch": 1.8201058839136044, "grad_norm": 1.275272011756897, "learning_rate": 0.0002533339329157673, "loss": 5.2661, "step": 301850 }, { "epoch": 1.8204073756949386, "grad_norm": 0.675865113735199, "learning_rate": 0.0002533191817497487, "loss": 4.9436, "step": 301900 }, { "epoch": 1.8207088674762726, "grad_norm": 1.0209640264511108, "learning_rate": 0.00025330442868228795, "loss": 5.261, "step": 301950 }, { "epoch": 1.8210103592576066, "grad_norm": 1.425218939781189, "learning_rate": 0.00025328967371365666, "loss": 5.0888, "step": 302000 }, { "epoch": 1.8213118510389408, "grad_norm": 2.2712574005126953, "learning_rate": 0.0002532749168441263, "loss": 5.2755, "step": 302050 }, { "epoch": 1.8216133428202748, "grad_norm": 3.1147046089172363, "learning_rate": 0.00025326015807396846, "loss": 5.1103, "step": 302100 }, { "epoch": 1.8219148346016087, "grad_norm": 1.2454761266708374, "learning_rate": 0.0002532453974034548, "loss": 5.3224, "step": 302150 }, { "epoch": 1.822216326382943, "grad_norm": 1.7371740341186523, "learning_rate": 0.0002532306348328569, "loss": 4.5643, "step": 302200 }, { "epoch": 1.8225178181642767, "grad_norm": 2.030564546585083, "learning_rate": 0.00025321587036244655, "loss": 4.9998, "step": 302250 }, { "epoch": 1.822819309945611, "grad_norm": 1.4609583616256714, "learning_rate": 0.0002532011039924954, "loss": 5.2799, "step": 302300 }, { "epoch": 1.823120801726945, "grad_norm": 2.0576796531677246, "learning_rate": 0.00025318633572327516, "loss": 4.5473, "step": 302350 }, { "epoch": 1.823422293508279, "grad_norm": 0.6849621534347534, "learning_rate": 0.00025317156555505765, "loss": 4.5651, "step": 302400 }, { "epoch": 1.823723785289613, "grad_norm": 1.624337077140808, "learning_rate": 0.0002531567934881148, "loss": 5.0135, "step": 302450 }, { "epoch": 1.824025277070947, "grad_norm": 1.4768038988113403, "learning_rate": 0.00025314201952271833, "loss": 5.2768, "step": 302500 }, { "epoch": 1.824326768852281, "grad_norm": 1.6016345024108887, "learning_rate": 0.00025312724365914014, "loss": 5.0067, "step": 302550 }, { "epoch": 1.8246282606336153, "grad_norm": 1.2211253643035889, "learning_rate": 0.00025311246589765227, "loss": 5.0059, "step": 302600 }, { "epoch": 1.824929752414949, "grad_norm": 2.263002872467041, "learning_rate": 0.00025309768623852654, "loss": 5.3171, "step": 302650 }, { "epoch": 1.8252312441962832, "grad_norm": 2.2348055839538574, "learning_rate": 0.0002530829046820351, "loss": 5.0147, "step": 302700 }, { "epoch": 1.8255327359776172, "grad_norm": 1.3479032516479492, "learning_rate": 0.00025306812122844987, "loss": 5.3725, "step": 302750 }, { "epoch": 1.8258342277589512, "grad_norm": 0.9551581144332886, "learning_rate": 0.0002530533358780429, "loss": 4.9215, "step": 302800 }, { "epoch": 1.8261357195402854, "grad_norm": 1.251035451889038, "learning_rate": 0.00025303854863108644, "loss": 5.4664, "step": 302850 }, { "epoch": 1.8264372113216194, "grad_norm": 0.3524314761161804, "learning_rate": 0.0002530237594878525, "loss": 5.1192, "step": 302900 }, { "epoch": 1.8267387031029534, "grad_norm": 2.099151134490967, "learning_rate": 0.00025300896844861323, "loss": 5.1047, "step": 302950 }, { "epoch": 1.8270401948842876, "grad_norm": 2.3623034954071045, "learning_rate": 0.000252994175513641, "loss": 5.1724, "step": 303000 }, { "epoch": 1.8270401948842876, "eval_loss": 5.520170211791992, "eval_runtime": 38.9634, "eval_samples_per_second": 13.141, "eval_steps_per_second": 6.57, "eval_tts_loss": 7.340573048752303, "step": 303000 }, { "epoch": 1.8273416866656214, "grad_norm": 1.2670478820800781, "learning_rate": 0.0002529793806832079, "loss": 5.2263, "step": 303050 }, { "epoch": 1.8276431784469556, "grad_norm": 1.3454958200454712, "learning_rate": 0.0002529645839575863, "loss": 5.2983, "step": 303100 }, { "epoch": 1.8279446702282895, "grad_norm": 1.957779884338379, "learning_rate": 0.0002529497853370484, "loss": 4.8473, "step": 303150 }, { "epoch": 1.8282461620096235, "grad_norm": 1.18436861038208, "learning_rate": 0.00025293498482186665, "loss": 4.9393, "step": 303200 }, { "epoch": 1.8285476537909577, "grad_norm": 0.7841783761978149, "learning_rate": 0.0002529201824123134, "loss": 4.5644, "step": 303250 }, { "epoch": 1.8288491455722917, "grad_norm": 1.377615213394165, "learning_rate": 0.00025290537810866107, "loss": 4.8974, "step": 303300 }, { "epoch": 1.8291506373536257, "grad_norm": 1.7306063175201416, "learning_rate": 0.0002528905719111821, "loss": 5.217, "step": 303350 }, { "epoch": 1.82945212913496, "grad_norm": 1.2373783588409424, "learning_rate": 0.000252875763820149, "loss": 5.0122, "step": 303400 }, { "epoch": 1.8297536209162937, "grad_norm": 1.7725293636322021, "learning_rate": 0.0002528609538358343, "loss": 4.9603, "step": 303450 }, { "epoch": 1.8300551126976279, "grad_norm": 2.4681830406188965, "learning_rate": 0.0002528461419585105, "loss": 5.0999, "step": 303500 }, { "epoch": 1.8303566044789619, "grad_norm": 1.5507928133010864, "learning_rate": 0.00025283132818845023, "loss": 4.9764, "step": 303550 }, { "epoch": 1.8306580962602959, "grad_norm": 1.1955424547195435, "learning_rate": 0.0002528165125259261, "loss": 4.7955, "step": 303600 }, { "epoch": 1.83095958804163, "grad_norm": 1.6246832609176636, "learning_rate": 0.0002528016949712108, "loss": 5.4577, "step": 303650 }, { "epoch": 1.831261079822964, "grad_norm": 1.5430856943130493, "learning_rate": 0.00025278687552457706, "loss": 5.2306, "step": 303700 }, { "epoch": 1.831562571604298, "grad_norm": 1.314378023147583, "learning_rate": 0.0002527720541862975, "loss": 4.6855, "step": 303750 }, { "epoch": 1.8318640633856322, "grad_norm": 1.146262526512146, "learning_rate": 0.00025275723095664497, "loss": 5.2074, "step": 303800 }, { "epoch": 1.832165555166966, "grad_norm": 2.221788167953491, "learning_rate": 0.0002527424058358922, "loss": 5.3628, "step": 303850 }, { "epoch": 1.8324670469483002, "grad_norm": 1.1420329809188843, "learning_rate": 0.00025272757882431217, "loss": 5.3479, "step": 303900 }, { "epoch": 1.8327685387296342, "grad_norm": 1.8266602754592896, "learning_rate": 0.00025271274992217757, "loss": 5.2436, "step": 303950 }, { "epoch": 1.8330700305109682, "grad_norm": 1.614977478981018, "learning_rate": 0.0002526979191297614, "loss": 5.4324, "step": 304000 }, { "epoch": 1.8333715222923024, "grad_norm": 1.6129441261291504, "learning_rate": 0.0002526830864473366, "loss": 5.4982, "step": 304050 }, { "epoch": 1.8336730140736364, "grad_norm": 0.8706692457199097, "learning_rate": 0.00025266825187517615, "loss": 4.9848, "step": 304100 }, { "epoch": 1.8339745058549703, "grad_norm": 1.4661104679107666, "learning_rate": 0.000252653415413553, "loss": 4.9295, "step": 304150 }, { "epoch": 1.8342759976363046, "grad_norm": 1.7571274042129517, "learning_rate": 0.0002526385770627402, "loss": 5.2057, "step": 304200 }, { "epoch": 1.8345774894176383, "grad_norm": 2.8402750492095947, "learning_rate": 0.00025262373682301097, "loss": 5.0225, "step": 304250 }, { "epoch": 1.8348789811989725, "grad_norm": 3.5413930416107178, "learning_rate": 0.00025260889469463823, "loss": 5.0132, "step": 304300 }, { "epoch": 1.8351804729803065, "grad_norm": 1.183648705482483, "learning_rate": 0.00025259405067789523, "loss": 4.6581, "step": 304350 }, { "epoch": 1.8354819647616405, "grad_norm": 3.042240619659424, "learning_rate": 0.0002525792047730552, "loss": 5.0827, "step": 304400 }, { "epoch": 1.8357834565429747, "grad_norm": 1.1799501180648804, "learning_rate": 0.0002525643569803912, "loss": 5.1506, "step": 304450 }, { "epoch": 1.8360849483243087, "grad_norm": 1.2033528089523315, "learning_rate": 0.0002525495073001766, "loss": 5.1168, "step": 304500 }, { "epoch": 1.8363864401056427, "grad_norm": 1.500463843345642, "learning_rate": 0.00025253465573268474, "loss": 5.3031, "step": 304550 }, { "epoch": 1.8366879318869769, "grad_norm": 1.2809149026870728, "learning_rate": 0.0002525198022781888, "loss": 5.2624, "step": 304600 }, { "epoch": 1.8369894236683106, "grad_norm": 1.7034960985183716, "learning_rate": 0.00025250494693696224, "loss": 5.3066, "step": 304650 }, { "epoch": 1.8372909154496448, "grad_norm": 1.3491075038909912, "learning_rate": 0.0002524900897092784, "loss": 5.2271, "step": 304700 }, { "epoch": 1.8375924072309788, "grad_norm": 1.320316195487976, "learning_rate": 0.0002524752305954106, "loss": 5.2665, "step": 304750 }, { "epoch": 1.8378938990123128, "grad_norm": 1.2392725944519043, "learning_rate": 0.00025246036959563256, "loss": 4.9195, "step": 304800 }, { "epoch": 1.838195390793647, "grad_norm": 1.4819374084472656, "learning_rate": 0.0002524455067102176, "loss": 5.3887, "step": 304850 }, { "epoch": 1.838496882574981, "grad_norm": 1.0964146852493286, "learning_rate": 0.0002524306419394393, "loss": 4.9209, "step": 304900 }, { "epoch": 1.838798374356315, "grad_norm": 1.2966054677963257, "learning_rate": 0.0002524157752835712, "loss": 4.9579, "step": 304950 }, { "epoch": 1.8390998661376492, "grad_norm": 3.1694226264953613, "learning_rate": 0.00025240090674288695, "loss": 4.9413, "step": 305000 }, { "epoch": 1.8394013579189832, "grad_norm": 1.4138145446777344, "learning_rate": 0.0002523860363176601, "loss": 5.177, "step": 305050 }, { "epoch": 1.8397028497003172, "grad_norm": 1.8188387155532837, "learning_rate": 0.00025237116400816435, "loss": 4.9491, "step": 305100 }, { "epoch": 1.8400043414816514, "grad_norm": 1.7948145866394043, "learning_rate": 0.0002523562898146734, "loss": 4.8219, "step": 305150 }, { "epoch": 1.8403058332629851, "grad_norm": 1.358388900756836, "learning_rate": 0.00025234141373746107, "loss": 5.2071, "step": 305200 }, { "epoch": 1.8406073250443193, "grad_norm": 1.419307827949524, "learning_rate": 0.00025232653577680105, "loss": 4.9699, "step": 305250 }, { "epoch": 1.8409088168256533, "grad_norm": 2.2339720726013184, "learning_rate": 0.0002523116559329672, "loss": 5.5522, "step": 305300 }, { "epoch": 1.8412103086069873, "grad_norm": 1.2025175094604492, "learning_rate": 0.00025229677420623325, "loss": 5.1746, "step": 305350 }, { "epoch": 1.8415118003883215, "grad_norm": 1.6530036926269531, "learning_rate": 0.00025228189059687325, "loss": 4.9126, "step": 305400 }, { "epoch": 1.8418132921696555, "grad_norm": 1.2555586099624634, "learning_rate": 0.0002522670051051609, "loss": 5.4376, "step": 305450 }, { "epoch": 1.8421147839509895, "grad_norm": 1.516679048538208, "learning_rate": 0.0002522521177313704, "loss": 4.8976, "step": 305500 }, { "epoch": 1.8424162757323237, "grad_norm": 1.9734967947006226, "learning_rate": 0.0002522372284757755, "loss": 4.6799, "step": 305550 }, { "epoch": 1.8427177675136575, "grad_norm": 1.603010654449463, "learning_rate": 0.0002522223373386503, "loss": 5.141, "step": 305600 }, { "epoch": 1.8430192592949917, "grad_norm": 1.3621315956115723, "learning_rate": 0.0002522074443202689, "loss": 4.8076, "step": 305650 }, { "epoch": 1.8433207510763256, "grad_norm": 1.096168041229248, "learning_rate": 0.00025219254942090535, "loss": 5.0104, "step": 305700 }, { "epoch": 1.8436222428576596, "grad_norm": 2.437946081161499, "learning_rate": 0.00025217765264083376, "loss": 5.3016, "step": 305750 }, { "epoch": 1.8439237346389938, "grad_norm": 1.421184778213501, "learning_rate": 0.00025216275398032826, "loss": 4.7386, "step": 305800 }, { "epoch": 1.8442252264203278, "grad_norm": 0.8313750624656677, "learning_rate": 0.0002521478534396631, "loss": 4.9096, "step": 305850 }, { "epoch": 1.8445267182016618, "grad_norm": 1.3484982252120972, "learning_rate": 0.00025213295101911244, "loss": 4.9637, "step": 305900 }, { "epoch": 1.844828209982996, "grad_norm": 0.7223657965660095, "learning_rate": 0.0002521180467189506, "loss": 4.7706, "step": 305950 }, { "epoch": 1.8451297017643298, "grad_norm": 1.364546537399292, "learning_rate": 0.00025210314053945187, "loss": 4.8929, "step": 306000 }, { "epoch": 1.8451297017643298, "eval_loss": 5.534787654876709, "eval_runtime": 39.1683, "eval_samples_per_second": 13.072, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.352124407336641, "step": 306000 }, { "epoch": 1.845431193545664, "grad_norm": 0.9262778759002686, "learning_rate": 0.00025208823248089047, "loss": 4.8877, "step": 306050 }, { "epoch": 1.845732685326998, "grad_norm": 0.5302439332008362, "learning_rate": 0.0002520733225435409, "loss": 4.9394, "step": 306100 }, { "epoch": 1.846034177108332, "grad_norm": 1.9041666984558105, "learning_rate": 0.00025205841072767745, "loss": 5.2075, "step": 306150 }, { "epoch": 1.8463356688896662, "grad_norm": 1.6671175956726074, "learning_rate": 0.0002520434970335746, "loss": 4.9778, "step": 306200 }, { "epoch": 1.8466371606710001, "grad_norm": 2.196503162384033, "learning_rate": 0.0002520285814615068, "loss": 5.0918, "step": 306250 }, { "epoch": 1.8469386524523341, "grad_norm": 1.5273998975753784, "learning_rate": 0.0002520136640117486, "loss": 5.2369, "step": 306300 }, { "epoch": 1.8472401442336683, "grad_norm": 1.1294511556625366, "learning_rate": 0.0002519987446845745, "loss": 5.192, "step": 306350 }, { "epoch": 1.847541636015002, "grad_norm": 1.3109780550003052, "learning_rate": 0.0002519838234802591, "loss": 5.0363, "step": 306400 }, { "epoch": 1.8478431277963363, "grad_norm": 3.048821449279785, "learning_rate": 0.00025196890039907684, "loss": 4.6487, "step": 306450 }, { "epoch": 1.8481446195776703, "grad_norm": 0.8250425457954407, "learning_rate": 0.00025195397544130256, "loss": 4.986, "step": 306500 }, { "epoch": 1.8484461113590043, "grad_norm": 1.4401577711105347, "learning_rate": 0.0002519390486072109, "loss": 4.8716, "step": 306550 }, { "epoch": 1.8487476031403385, "grad_norm": 1.945588231086731, "learning_rate": 0.0002519241198970764, "loss": 5.1098, "step": 306600 }, { "epoch": 1.8490490949216725, "grad_norm": 1.7270102500915527, "learning_rate": 0.0002519091893111741, "loss": 4.9294, "step": 306650 }, { "epoch": 1.8493505867030064, "grad_norm": 1.1385233402252197, "learning_rate": 0.0002518942568497785, "loss": 4.9206, "step": 306700 }, { "epoch": 1.8496520784843407, "grad_norm": 2.240539789199829, "learning_rate": 0.0002518793225131645, "loss": 4.9982, "step": 306750 }, { "epoch": 1.8499535702656744, "grad_norm": 1.0906099081039429, "learning_rate": 0.000251864386301607, "loss": 4.9448, "step": 306800 }, { "epoch": 1.8502550620470086, "grad_norm": 1.747395634651184, "learning_rate": 0.00025184944821538083, "loss": 5.3227, "step": 306850 }, { "epoch": 1.8505565538283426, "grad_norm": 1.330120325088501, "learning_rate": 0.0002518345082547609, "loss": 5.485, "step": 306900 }, { "epoch": 1.8508580456096766, "grad_norm": 1.2158167362213135, "learning_rate": 0.0002518195664200222, "loss": 5.1412, "step": 306950 }, { "epoch": 1.8511595373910108, "grad_norm": 2.08608078956604, "learning_rate": 0.00025180462271143965, "loss": 4.8383, "step": 307000 }, { "epoch": 1.8514610291723448, "grad_norm": 0.7088283896446228, "learning_rate": 0.0002517896771292883, "loss": 4.8979, "step": 307050 }, { "epoch": 1.8517625209536788, "grad_norm": 2.1641125679016113, "learning_rate": 0.0002517747296738432, "loss": 5.2456, "step": 307100 }, { "epoch": 1.852064012735013, "grad_norm": 1.751407265663147, "learning_rate": 0.00025175978034537946, "loss": 5.3615, "step": 307150 }, { "epoch": 1.8523655045163467, "grad_norm": 1.7794623374938965, "learning_rate": 0.0002517448291441722, "loss": 5.3074, "step": 307200 }, { "epoch": 1.852666996297681, "grad_norm": 1.4636337757110596, "learning_rate": 0.0002517298760704965, "loss": 5.0862, "step": 307250 }, { "epoch": 1.852968488079015, "grad_norm": 1.2482833862304688, "learning_rate": 0.0002517149211246277, "loss": 5.214, "step": 307300 }, { "epoch": 1.853269979860349, "grad_norm": 2.1864914894104004, "learning_rate": 0.00025169996430684084, "loss": 5.2474, "step": 307350 }, { "epoch": 1.8535714716416831, "grad_norm": 2.3544604778289795, "learning_rate": 0.00025168500561741135, "loss": 4.8474, "step": 307400 }, { "epoch": 1.853872963423017, "grad_norm": 1.0229319334030151, "learning_rate": 0.00025167004505661437, "loss": 5.2385, "step": 307450 }, { "epoch": 1.854174455204351, "grad_norm": 2.545151948928833, "learning_rate": 0.00025165508262472536, "loss": 5.0243, "step": 307500 }, { "epoch": 1.8544759469856853, "grad_norm": 1.5481868982315063, "learning_rate": 0.0002516401183220196, "loss": 5.1162, "step": 307550 }, { "epoch": 1.854777438767019, "grad_norm": 2.0172228813171387, "learning_rate": 0.0002516251521487725, "loss": 5.0564, "step": 307600 }, { "epoch": 1.8550789305483533, "grad_norm": 1.3533451557159424, "learning_rate": 0.0002516101841052595, "loss": 5.4642, "step": 307650 }, { "epoch": 1.8553804223296873, "grad_norm": 0.7763773798942566, "learning_rate": 0.0002515952141917561, "loss": 4.7674, "step": 307700 }, { "epoch": 1.8556819141110212, "grad_norm": 0.9272885322570801, "learning_rate": 0.0002515802424085377, "loss": 4.965, "step": 307750 }, { "epoch": 1.8559834058923554, "grad_norm": 2.0386295318603516, "learning_rate": 0.00025156526875587994, "loss": 4.8575, "step": 307800 }, { "epoch": 1.8562848976736894, "grad_norm": 1.3861122131347656, "learning_rate": 0.00025155029323405836, "loss": 4.9822, "step": 307850 }, { "epoch": 1.8565863894550234, "grad_norm": 1.6990946531295776, "learning_rate": 0.00025153531584334854, "loss": 5.3993, "step": 307900 }, { "epoch": 1.8568878812363576, "grad_norm": 1.6461795568466187, "learning_rate": 0.00025152033658402614, "loss": 4.8221, "step": 307950 }, { "epoch": 1.8571893730176914, "grad_norm": 1.3680918216705322, "learning_rate": 0.0002515053554563668, "loss": 4.856, "step": 308000 }, { "epoch": 1.8574908647990256, "grad_norm": 1.2620491981506348, "learning_rate": 0.0002514903724606463, "loss": 4.9066, "step": 308050 }, { "epoch": 1.8577923565803596, "grad_norm": 1.7097196578979492, "learning_rate": 0.0002514753875971403, "loss": 4.8084, "step": 308100 }, { "epoch": 1.8580938483616936, "grad_norm": 1.7519619464874268, "learning_rate": 0.0002514604008661246, "loss": 5.1646, "step": 308150 }, { "epoch": 1.8583953401430278, "grad_norm": 1.5122672319412231, "learning_rate": 0.000251445412267875, "loss": 5.0611, "step": 308200 }, { "epoch": 1.8586968319243617, "grad_norm": 1.9605188369750977, "learning_rate": 0.0002514304218026674, "loss": 4.7769, "step": 308250 }, { "epoch": 1.8589983237056957, "grad_norm": 1.9829039573669434, "learning_rate": 0.0002514154294707776, "loss": 4.9813, "step": 308300 }, { "epoch": 1.85929981548703, "grad_norm": 2.1928839683532715, "learning_rate": 0.0002514004352724815, "loss": 4.7267, "step": 308350 }, { "epoch": 1.8596013072683637, "grad_norm": 1.0619195699691772, "learning_rate": 0.0002513854392080552, "loss": 5.3456, "step": 308400 }, { "epoch": 1.859902799049698, "grad_norm": 0.7588275671005249, "learning_rate": 0.0002513704412777745, "loss": 4.7777, "step": 308450 }, { "epoch": 1.860204290831032, "grad_norm": 2.2773852348327637, "learning_rate": 0.0002513554414819155, "loss": 4.9767, "step": 308500 }, { "epoch": 1.8605057826123659, "grad_norm": 1.680938482284546, "learning_rate": 0.00025134043982075427, "loss": 5.3593, "step": 308550 }, { "epoch": 1.8608072743937, "grad_norm": 1.8236536979675293, "learning_rate": 0.00025132543629456693, "loss": 5.0808, "step": 308600 }, { "epoch": 1.861108766175034, "grad_norm": 2.356583595275879, "learning_rate": 0.00025131043090362943, "loss": 4.9961, "step": 308650 }, { "epoch": 1.861410257956368, "grad_norm": 0.5836613774299622, "learning_rate": 0.000251295423648218, "loss": 4.983, "step": 308700 }, { "epoch": 1.8617117497377023, "grad_norm": 1.4610369205474854, "learning_rate": 0.00025128041452860897, "loss": 5.2675, "step": 308750 }, { "epoch": 1.8620132415190362, "grad_norm": 2.089723587036133, "learning_rate": 0.0002512654035450784, "loss": 5.222, "step": 308800 }, { "epoch": 1.8623147333003702, "grad_norm": 1.2679916620254517, "learning_rate": 0.0002512503906979026, "loss": 4.8478, "step": 308850 }, { "epoch": 1.8626162250817044, "grad_norm": 0.831847608089447, "learning_rate": 0.00025123537598735783, "loss": 5.241, "step": 308900 }, { "epoch": 1.8629177168630382, "grad_norm": 0.5778084397315979, "learning_rate": 0.0002512203594137204, "loss": 5.0264, "step": 308950 }, { "epoch": 1.8632192086443724, "grad_norm": 1.178700566291809, "learning_rate": 0.00025120534097726677, "loss": 4.9308, "step": 309000 }, { "epoch": 1.8632192086443724, "eval_loss": 5.5182390213012695, "eval_runtime": 38.929, "eval_samples_per_second": 13.152, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.340053132106513, "step": 309000 }, { "epoch": 1.8635207004257064, "grad_norm": 1.4880664348602295, "learning_rate": 0.0002511903206782732, "loss": 4.8446, "step": 309050 }, { "epoch": 1.8638221922070404, "grad_norm": 1.3613202571868896, "learning_rate": 0.00025117529851701625, "loss": 4.9543, "step": 309100 }, { "epoch": 1.8641236839883746, "grad_norm": 2.002089262008667, "learning_rate": 0.0002511602744937723, "loss": 4.7462, "step": 309150 }, { "epoch": 1.8644251757697086, "grad_norm": 1.392925500869751, "learning_rate": 0.0002511452486088178, "loss": 5.225, "step": 309200 }, { "epoch": 1.8647266675510425, "grad_norm": 1.9864152669906616, "learning_rate": 0.00025113022086242944, "loss": 5.3096, "step": 309250 }, { "epoch": 1.8650281593323768, "grad_norm": 0.6058738231658936, "learning_rate": 0.0002511151912548836, "loss": 5.0545, "step": 309300 }, { "epoch": 1.8653296511137105, "grad_norm": 1.6291979551315308, "learning_rate": 0.00025110015978645705, "loss": 5.2829, "step": 309350 }, { "epoch": 1.8656311428950447, "grad_norm": 1.6107017993927002, "learning_rate": 0.00025108512645742625, "loss": 5.0024, "step": 309400 }, { "epoch": 1.8659326346763787, "grad_norm": 1.2201545238494873, "learning_rate": 0.00025107009126806797, "loss": 5.084, "step": 309450 }, { "epoch": 1.8662341264577127, "grad_norm": 0.7911856174468994, "learning_rate": 0.0002510550542186589, "loss": 5.0367, "step": 309500 }, { "epoch": 1.866535618239047, "grad_norm": 1.8618459701538086, "learning_rate": 0.0002510400153094758, "loss": 5.0395, "step": 309550 }, { "epoch": 1.8668371100203809, "grad_norm": 1.033677577972412, "learning_rate": 0.0002510249745407954, "loss": 5.1252, "step": 309600 }, { "epoch": 1.8671386018017149, "grad_norm": 1.18272864818573, "learning_rate": 0.0002510099319128945, "loss": 5.2281, "step": 309650 }, { "epoch": 1.867440093583049, "grad_norm": 0.813994824886322, "learning_rate": 0.00025099488742605, "loss": 4.7108, "step": 309700 }, { "epoch": 1.8677415853643828, "grad_norm": 0.916269838809967, "learning_rate": 0.0002509798410805387, "loss": 5.1494, "step": 309750 }, { "epoch": 1.868043077145717, "grad_norm": 0.38804909586906433, "learning_rate": 0.0002509647928766375, "loss": 5.3416, "step": 309800 }, { "epoch": 1.868344568927051, "grad_norm": 1.6745452880859375, "learning_rate": 0.0002509497428146234, "loss": 5.1728, "step": 309850 }, { "epoch": 1.868646060708385, "grad_norm": 2.1365480422973633, "learning_rate": 0.0002509346908947733, "loss": 4.7958, "step": 309900 }, { "epoch": 1.8689475524897192, "grad_norm": 1.3894968032836914, "learning_rate": 0.0002509196371173643, "loss": 5.4208, "step": 309950 }, { "epoch": 1.8692490442710532, "grad_norm": 1.2760074138641357, "learning_rate": 0.0002509045814826734, "loss": 4.9147, "step": 310000 }, { "epoch": 1.8695505360523872, "grad_norm": 2.8511385917663574, "learning_rate": 0.0002508895239909777, "loss": 5.0808, "step": 310050 }, { "epoch": 1.8698520278337214, "grad_norm": 1.1166627407073975, "learning_rate": 0.0002508744646425542, "loss": 4.8101, "step": 310100 }, { "epoch": 1.8701535196150552, "grad_norm": 2.0453102588653564, "learning_rate": 0.00025085940343768016, "loss": 5.4925, "step": 310150 }, { "epoch": 1.8704550113963894, "grad_norm": 1.602921724319458, "learning_rate": 0.0002508443403766328, "loss": 5.0964, "step": 310200 }, { "epoch": 1.8707565031777234, "grad_norm": 1.7455590963363647, "learning_rate": 0.00025082927545968923, "loss": 5.6935, "step": 310250 }, { "epoch": 1.8710579949590573, "grad_norm": 0.7515433430671692, "learning_rate": 0.00025081420868712674, "loss": 4.8834, "step": 310300 }, { "epoch": 1.8713594867403915, "grad_norm": 2.01792311668396, "learning_rate": 0.00025079914005922264, "loss": 4.5708, "step": 310350 }, { "epoch": 1.8716609785217255, "grad_norm": 0.703222930431366, "learning_rate": 0.00025078406957625414, "loss": 5.249, "step": 310400 }, { "epoch": 1.8719624703030595, "grad_norm": 1.8997410535812378, "learning_rate": 0.0002507689972384987, "loss": 4.6448, "step": 310450 }, { "epoch": 1.8722639620843937, "grad_norm": 1.5336599349975586, "learning_rate": 0.0002507539230462337, "loss": 4.7817, "step": 310500 }, { "epoch": 1.8725654538657275, "grad_norm": 0.6350359320640564, "learning_rate": 0.0002507388469997365, "loss": 4.7888, "step": 310550 }, { "epoch": 1.8728669456470617, "grad_norm": 1.2101149559020996, "learning_rate": 0.0002507237690992846, "loss": 4.7195, "step": 310600 }, { "epoch": 1.8731684374283957, "grad_norm": 1.679758071899414, "learning_rate": 0.00025070868934515543, "loss": 5.2811, "step": 310650 }, { "epoch": 1.8734699292097297, "grad_norm": 1.309261679649353, "learning_rate": 0.0002506936077376266, "loss": 5.2319, "step": 310700 }, { "epoch": 1.8737714209910639, "grad_norm": 1.1995118856430054, "learning_rate": 0.00025067852427697557, "loss": 4.8348, "step": 310750 }, { "epoch": 1.8740729127723978, "grad_norm": 1.0604816675186157, "learning_rate": 0.00025066343896348, "loss": 4.8631, "step": 310800 }, { "epoch": 1.8743744045537318, "grad_norm": 1.2689590454101562, "learning_rate": 0.00025064835179741755, "loss": 4.864, "step": 310850 }, { "epoch": 1.874675896335066, "grad_norm": 1.8307877779006958, "learning_rate": 0.00025063326277906576, "loss": 4.9237, "step": 310900 }, { "epoch": 1.8749773881163998, "grad_norm": 1.9754704236984253, "learning_rate": 0.00025061817190870237, "loss": 5.1038, "step": 310950 }, { "epoch": 1.875278879897734, "grad_norm": 1.6296417713165283, "learning_rate": 0.0002506030791866051, "loss": 4.9393, "step": 311000 }, { "epoch": 1.875580371679068, "grad_norm": 2.301593780517578, "learning_rate": 0.00025058798461305177, "loss": 4.98, "step": 311050 }, { "epoch": 1.875881863460402, "grad_norm": 1.3395894765853882, "learning_rate": 0.0002505728881883201, "loss": 4.9952, "step": 311100 }, { "epoch": 1.8761833552417362, "grad_norm": 1.2561304569244385, "learning_rate": 0.000250557789912688, "loss": 5.0951, "step": 311150 }, { "epoch": 1.8764848470230702, "grad_norm": 2.366856575012207, "learning_rate": 0.00025054268978643324, "loss": 4.9118, "step": 311200 }, { "epoch": 1.8767863388044042, "grad_norm": 1.6664279699325562, "learning_rate": 0.00025052758780983375, "loss": 5.0016, "step": 311250 }, { "epoch": 1.8770878305857384, "grad_norm": 1.3663573265075684, "learning_rate": 0.00025051248398316744, "loss": 5.054, "step": 311300 }, { "epoch": 1.8773893223670721, "grad_norm": 1.3001840114593506, "learning_rate": 0.00025049737830671237, "loss": 4.9451, "step": 311350 }, { "epoch": 1.8776908141484063, "grad_norm": 0.9445094466209412, "learning_rate": 0.00025048227078074646, "loss": 4.8234, "step": 311400 }, { "epoch": 1.8779923059297403, "grad_norm": 0.7316136956214905, "learning_rate": 0.00025046716140554777, "loss": 5.1222, "step": 311450 }, { "epoch": 1.8782937977110743, "grad_norm": 1.4069174528121948, "learning_rate": 0.00025045205018139427, "loss": 4.7976, "step": 311500 }, { "epoch": 1.8785952894924085, "grad_norm": 1.3524880409240723, "learning_rate": 0.0002504369371085642, "loss": 4.7237, "step": 311550 }, { "epoch": 1.8788967812737425, "grad_norm": 1.9893128871917725, "learning_rate": 0.0002504218221873356, "loss": 4.9986, "step": 311600 }, { "epoch": 1.8791982730550765, "grad_norm": 0.8841208219528198, "learning_rate": 0.00025040670541798667, "loss": 5.1414, "step": 311650 }, { "epoch": 1.8794997648364107, "grad_norm": 1.6191624402999878, "learning_rate": 0.0002503915868007956, "loss": 5.0175, "step": 311700 }, { "epoch": 1.8798012566177444, "grad_norm": 2.2065980434417725, "learning_rate": 0.0002503764663360407, "loss": 5.0976, "step": 311750 }, { "epoch": 1.8801027483990786, "grad_norm": 2.1402735710144043, "learning_rate": 0.00025036134402400015, "loss": 5.0007, "step": 311800 }, { "epoch": 1.8804042401804126, "grad_norm": 1.2712122201919556, "learning_rate": 0.0002503462198649523, "loss": 5.2616, "step": 311850 }, { "epoch": 1.8807057319617466, "grad_norm": 0.7223392724990845, "learning_rate": 0.00025033109385917544, "loss": 5.1146, "step": 311900 }, { "epoch": 1.8810072237430808, "grad_norm": 1.052035927772522, "learning_rate": 0.000250315966006948, "loss": 4.7217, "step": 311950 }, { "epoch": 1.8813087155244148, "grad_norm": 0.5943346619606018, "learning_rate": 0.00025030083630854836, "loss": 5.0739, "step": 312000 }, { "epoch": 1.8813087155244148, "eval_loss": 5.516366958618164, "eval_runtime": 39.261, "eval_samples_per_second": 13.041, "eval_steps_per_second": 6.52, "eval_tts_loss": 7.388077819708021, "step": 312000 }, { "epoch": 1.8816102073057488, "grad_norm": 1.9451414346694946, "learning_rate": 0.000250285704764255, "loss": 4.9914, "step": 312050 }, { "epoch": 1.881911699087083, "grad_norm": 1.933398962020874, "learning_rate": 0.00025027057137434636, "loss": 4.9231, "step": 312100 }, { "epoch": 1.8822131908684168, "grad_norm": 1.4186338186264038, "learning_rate": 0.00025025543613910086, "loss": 5.4358, "step": 312150 }, { "epoch": 1.882514682649751, "grad_norm": 1.4130631685256958, "learning_rate": 0.0002502402990587972, "loss": 4.5313, "step": 312200 }, { "epoch": 1.882816174431085, "grad_norm": 1.9797297716140747, "learning_rate": 0.0002502251601337139, "loss": 5.0338, "step": 312250 }, { "epoch": 1.883117666212419, "grad_norm": 1.4033472537994385, "learning_rate": 0.0002502100193641295, "loss": 4.8962, "step": 312300 }, { "epoch": 1.8834191579937531, "grad_norm": 1.9164682626724243, "learning_rate": 0.00025019487675032276, "loss": 5.1364, "step": 312350 }, { "epoch": 1.8837206497750871, "grad_norm": 0.42783215641975403, "learning_rate": 0.0002501797322925723, "loss": 5.0532, "step": 312400 }, { "epoch": 1.8840221415564211, "grad_norm": 1.751720666885376, "learning_rate": 0.0002501645859911568, "loss": 4.8801, "step": 312450 }, { "epoch": 1.8843236333377553, "grad_norm": 1.5698885917663574, "learning_rate": 0.000250149437846355, "loss": 5.1951, "step": 312500 }, { "epoch": 1.8846251251190893, "grad_norm": 1.575459599494934, "learning_rate": 0.0002501342878584458, "loss": 5.1998, "step": 312550 }, { "epoch": 1.8849266169004233, "grad_norm": 1.2572909593582153, "learning_rate": 0.00025011913602770785, "loss": 5.0166, "step": 312600 }, { "epoch": 1.8852281086817575, "grad_norm": 2.1226890087127686, "learning_rate": 0.0002501039823544201, "loss": 5.0833, "step": 312650 }, { "epoch": 1.8855296004630913, "grad_norm": 0.9793957471847534, "learning_rate": 0.00025008882683886147, "loss": 5.0747, "step": 312700 }, { "epoch": 1.8858310922444255, "grad_norm": 2.0702240467071533, "learning_rate": 0.00025007366948131074, "loss": 5.0783, "step": 312750 }, { "epoch": 1.8861325840257595, "grad_norm": 0.5472849607467651, "learning_rate": 0.000250058510282047, "loss": 4.909, "step": 312800 }, { "epoch": 1.8864340758070934, "grad_norm": 1.5672798156738281, "learning_rate": 0.0002500433492413491, "loss": 5.0457, "step": 312850 }, { "epoch": 1.8867355675884276, "grad_norm": 1.7795900106430054, "learning_rate": 0.00025002818635949615, "loss": 5.2974, "step": 312900 }, { "epoch": 1.8870370593697616, "grad_norm": 1.3212474584579468, "learning_rate": 0.00025001302163676724, "loss": 4.9077, "step": 312950 }, { "epoch": 1.8873385511510956, "grad_norm": 2.0829904079437256, "learning_rate": 0.0002499978550734413, "loss": 5.6205, "step": 313000 }, { "epoch": 1.8876400429324298, "grad_norm": 2.042788505554199, "learning_rate": 0.0002499826866697977, "loss": 4.8936, "step": 313050 }, { "epoch": 1.8879415347137636, "grad_norm": 1.6887903213500977, "learning_rate": 0.0002499675164261153, "loss": 5.2833, "step": 313100 }, { "epoch": 1.8882430264950978, "grad_norm": 1.36870276927948, "learning_rate": 0.0002499523443426734, "loss": 5.2761, "step": 313150 }, { "epoch": 1.8885445182764318, "grad_norm": 1.469660758972168, "learning_rate": 0.00024993717041975127, "loss": 5.3493, "step": 313200 }, { "epoch": 1.8888460100577658, "grad_norm": 2.0518577098846436, "learning_rate": 0.0002499219946576282, "loss": 4.8516, "step": 313250 }, { "epoch": 1.8891475018391, "grad_norm": 1.1707655191421509, "learning_rate": 0.00024990681705658335, "loss": 4.9965, "step": 313300 }, { "epoch": 1.889448993620434, "grad_norm": 1.3118118047714233, "learning_rate": 0.0002498916376168961, "loss": 4.9707, "step": 313350 }, { "epoch": 1.889750485401768, "grad_norm": 1.289105772972107, "learning_rate": 0.00024987645633884584, "loss": 4.9106, "step": 313400 }, { "epoch": 1.8900519771831021, "grad_norm": 1.052148461341858, "learning_rate": 0.00024986127322271196, "loss": 5.1329, "step": 313450 }, { "epoch": 1.890353468964436, "grad_norm": 0.7200232148170471, "learning_rate": 0.0002498460882687738, "loss": 4.4497, "step": 313500 }, { "epoch": 1.89065496074577, "grad_norm": 1.6081581115722656, "learning_rate": 0.0002498309014773109, "loss": 4.9114, "step": 313550 }, { "epoch": 1.890956452527104, "grad_norm": 1.2195820808410645, "learning_rate": 0.00024981571284860276, "loss": 4.9207, "step": 313600 }, { "epoch": 1.891257944308438, "grad_norm": 0.28519847989082336, "learning_rate": 0.00024980052238292885, "loss": 5.2807, "step": 313650 }, { "epoch": 1.8915594360897723, "grad_norm": 1.987677812576294, "learning_rate": 0.0002497853300805687, "loss": 5.4333, "step": 313700 }, { "epoch": 1.8918609278711063, "grad_norm": 1.3396953344345093, "learning_rate": 0.000249770135941802, "loss": 5.0362, "step": 313750 }, { "epoch": 1.8921624196524403, "grad_norm": 2.2576379776000977, "learning_rate": 0.0002497549399669083, "loss": 5.2958, "step": 313800 }, { "epoch": 1.8924639114337745, "grad_norm": 2.11354923248291, "learning_rate": 0.0002497397421561673, "loss": 4.7846, "step": 313850 }, { "epoch": 1.8927654032151082, "grad_norm": 1.0982319116592407, "learning_rate": 0.0002497245425098587, "loss": 4.9099, "step": 313900 }, { "epoch": 1.8930668949964424, "grad_norm": 2.0870277881622314, "learning_rate": 0.0002497093410282622, "loss": 4.9965, "step": 313950 }, { "epoch": 1.8933683867777764, "grad_norm": 2.4484190940856934, "learning_rate": 0.0002496941377116576, "loss": 5.3436, "step": 314000 }, { "epoch": 1.8936698785591104, "grad_norm": 1.6933436393737793, "learning_rate": 0.00024967893256032457, "loss": 5.0878, "step": 314050 }, { "epoch": 1.8939713703404446, "grad_norm": 1.558496356010437, "learning_rate": 0.0002496637255745431, "loss": 5.0626, "step": 314100 }, { "epoch": 1.8942728621217786, "grad_norm": 1.1428108215332031, "learning_rate": 0.00024964851675459294, "loss": 5.3921, "step": 314150 }, { "epoch": 1.8945743539031126, "grad_norm": 1.6349339485168457, "learning_rate": 0.0002496333061007541, "loss": 5.21, "step": 314200 }, { "epoch": 1.8948758456844468, "grad_norm": 1.9235687255859375, "learning_rate": 0.00024961809361330646, "loss": 4.6966, "step": 314250 }, { "epoch": 1.8951773374657805, "grad_norm": 2.1006293296813965, "learning_rate": 0.00024960287929252993, "loss": 4.9441, "step": 314300 }, { "epoch": 1.8954788292471147, "grad_norm": 1.450727105140686, "learning_rate": 0.00024958766313870455, "loss": 5.5629, "step": 314350 }, { "epoch": 1.8957803210284487, "grad_norm": 1.5047568082809448, "learning_rate": 0.00024957244515211037, "loss": 5.0178, "step": 314400 }, { "epoch": 1.8960818128097827, "grad_norm": 0.7225234508514404, "learning_rate": 0.00024955722533302743, "loss": 5.1315, "step": 314450 }, { "epoch": 1.896383304591117, "grad_norm": 1.680594563484192, "learning_rate": 0.0002495420036817358, "loss": 5.2912, "step": 314500 }, { "epoch": 1.896684796372451, "grad_norm": 1.5928490161895752, "learning_rate": 0.0002495267801985156, "loss": 4.745, "step": 314550 }, { "epoch": 1.896986288153785, "grad_norm": 1.7447104454040527, "learning_rate": 0.00024951155488364717, "loss": 5.0454, "step": 314600 }, { "epoch": 1.897287779935119, "grad_norm": 1.7678942680358887, "learning_rate": 0.00024949632773741055, "loss": 5.1279, "step": 314650 }, { "epoch": 1.8975892717164529, "grad_norm": 1.3929646015167236, "learning_rate": 0.0002494810987600859, "loss": 5.0265, "step": 314700 }, { "epoch": 1.897890763497787, "grad_norm": 1.611208200454712, "learning_rate": 0.0002494658679519537, "loss": 4.9605, "step": 314750 }, { "epoch": 1.898192255279121, "grad_norm": 1.246377944946289, "learning_rate": 0.0002494506353132941, "loss": 4.4731, "step": 314800 }, { "epoch": 1.898493747060455, "grad_norm": 1.3710594177246094, "learning_rate": 0.0002494354008443875, "loss": 4.9443, "step": 314850 }, { "epoch": 1.8987952388417892, "grad_norm": 2.4245197772979736, "learning_rate": 0.0002494201645455143, "loss": 5.3788, "step": 314900 }, { "epoch": 1.8990967306231232, "grad_norm": 1.9338147640228271, "learning_rate": 0.0002494049264169548, "loss": 5.072, "step": 314950 }, { "epoch": 1.8993982224044572, "grad_norm": 1.139269232749939, "learning_rate": 0.00024938968645898946, "loss": 5.0816, "step": 315000 }, { "epoch": 1.8993982224044572, "eval_loss": 5.529683589935303, "eval_runtime": 39.2085, "eval_samples_per_second": 13.058, "eval_steps_per_second": 6.529, "eval_tts_loss": 7.250932620776029, "step": 315000 }, { "epoch": 1.8996997141857914, "grad_norm": 1.3263046741485596, "learning_rate": 0.0002493744446718988, "loss": 4.7433, "step": 315050 }, { "epoch": 1.9000012059671252, "grad_norm": 1.6044598817825317, "learning_rate": 0.0002493592010559633, "loss": 5.0964, "step": 315100 }, { "epoch": 1.9003026977484594, "grad_norm": 1.857555627822876, "learning_rate": 0.00024934395561146356, "loss": 5.0941, "step": 315150 }, { "epoch": 1.9006041895297934, "grad_norm": 1.3656076192855835, "learning_rate": 0.00024932870833867997, "loss": 5.3082, "step": 315200 }, { "epoch": 1.9009056813111274, "grad_norm": 1.1433864831924438, "learning_rate": 0.00024931345923789335, "loss": 4.7508, "step": 315250 }, { "epoch": 1.9012071730924616, "grad_norm": 1.671104073524475, "learning_rate": 0.0002492982083093842, "loss": 4.8719, "step": 315300 }, { "epoch": 1.9015086648737956, "grad_norm": 2.1754424571990967, "learning_rate": 0.00024928295555343323, "loss": 5.3818, "step": 315350 }, { "epoch": 1.9018101566551295, "grad_norm": 1.4462562799453735, "learning_rate": 0.00024926770097032113, "loss": 5.1461, "step": 315400 }, { "epoch": 1.9021116484364637, "grad_norm": 1.8592743873596191, "learning_rate": 0.0002492524445603287, "loss": 5.0606, "step": 315450 }, { "epoch": 1.9024131402177975, "grad_norm": 1.845716118812561, "learning_rate": 0.0002492371863237367, "loss": 5.2498, "step": 315500 }, { "epoch": 1.9027146319991317, "grad_norm": 1.4354639053344727, "learning_rate": 0.00024922192626082584, "loss": 4.9039, "step": 315550 }, { "epoch": 1.9030161237804657, "grad_norm": 1.3789528608322144, "learning_rate": 0.00024920666437187704, "loss": 4.9565, "step": 315600 }, { "epoch": 1.9033176155617997, "grad_norm": 2.3351428508758545, "learning_rate": 0.00024919140065717113, "loss": 5.3907, "step": 315650 }, { "epoch": 1.9036191073431339, "grad_norm": 1.7065565586090088, "learning_rate": 0.0002491761351169891, "loss": 5.3833, "step": 315700 }, { "epoch": 1.9039205991244679, "grad_norm": 0.484394371509552, "learning_rate": 0.0002491608677516118, "loss": 4.6385, "step": 315750 }, { "epoch": 1.9042220909058019, "grad_norm": 2.38742733001709, "learning_rate": 0.00024914559856132023, "loss": 5.4261, "step": 315800 }, { "epoch": 1.904523582687136, "grad_norm": 1.3787366151809692, "learning_rate": 0.0002491303275463954, "loss": 5.2287, "step": 315850 }, { "epoch": 1.9048250744684698, "grad_norm": 1.5478427410125732, "learning_rate": 0.0002491150547071184, "loss": 5.3477, "step": 315900 }, { "epoch": 1.905126566249804, "grad_norm": 1.208040475845337, "learning_rate": 0.0002490997800437702, "loss": 5.1442, "step": 315950 }, { "epoch": 1.905428058031138, "grad_norm": 1.4968886375427246, "learning_rate": 0.00024908450355663205, "loss": 5.2915, "step": 316000 }, { "epoch": 1.905729549812472, "grad_norm": 1.5645548105239868, "learning_rate": 0.000249069225245985, "loss": 5.2686, "step": 316050 }, { "epoch": 1.9060310415938062, "grad_norm": 1.8757014274597168, "learning_rate": 0.0002490539451121102, "loss": 5.434, "step": 316100 }, { "epoch": 1.9063325333751402, "grad_norm": 1.6222349405288696, "learning_rate": 0.0002490386631552889, "loss": 5.0516, "step": 316150 }, { "epoch": 1.9066340251564742, "grad_norm": 1.9742145538330078, "learning_rate": 0.0002490233793758024, "loss": 5.0212, "step": 316200 }, { "epoch": 1.9069355169378084, "grad_norm": 2.2284152507781982, "learning_rate": 0.00024900809377393187, "loss": 5.2366, "step": 316250 }, { "epoch": 1.9072370087191424, "grad_norm": 1.569120168685913, "learning_rate": 0.00024899280634995866, "loss": 5.214, "step": 316300 }, { "epoch": 1.9075385005004764, "grad_norm": 2.0533087253570557, "learning_rate": 0.0002489775171041641, "loss": 4.883, "step": 316350 }, { "epoch": 1.9078399922818106, "grad_norm": 1.4743576049804688, "learning_rate": 0.0002489622260368296, "loss": 4.7713, "step": 316400 }, { "epoch": 1.9081414840631443, "grad_norm": 2.37616229057312, "learning_rate": 0.0002489469331482366, "loss": 4.9146, "step": 316450 }, { "epoch": 1.9084429758444785, "grad_norm": 1.3759758472442627, "learning_rate": 0.0002489316384386665, "loss": 5.2632, "step": 316500 }, { "epoch": 1.9087444676258125, "grad_norm": 2.3353593349456787, "learning_rate": 0.0002489163419084007, "loss": 5.3061, "step": 316550 }, { "epoch": 1.9090459594071465, "grad_norm": 1.266197681427002, "learning_rate": 0.00024890104355772085, "loss": 5.0058, "step": 316600 }, { "epoch": 1.9093474511884807, "grad_norm": 1.8684016466140747, "learning_rate": 0.0002488857433869084, "loss": 5.073, "step": 316650 }, { "epoch": 1.9096489429698147, "grad_norm": 1.2566783428192139, "learning_rate": 0.000248870441396245, "loss": 4.6892, "step": 316700 }, { "epoch": 1.9099504347511487, "grad_norm": 1.4862316846847534, "learning_rate": 0.0002488551375860122, "loss": 4.9731, "step": 316750 }, { "epoch": 1.9102519265324829, "grad_norm": 1.6986700296401978, "learning_rate": 0.00024883983195649165, "loss": 4.285, "step": 316800 }, { "epoch": 1.9105534183138166, "grad_norm": 1.74150812625885, "learning_rate": 0.0002488245245079651, "loss": 5.4076, "step": 316850 }, { "epoch": 1.9108549100951508, "grad_norm": 1.3930222988128662, "learning_rate": 0.00024880921524071415, "loss": 4.856, "step": 316900 }, { "epoch": 1.9111564018764848, "grad_norm": 1.7132394313812256, "learning_rate": 0.00024879390415502063, "loss": 5.6393, "step": 316950 }, { "epoch": 1.9114578936578188, "grad_norm": 1.74297034740448, "learning_rate": 0.00024877859125116625, "loss": 5.1034, "step": 317000 }, { "epoch": 1.911759385439153, "grad_norm": 1.953493595123291, "learning_rate": 0.0002487632765294329, "loss": 5.2267, "step": 317050 }, { "epoch": 1.912060877220487, "grad_norm": 1.896254301071167, "learning_rate": 0.0002487479599901024, "loss": 5.2408, "step": 317100 }, { "epoch": 1.912362369001821, "grad_norm": 1.4679303169250488, "learning_rate": 0.0002487326416334566, "loss": 5.1059, "step": 317150 }, { "epoch": 1.9126638607831552, "grad_norm": 1.5187623500823975, "learning_rate": 0.00024871732145977744, "loss": 5.4677, "step": 317200 }, { "epoch": 1.912965352564489, "grad_norm": 1.251627802848816, "learning_rate": 0.00024870199946934685, "loss": 4.9584, "step": 317250 }, { "epoch": 1.9132668443458232, "grad_norm": 1.4408172369003296, "learning_rate": 0.0002486866756624468, "loss": 5.2273, "step": 317300 }, { "epoch": 1.9135683361271572, "grad_norm": 1.7117278575897217, "learning_rate": 0.0002486713500393593, "loss": 5.1699, "step": 317350 }, { "epoch": 1.9138698279084911, "grad_norm": 2.3095386028289795, "learning_rate": 0.00024865602260036644, "loss": 5.209, "step": 317400 }, { "epoch": 1.9141713196898253, "grad_norm": 1.2172120809555054, "learning_rate": 0.00024864069334575034, "loss": 5.1254, "step": 317450 }, { "epoch": 1.9144728114711593, "grad_norm": 1.643635869026184, "learning_rate": 0.00024862536227579295, "loss": 5.2145, "step": 317500 }, { "epoch": 1.9147743032524933, "grad_norm": 2.0041258335113525, "learning_rate": 0.00024861002939077656, "loss": 5.0313, "step": 317550 }, { "epoch": 1.9150757950338275, "grad_norm": 2.4559621810913086, "learning_rate": 0.00024859469469098327, "loss": 4.7067, "step": 317600 }, { "epoch": 1.9153772868151613, "grad_norm": 0.8350913524627686, "learning_rate": 0.00024857935817669535, "loss": 4.8766, "step": 317650 }, { "epoch": 1.9156787785964955, "grad_norm": 1.17795729637146, "learning_rate": 0.00024856401984819503, "loss": 4.7108, "step": 317700 }, { "epoch": 1.9159802703778295, "grad_norm": 0.7666353583335876, "learning_rate": 0.00024854867970576457, "loss": 5.0432, "step": 317750 }, { "epoch": 1.9162817621591635, "grad_norm": 1.2600256204605103, "learning_rate": 0.0002485333377496863, "loss": 4.9021, "step": 317800 }, { "epoch": 1.9165832539404977, "grad_norm": 1.3065036535263062, "learning_rate": 0.00024851799398024253, "loss": 4.8817, "step": 317850 }, { "epoch": 1.9168847457218317, "grad_norm": 0.7576966285705566, "learning_rate": 0.00024850264839771573, "loss": 5.0162, "step": 317900 }, { "epoch": 1.9171862375031656, "grad_norm": 1.62738037109375, "learning_rate": 0.00024848730100238825, "loss": 5.0719, "step": 317950 }, { "epoch": 1.9174877292844998, "grad_norm": 1.5252028703689575, "learning_rate": 0.0002484719517945425, "loss": 5.3545, "step": 318000 }, { "epoch": 1.9174877292844998, "eval_loss": 5.508120536804199, "eval_runtime": 38.9728, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.569, "eval_tts_loss": 7.328344843955171, "step": 318000 }, { "epoch": 1.9177892210658336, "grad_norm": 1.8969730138778687, "learning_rate": 0.00024845660077446103, "loss": 5.2518, "step": 318050 }, { "epoch": 1.9180907128471678, "grad_norm": 1.4034888744354248, "learning_rate": 0.0002484412479424263, "loss": 5.3284, "step": 318100 }, { "epoch": 1.9183922046285018, "grad_norm": 1.5126172304153442, "learning_rate": 0.00024842589329872094, "loss": 5.0308, "step": 318150 }, { "epoch": 1.9186936964098358, "grad_norm": 1.5906933546066284, "learning_rate": 0.0002484105368436275, "loss": 4.9723, "step": 318200 }, { "epoch": 1.91899518819117, "grad_norm": 1.3341706991195679, "learning_rate": 0.0002483951785774285, "loss": 5.2816, "step": 318250 }, { "epoch": 1.919296679972504, "grad_norm": 1.2697529792785645, "learning_rate": 0.00024837981850040663, "loss": 5.1436, "step": 318300 }, { "epoch": 1.919598171753838, "grad_norm": 1.3173513412475586, "learning_rate": 0.0002483644566128447, "loss": 5.0713, "step": 318350 }, { "epoch": 1.9198996635351722, "grad_norm": 1.9951523542404175, "learning_rate": 0.0002483490929150252, "loss": 4.9701, "step": 318400 }, { "epoch": 1.920201155316506, "grad_norm": 2.019516706466675, "learning_rate": 0.00024833372740723107, "loss": 5.2254, "step": 318450 }, { "epoch": 1.9205026470978401, "grad_norm": 1.5060330629348755, "learning_rate": 0.000248318360089745, "loss": 4.9053, "step": 318500 }, { "epoch": 1.9208041388791741, "grad_norm": 1.4277794361114502, "learning_rate": 0.0002483029909628498, "loss": 4.6208, "step": 318550 }, { "epoch": 1.921105630660508, "grad_norm": 1.4518994092941284, "learning_rate": 0.0002482876200268284, "loss": 5.239, "step": 318600 }, { "epoch": 1.9214071224418423, "grad_norm": 1.5362588167190552, "learning_rate": 0.0002482722472819636, "loss": 5.2273, "step": 318650 }, { "epoch": 1.9217086142231763, "grad_norm": 1.3877978324890137, "learning_rate": 0.0002482568727285383, "loss": 5.0104, "step": 318700 }, { "epoch": 1.9220101060045103, "grad_norm": 2.7437942028045654, "learning_rate": 0.00024824149636683545, "loss": 5.2583, "step": 318750 }, { "epoch": 1.9223115977858445, "grad_norm": 1.2619937658309937, "learning_rate": 0.00024822611819713816, "loss": 5.7451, "step": 318800 }, { "epoch": 1.9226130895671782, "grad_norm": 1.7641527652740479, "learning_rate": 0.0002482107382197292, "loss": 5.0442, "step": 318850 }, { "epoch": 1.9229145813485125, "grad_norm": 1.6170175075531006, "learning_rate": 0.00024819535643489185, "loss": 5.1071, "step": 318900 }, { "epoch": 1.9232160731298464, "grad_norm": 0.4995718002319336, "learning_rate": 0.0002481799728429091, "loss": 4.4036, "step": 318950 }, { "epoch": 1.9235175649111804, "grad_norm": 1.679093599319458, "learning_rate": 0.00024816458744406403, "loss": 4.8745, "step": 319000 }, { "epoch": 1.9238190566925146, "grad_norm": 1.994027853012085, "learning_rate": 0.0002481492002386398, "loss": 5.1449, "step": 319050 }, { "epoch": 1.9241205484738486, "grad_norm": 1.1592472791671753, "learning_rate": 0.00024813381122691967, "loss": 5.1837, "step": 319100 }, { "epoch": 1.9244220402551826, "grad_norm": 1.5350098609924316, "learning_rate": 0.0002481184204091867, "loss": 4.9215, "step": 319150 }, { "epoch": 1.9247235320365168, "grad_norm": 1.621509075164795, "learning_rate": 0.0002481030277857243, "loss": 5.3366, "step": 319200 }, { "epoch": 1.9250250238178506, "grad_norm": 1.372517704963684, "learning_rate": 0.0002480876333568157, "loss": 4.7647, "step": 319250 }, { "epoch": 1.9253265155991848, "grad_norm": 1.5479817390441895, "learning_rate": 0.0002480722371227441, "loss": 4.7796, "step": 319300 }, { "epoch": 1.9256280073805188, "grad_norm": 1.1241117715835571, "learning_rate": 0.000248056839083793, "loss": 5.4935, "step": 319350 }, { "epoch": 1.9259294991618527, "grad_norm": 1.7426440715789795, "learning_rate": 0.00024804143924024564, "loss": 5.0172, "step": 319400 }, { "epoch": 1.926230990943187, "grad_norm": 1.326737880706787, "learning_rate": 0.00024802603759238557, "loss": 4.7654, "step": 319450 }, { "epoch": 1.926532482724521, "grad_norm": 2.2111058235168457, "learning_rate": 0.0002480106341404961, "loss": 4.9153, "step": 319500 }, { "epoch": 1.926833974505855, "grad_norm": 2.815544605255127, "learning_rate": 0.00024799522888486087, "loss": 5.3836, "step": 319550 }, { "epoch": 1.9271354662871891, "grad_norm": 1.2278385162353516, "learning_rate": 0.0002479798218257632, "loss": 5.4966, "step": 319600 }, { "epoch": 1.9274369580685229, "grad_norm": 1.1909863948822021, "learning_rate": 0.0002479644129634868, "loss": 4.7951, "step": 319650 }, { "epoch": 1.927738449849857, "grad_norm": 0.33808550238609314, "learning_rate": 0.0002479490022983151, "loss": 5.0505, "step": 319700 }, { "epoch": 1.928039941631191, "grad_norm": 1.2095813751220703, "learning_rate": 0.0002479335898305319, "loss": 5.3716, "step": 319750 }, { "epoch": 1.928341433412525, "grad_norm": 2.0433506965637207, "learning_rate": 0.0002479181755604207, "loss": 5.0882, "step": 319800 }, { "epoch": 1.9286429251938593, "grad_norm": 1.8252662420272827, "learning_rate": 0.00024790275948826523, "loss": 5.0111, "step": 319850 }, { "epoch": 1.9289444169751933, "grad_norm": 1.3902931213378906, "learning_rate": 0.00024788734161434913, "loss": 5.0729, "step": 319900 }, { "epoch": 1.9292459087565272, "grad_norm": 0.9640693664550781, "learning_rate": 0.00024787192193895626, "loss": 4.6883, "step": 319950 }, { "epoch": 1.9295474005378614, "grad_norm": 1.4819703102111816, "learning_rate": 0.0002478565004623703, "loss": 5.0539, "step": 320000 }, { "epoch": 1.9298488923191954, "grad_norm": 2.3088841438293457, "learning_rate": 0.00024784107718487514, "loss": 5.3641, "step": 320050 }, { "epoch": 1.9301503841005294, "grad_norm": 1.438851237297058, "learning_rate": 0.0002478256521067546, "loss": 4.8369, "step": 320100 }, { "epoch": 1.9304518758818636, "grad_norm": 1.5891302824020386, "learning_rate": 0.0002478102252282924, "loss": 5.1757, "step": 320150 }, { "epoch": 1.9307533676631974, "grad_norm": 2.324437141418457, "learning_rate": 0.0002477947965497727, "loss": 5.2705, "step": 320200 }, { "epoch": 1.9310548594445316, "grad_norm": 1.1491178274154663, "learning_rate": 0.00024777936607147934, "loss": 5.2518, "step": 320250 }, { "epoch": 1.9313563512258656, "grad_norm": 1.4928703308105469, "learning_rate": 0.0002477639337936962, "loss": 4.65, "step": 320300 }, { "epoch": 1.9316578430071996, "grad_norm": 1.610618233680725, "learning_rate": 0.00024774849971670747, "loss": 4.6173, "step": 320350 }, { "epoch": 1.9319593347885338, "grad_norm": 1.2652558088302612, "learning_rate": 0.00024773306384079705, "loss": 5.3548, "step": 320400 }, { "epoch": 1.9322608265698678, "grad_norm": 1.931390404701233, "learning_rate": 0.00024771762616624906, "loss": 4.7128, "step": 320450 }, { "epoch": 1.9325623183512017, "grad_norm": 2.2639822959899902, "learning_rate": 0.0002477021866933476, "loss": 5.253, "step": 320500 }, { "epoch": 1.932863810132536, "grad_norm": 1.7460839748382568, "learning_rate": 0.0002476867454223769, "loss": 5.154, "step": 320550 }, { "epoch": 1.9331653019138697, "grad_norm": 1.4259179830551147, "learning_rate": 0.00024767130235362096, "loss": 5.1056, "step": 320600 }, { "epoch": 1.933466793695204, "grad_norm": 2.6350345611572266, "learning_rate": 0.00024765585748736415, "loss": 5.3656, "step": 320650 }, { "epoch": 1.933768285476538, "grad_norm": 0.7519969344139099, "learning_rate": 0.00024764041082389057, "loss": 5.2591, "step": 320700 }, { "epoch": 1.9340697772578719, "grad_norm": 1.40339195728302, "learning_rate": 0.00024762496236348464, "loss": 5.2323, "step": 320750 }, { "epoch": 1.934371269039206, "grad_norm": 2.054314374923706, "learning_rate": 0.0002476095121064306, "loss": 5.5473, "step": 320800 }, { "epoch": 1.93467276082054, "grad_norm": 1.2627817392349243, "learning_rate": 0.00024759406005301276, "loss": 5.1916, "step": 320850 }, { "epoch": 1.934974252601874, "grad_norm": 0.7346835136413574, "learning_rate": 0.0002475786062035155, "loss": 5.5026, "step": 320900 }, { "epoch": 1.9352757443832083, "grad_norm": 1.8468918800354004, "learning_rate": 0.0002475631505582233, "loss": 5.3004, "step": 320950 }, { "epoch": 1.935577236164542, "grad_norm": 2.1407597064971924, "learning_rate": 0.0002475476931174205, "loss": 4.9286, "step": 321000 }, { "epoch": 1.935577236164542, "eval_loss": 5.523394584655762, "eval_runtime": 39.1015, "eval_samples_per_second": 13.094, "eval_steps_per_second": 6.547, "eval_tts_loss": 7.321309853923885, "step": 321000 }, { "epoch": 1.9358787279458762, "grad_norm": 1.863466501235962, "learning_rate": 0.00024753223388139164, "loss": 5.1403, "step": 321050 }, { "epoch": 1.9361802197272102, "grad_norm": 1.2554746866226196, "learning_rate": 0.00024751677285042117, "loss": 4.9362, "step": 321100 }, { "epoch": 1.9364817115085442, "grad_norm": 0.9729339480400085, "learning_rate": 0.00024750131002479367, "loss": 5.1627, "step": 321150 }, { "epoch": 1.9367832032898784, "grad_norm": 1.1993614435195923, "learning_rate": 0.00024748584540479373, "loss": 4.6223, "step": 321200 }, { "epoch": 1.9370846950712124, "grad_norm": 1.2583128213882446, "learning_rate": 0.0002474703789907059, "loss": 4.9175, "step": 321250 }, { "epoch": 1.9373861868525464, "grad_norm": 1.509033441543579, "learning_rate": 0.0002474549107828148, "loss": 4.9833, "step": 321300 }, { "epoch": 1.9376876786338806, "grad_norm": 0.9324595928192139, "learning_rate": 0.0002474394407814052, "loss": 5.1814, "step": 321350 }, { "epoch": 1.9379891704152143, "grad_norm": 1.716727375984192, "learning_rate": 0.00024742396898676174, "loss": 5.0917, "step": 321400 }, { "epoch": 1.9382906621965486, "grad_norm": 2.6608517169952393, "learning_rate": 0.00024740849539916913, "loss": 4.9964, "step": 321450 }, { "epoch": 1.9385921539778825, "grad_norm": 1.342793345451355, "learning_rate": 0.0002473930200189122, "loss": 4.5426, "step": 321500 }, { "epoch": 1.9388936457592165, "grad_norm": 2.6659770011901855, "learning_rate": 0.00024737754284627576, "loss": 5.1198, "step": 321550 }, { "epoch": 1.9391951375405507, "grad_norm": 2.295457363128662, "learning_rate": 0.0002473620638815445, "loss": 5.2855, "step": 321600 }, { "epoch": 1.9394966293218847, "grad_norm": 1.4848747253417969, "learning_rate": 0.0002473465831250035, "loss": 5.0875, "step": 321650 }, { "epoch": 1.9397981211032187, "grad_norm": 1.344787836074829, "learning_rate": 0.00024733110057693756, "loss": 4.9018, "step": 321700 }, { "epoch": 1.940099612884553, "grad_norm": 1.957626223564148, "learning_rate": 0.0002473156162376315, "loss": 4.9884, "step": 321750 }, { "epoch": 1.9404011046658867, "grad_norm": 2.2988648414611816, "learning_rate": 0.0002473001301073705, "loss": 4.9508, "step": 321800 }, { "epoch": 1.9407025964472209, "grad_norm": 1.7792681455612183, "learning_rate": 0.0002472846421864395, "loss": 5.1811, "step": 321850 }, { "epoch": 1.9410040882285549, "grad_norm": 2.0279901027679443, "learning_rate": 0.0002472691524751234, "loss": 5.2086, "step": 321900 }, { "epoch": 1.9413055800098888, "grad_norm": 1.9139002561569214, "learning_rate": 0.0002472536609737074, "loss": 4.9841, "step": 321950 }, { "epoch": 1.941607071791223, "grad_norm": 1.770529866218567, "learning_rate": 0.00024723816768247657, "loss": 4.946, "step": 322000 }, { "epoch": 1.941908563572557, "grad_norm": 1.2288658618927002, "learning_rate": 0.000247222672601716, "loss": 5.0233, "step": 322050 }, { "epoch": 1.942210055353891, "grad_norm": 1.3368414640426636, "learning_rate": 0.0002472071757317109, "loss": 4.6111, "step": 322100 }, { "epoch": 1.9425115471352252, "grad_norm": 1.0465943813323975, "learning_rate": 0.00024719167707274646, "loss": 4.9341, "step": 322150 }, { "epoch": 1.942813038916559, "grad_norm": 1.4244179725646973, "learning_rate": 0.00024717617662510787, "loss": 5.1879, "step": 322200 }, { "epoch": 1.9431145306978932, "grad_norm": 1.556726336479187, "learning_rate": 0.00024716067438908045, "loss": 4.8896, "step": 322250 }, { "epoch": 1.9434160224792272, "grad_norm": 0.9418824315071106, "learning_rate": 0.00024714517036494946, "loss": 4.7155, "step": 322300 }, { "epoch": 1.9437175142605612, "grad_norm": 1.359824299812317, "learning_rate": 0.00024712966455300026, "loss": 5.0896, "step": 322350 }, { "epoch": 1.9440190060418954, "grad_norm": 1.874659776687622, "learning_rate": 0.0002471141569535182, "loss": 4.9021, "step": 322400 }, { "epoch": 1.9443204978232294, "grad_norm": 1.226171612739563, "learning_rate": 0.00024709864756678867, "loss": 5.2115, "step": 322450 }, { "epoch": 1.9446219896045633, "grad_norm": 1.46921706199646, "learning_rate": 0.00024708313639309706, "loss": 5.1961, "step": 322500 }, { "epoch": 1.9449234813858975, "grad_norm": 0.7405058145523071, "learning_rate": 0.0002470676234327289, "loss": 4.5097, "step": 322550 }, { "epoch": 1.9452249731672313, "grad_norm": 1.9173005819320679, "learning_rate": 0.00024705210868596967, "loss": 4.9402, "step": 322600 }, { "epoch": 1.9455264649485655, "grad_norm": 0.541588544845581, "learning_rate": 0.0002470365921531049, "loss": 5.2298, "step": 322650 }, { "epoch": 1.9458279567298995, "grad_norm": 2.6298916339874268, "learning_rate": 0.00024702107383442007, "loss": 4.909, "step": 322700 }, { "epoch": 1.9461294485112335, "grad_norm": 1.3210078477859497, "learning_rate": 0.00024700555373020086, "loss": 5.2276, "step": 322750 }, { "epoch": 1.9464309402925677, "grad_norm": 1.9464111328125, "learning_rate": 0.00024699003184073284, "loss": 5.2706, "step": 322800 }, { "epoch": 1.9467324320739017, "grad_norm": 1.7088098526000977, "learning_rate": 0.00024697450816630175, "loss": 5.6194, "step": 322850 }, { "epoch": 1.9470339238552357, "grad_norm": 1.4720138311386108, "learning_rate": 0.00024695898270719316, "loss": 5.1799, "step": 322900 }, { "epoch": 1.9473354156365699, "grad_norm": 1.3653534650802612, "learning_rate": 0.0002469434554636929, "loss": 5.0231, "step": 322950 }, { "epoch": 1.9476369074179036, "grad_norm": 0.7894729375839233, "learning_rate": 0.0002469279264360867, "loss": 5.2606, "step": 323000 }, { "epoch": 1.9479383991992378, "grad_norm": 1.7452212572097778, "learning_rate": 0.0002469123956246603, "loss": 5.4138, "step": 323050 }, { "epoch": 1.9482398909805718, "grad_norm": 1.4124888181686401, "learning_rate": 0.0002468968630296996, "loss": 5.2148, "step": 323100 }, { "epoch": 1.9485413827619058, "grad_norm": 1.4714405536651611, "learning_rate": 0.0002468813286514904, "loss": 5.5904, "step": 323150 }, { "epoch": 1.94884287454324, "grad_norm": 1.7668625116348267, "learning_rate": 0.0002468657924903186, "loss": 4.7909, "step": 323200 }, { "epoch": 1.949144366324574, "grad_norm": 1.8990533351898193, "learning_rate": 0.0002468502545464701, "loss": 5.1755, "step": 323250 }, { "epoch": 1.949445858105908, "grad_norm": 1.308420181274414, "learning_rate": 0.0002468347148202309, "loss": 5.3901, "step": 323300 }, { "epoch": 1.9497473498872422, "grad_norm": 2.0919132232666016, "learning_rate": 0.000246819173311887, "loss": 5.3359, "step": 323350 }, { "epoch": 1.950048841668576, "grad_norm": 1.3073275089263916, "learning_rate": 0.0002468036300217243, "loss": 4.9608, "step": 323400 }, { "epoch": 1.9503503334499102, "grad_norm": 2.0481560230255127, "learning_rate": 0.00024678808495002906, "loss": 5.4708, "step": 323450 }, { "epoch": 1.9506518252312441, "grad_norm": 1.6984933614730835, "learning_rate": 0.00024677253809708715, "loss": 5.0825, "step": 323500 }, { "epoch": 1.9509533170125781, "grad_norm": 1.4072688817977905, "learning_rate": 0.00024675698946318483, "loss": 5.2301, "step": 323550 }, { "epoch": 1.9512548087939123, "grad_norm": 2.6339080333709717, "learning_rate": 0.0002467414390486081, "loss": 5.1476, "step": 323600 }, { "epoch": 1.9515563005752463, "grad_norm": 1.2492728233337402, "learning_rate": 0.00024672588685364336, "loss": 5.1489, "step": 323650 }, { "epoch": 1.9518577923565803, "grad_norm": 0.7494859099388123, "learning_rate": 0.0002467103328785766, "loss": 4.976, "step": 323700 }, { "epoch": 1.9521592841379145, "grad_norm": 2.38767147064209, "learning_rate": 0.0002466947771236943, "loss": 5.2347, "step": 323750 }, { "epoch": 1.9524607759192485, "grad_norm": 1.3795132637023926, "learning_rate": 0.0002466792195892825, "loss": 4.6611, "step": 323800 }, { "epoch": 1.9527622677005825, "grad_norm": 0.559702455997467, "learning_rate": 0.00024666366027562773, "loss": 5.2201, "step": 323850 }, { "epoch": 1.9530637594819167, "grad_norm": 1.550021767616272, "learning_rate": 0.00024664809918301616, "loss": 5.4526, "step": 323900 }, { "epoch": 1.9533652512632504, "grad_norm": 0.8199068307876587, "learning_rate": 0.0002466325363117343, "loss": 5.0452, "step": 323950 }, { "epoch": 1.9536667430445847, "grad_norm": 1.3070957660675049, "learning_rate": 0.0002466169716620684, "loss": 5.0755, "step": 324000 }, { "epoch": 1.9536667430445847, "eval_loss": 5.523186683654785, "eval_runtime": 39.2304, "eval_samples_per_second": 13.051, "eval_steps_per_second": 6.526, "eval_tts_loss": 7.3110322345993355, "step": 324000 }, { "epoch": 1.9539682348259186, "grad_norm": 2.412703037261963, "learning_rate": 0.0002466014052343052, "loss": 4.9929, "step": 324050 }, { "epoch": 1.9542697266072526, "grad_norm": 1.444990634918213, "learning_rate": 0.00024658583702873084, "loss": 5.1178, "step": 324100 }, { "epoch": 1.9545712183885868, "grad_norm": 1.347989559173584, "learning_rate": 0.0002465702670456321, "loss": 5.3795, "step": 324150 }, { "epoch": 1.9548727101699208, "grad_norm": 1.0934830904006958, "learning_rate": 0.00024655469528529533, "loss": 5.0228, "step": 324200 }, { "epoch": 1.9551742019512548, "grad_norm": 1.980638861656189, "learning_rate": 0.00024653912174800723, "loss": 5.3502, "step": 324250 }, { "epoch": 1.955475693732589, "grad_norm": 0.7506694197654724, "learning_rate": 0.0002465235464340543, "loss": 4.5187, "step": 324300 }, { "epoch": 1.9557771855139228, "grad_norm": 1.4685373306274414, "learning_rate": 0.0002465079693437233, "loss": 5.0862, "step": 324350 }, { "epoch": 1.956078677295257, "grad_norm": 1.38614821434021, "learning_rate": 0.00024649239047730084, "loss": 4.5832, "step": 324400 }, { "epoch": 1.956380169076591, "grad_norm": 0.6133763790130615, "learning_rate": 0.00024647680983507365, "loss": 4.9781, "step": 324450 }, { "epoch": 1.956681660857925, "grad_norm": 0.3446091413497925, "learning_rate": 0.0002464612274173284, "loss": 4.8095, "step": 324500 }, { "epoch": 1.9569831526392591, "grad_norm": 2.8220834732055664, "learning_rate": 0.00024644564322435196, "loss": 5.382, "step": 324550 }, { "epoch": 1.9572846444205931, "grad_norm": 1.3944159746170044, "learning_rate": 0.00024643005725643105, "loss": 5.6001, "step": 324600 }, { "epoch": 1.9575861362019271, "grad_norm": 1.730661392211914, "learning_rate": 0.00024641446951385263, "loss": 5.0643, "step": 324650 }, { "epoch": 1.9578876279832613, "grad_norm": 1.444913625717163, "learning_rate": 0.0002463988799969034, "loss": 4.9177, "step": 324700 }, { "epoch": 1.958189119764595, "grad_norm": 1.5032461881637573, "learning_rate": 0.00024638328870587044, "loss": 5.2473, "step": 324750 }, { "epoch": 1.9584906115459293, "grad_norm": 2.3509435653686523, "learning_rate": 0.00024636769564104053, "loss": 5.1737, "step": 324800 }, { "epoch": 1.9587921033272633, "grad_norm": 1.4343712329864502, "learning_rate": 0.00024635210080270074, "loss": 4.9501, "step": 324850 }, { "epoch": 1.9590935951085973, "grad_norm": 1.680198073387146, "learning_rate": 0.000246336504191138, "loss": 4.9017, "step": 324900 }, { "epoch": 1.9593950868899315, "grad_norm": 2.738692283630371, "learning_rate": 0.0002463209058066394, "loss": 4.5958, "step": 324950 }, { "epoch": 1.9596965786712655, "grad_norm": 1.9397082328796387, "learning_rate": 0.000246305305649492, "loss": 4.919, "step": 325000 }, { "epoch": 1.9599980704525994, "grad_norm": 2.552278995513916, "learning_rate": 0.00024628970371998285, "loss": 5.3103, "step": 325050 }, { "epoch": 1.9602995622339336, "grad_norm": 1.8690754175186157, "learning_rate": 0.00024627410001839916, "loss": 5.0996, "step": 325100 }, { "epoch": 1.9606010540152674, "grad_norm": 1.5027318000793457, "learning_rate": 0.00024625849454502804, "loss": 4.8635, "step": 325150 }, { "epoch": 1.9609025457966016, "grad_norm": 1.4707846641540527, "learning_rate": 0.00024624288730015667, "loss": 5.1693, "step": 325200 }, { "epoch": 1.9612040375779356, "grad_norm": 2.356832504272461, "learning_rate": 0.00024622727828407234, "loss": 5.2609, "step": 325250 }, { "epoch": 1.9615055293592696, "grad_norm": 1.0734165906906128, "learning_rate": 0.0002462116674970622, "loss": 5.0065, "step": 325300 }, { "epoch": 1.9618070211406038, "grad_norm": 1.1971843242645264, "learning_rate": 0.0002461960549394137, "loss": 4.5494, "step": 325350 }, { "epoch": 1.9621085129219378, "grad_norm": 1.1803638935089111, "learning_rate": 0.0002461804406114141, "loss": 5.0444, "step": 325400 }, { "epoch": 1.9624100047032718, "grad_norm": 1.9280285835266113, "learning_rate": 0.0002461648245133507, "loss": 5.3304, "step": 325450 }, { "epoch": 1.962711496484606, "grad_norm": 0.3966960608959198, "learning_rate": 0.00024614920664551096, "loss": 5.2321, "step": 325500 }, { "epoch": 1.9630129882659397, "grad_norm": 1.251265525817871, "learning_rate": 0.00024613358700818224, "loss": 4.7531, "step": 325550 }, { "epoch": 1.963314480047274, "grad_norm": 1.3336608409881592, "learning_rate": 0.00024611796560165215, "loss": 4.5618, "step": 325600 }, { "epoch": 1.963615971828608, "grad_norm": 0.6877729892730713, "learning_rate": 0.000246102342426208, "loss": 5.4998, "step": 325650 }, { "epoch": 1.963917463609942, "grad_norm": 1.0032466650009155, "learning_rate": 0.00024608671748213736, "loss": 4.6279, "step": 325700 }, { "epoch": 1.9642189553912761, "grad_norm": 1.2373701333999634, "learning_rate": 0.00024607109076972783, "loss": 4.4169, "step": 325750 }, { "epoch": 1.96452044717261, "grad_norm": 1.3634939193725586, "learning_rate": 0.00024605546228926694, "loss": 4.7484, "step": 325800 }, { "epoch": 1.964821938953944, "grad_norm": 1.4027611017227173, "learning_rate": 0.0002460398320410424, "loss": 5.2284, "step": 325850 }, { "epoch": 1.9651234307352783, "grad_norm": 1.385784387588501, "learning_rate": 0.0002460242000253418, "loss": 4.7949, "step": 325900 }, { "epoch": 1.965424922516612, "grad_norm": 1.7255228757858276, "learning_rate": 0.00024600856624245286, "loss": 4.8264, "step": 325950 }, { "epoch": 1.9657264142979463, "grad_norm": 1.4827659130096436, "learning_rate": 0.00024599293069266324, "loss": 5.4711, "step": 326000 }, { "epoch": 1.9660279060792802, "grad_norm": 1.699796438217163, "learning_rate": 0.00024597729337626066, "loss": 4.6952, "step": 326050 }, { "epoch": 1.9663293978606142, "grad_norm": 1.2679641246795654, "learning_rate": 0.00024596165429353303, "loss": 4.802, "step": 326100 }, { "epoch": 1.9666308896419484, "grad_norm": 1.2921478748321533, "learning_rate": 0.00024594601344476806, "loss": 5.131, "step": 326150 }, { "epoch": 1.9669323814232824, "grad_norm": 1.533624291419983, "learning_rate": 0.00024593037083025366, "loss": 5.0981, "step": 326200 }, { "epoch": 1.9672338732046164, "grad_norm": 1.4541741609573364, "learning_rate": 0.0002459147264502777, "loss": 4.928, "step": 326250 }, { "epoch": 1.9675353649859506, "grad_norm": 1.9890538454055786, "learning_rate": 0.000245899080305128, "loss": 4.7865, "step": 326300 }, { "epoch": 1.9678368567672844, "grad_norm": 1.6019926071166992, "learning_rate": 0.0002458834323950927, "loss": 4.9927, "step": 326350 }, { "epoch": 1.9681383485486186, "grad_norm": 0.642747700214386, "learning_rate": 0.00024586778272045956, "loss": 5.0331, "step": 326400 }, { "epoch": 1.9684398403299526, "grad_norm": 1.6422065496444702, "learning_rate": 0.00024585213128151676, "loss": 4.8661, "step": 326450 }, { "epoch": 1.9687413321112865, "grad_norm": 1.361725091934204, "learning_rate": 0.0002458364780785522, "loss": 5.1621, "step": 326500 }, { "epoch": 1.9690428238926208, "grad_norm": 1.786818504333496, "learning_rate": 0.00024582082311185407, "loss": 4.8169, "step": 326550 }, { "epoch": 1.9693443156739547, "grad_norm": 1.6480337381362915, "learning_rate": 0.0002458051663817105, "loss": 4.9248, "step": 326600 }, { "epoch": 1.9696458074552887, "grad_norm": 1.5459866523742676, "learning_rate": 0.0002457895078884095, "loss": 5.289, "step": 326650 }, { "epoch": 1.969947299236623, "grad_norm": 1.893025517463684, "learning_rate": 0.00024577384763223927, "loss": 4.9615, "step": 326700 }, { "epoch": 1.9702487910179567, "grad_norm": 2.3774709701538086, "learning_rate": 0.00024575818561348807, "loss": 5.1973, "step": 326750 }, { "epoch": 1.970550282799291, "grad_norm": 2.6167006492614746, "learning_rate": 0.00024574252183244414, "loss": 5.1368, "step": 326800 }, { "epoch": 1.9708517745806249, "grad_norm": 0.6278397440910339, "learning_rate": 0.0002457268562893957, "loss": 5.6008, "step": 326850 }, { "epoch": 1.9711532663619589, "grad_norm": 1.0976533889770508, "learning_rate": 0.00024571118898463113, "loss": 5.5851, "step": 326900 }, { "epoch": 1.971454758143293, "grad_norm": 1.7867629528045654, "learning_rate": 0.0002456955199184387, "loss": 5.2565, "step": 326950 }, { "epoch": 1.971756249924627, "grad_norm": 1.2073665857315063, "learning_rate": 0.00024567984909110673, "loss": 5.133, "step": 327000 }, { "epoch": 1.971756249924627, "eval_loss": 5.522180080413818, "eval_runtime": 39.0674, "eval_samples_per_second": 13.106, "eval_steps_per_second": 6.553, "eval_tts_loss": 7.321185657132254, "step": 327000 }, { "epoch": 1.972057741705961, "grad_norm": 1.3536081314086914, "learning_rate": 0.00024566417650292375, "loss": 5.1128, "step": 327050 }, { "epoch": 1.9723592334872952, "grad_norm": 1.3443080186843872, "learning_rate": 0.00024564850215417805, "loss": 4.8832, "step": 327100 }, { "epoch": 1.972660725268629, "grad_norm": 1.8214651346206665, "learning_rate": 0.00024563282604515824, "loss": 4.7449, "step": 327150 }, { "epoch": 1.9729622170499632, "grad_norm": 3.2200398445129395, "learning_rate": 0.0002456171481761527, "loss": 5.2432, "step": 327200 }, { "epoch": 1.9732637088312974, "grad_norm": 1.143525242805481, "learning_rate": 0.00024560146854745, "loss": 5.3743, "step": 327250 }, { "epoch": 1.9735652006126312, "grad_norm": 1.6780445575714111, "learning_rate": 0.00024558578715933873, "loss": 5.148, "step": 327300 }, { "epoch": 1.9738666923939654, "grad_norm": 0.3437446057796478, "learning_rate": 0.00024557010401210743, "loss": 4.6094, "step": 327350 }, { "epoch": 1.9741681841752994, "grad_norm": 1.3903502225875854, "learning_rate": 0.0002455544191060448, "loss": 5.2426, "step": 327400 }, { "epoch": 1.9744696759566334, "grad_norm": 1.2334507703781128, "learning_rate": 0.0002455387324414394, "loss": 4.6318, "step": 327450 }, { "epoch": 1.9747711677379676, "grad_norm": 1.338446021080017, "learning_rate": 0.00024552304401858, "loss": 5.0178, "step": 327500 }, { "epoch": 1.9750726595193016, "grad_norm": 1.1593304872512817, "learning_rate": 0.00024550735383775527, "loss": 5.1984, "step": 327550 }, { "epoch": 1.9753741513006355, "grad_norm": 1.6495189666748047, "learning_rate": 0.00024549166189925396, "loss": 4.6612, "step": 327600 }, { "epoch": 1.9756756430819697, "grad_norm": 1.7739098072052002, "learning_rate": 0.0002454759682033649, "loss": 5.002, "step": 327650 }, { "epoch": 1.9759771348633035, "grad_norm": 2.044337749481201, "learning_rate": 0.00024546027275037694, "loss": 5.1391, "step": 327700 }, { "epoch": 1.9762786266446377, "grad_norm": 1.2132036685943604, "learning_rate": 0.0002454445755405789, "loss": 5.2883, "step": 327750 }, { "epoch": 1.9765801184259717, "grad_norm": 0.5226259827613831, "learning_rate": 0.0002454288765742596, "loss": 5.1916, "step": 327800 }, { "epoch": 1.9768816102073057, "grad_norm": 1.1096580028533936, "learning_rate": 0.000245413175851708, "loss": 5.1236, "step": 327850 }, { "epoch": 1.97718310198864, "grad_norm": 1.4317206144332886, "learning_rate": 0.00024539747337321314, "loss": 4.9765, "step": 327900 }, { "epoch": 1.9774845937699739, "grad_norm": 1.683730125427246, "learning_rate": 0.00024538176913906383, "loss": 5.246, "step": 327950 }, { "epoch": 1.9777860855513079, "grad_norm": 1.367089867591858, "learning_rate": 0.00024536606314954924, "loss": 5.4093, "step": 328000 }, { "epoch": 1.978087577332642, "grad_norm": 2.3698620796203613, "learning_rate": 0.0002453503554049583, "loss": 4.9625, "step": 328050 }, { "epoch": 1.9783890691139758, "grad_norm": 2.9235408306121826, "learning_rate": 0.0002453346459055802, "loss": 5.3274, "step": 328100 }, { "epoch": 1.97869056089531, "grad_norm": 1.404961347579956, "learning_rate": 0.00024531893465170397, "loss": 5.5545, "step": 328150 }, { "epoch": 1.978992052676644, "grad_norm": 3.467008590698242, "learning_rate": 0.00024530322164361876, "loss": 5.0802, "step": 328200 }, { "epoch": 1.979293544457978, "grad_norm": 1.7361083030700684, "learning_rate": 0.0002452875068816138, "loss": 5.389, "step": 328250 }, { "epoch": 1.9795950362393122, "grad_norm": 1.967160940170288, "learning_rate": 0.0002452717903659782, "loss": 5.1193, "step": 328300 }, { "epoch": 1.9798965280206462, "grad_norm": 0.7665075659751892, "learning_rate": 0.00024525607209700134, "loss": 4.9664, "step": 328350 }, { "epoch": 1.9801980198019802, "grad_norm": 1.2102305889129639, "learning_rate": 0.00024524035207497234, "loss": 5.1722, "step": 328400 }, { "epoch": 1.9804995115833144, "grad_norm": 1.2171196937561035, "learning_rate": 0.0002452246303001806, "loss": 4.6484, "step": 328450 }, { "epoch": 1.9808010033646481, "grad_norm": 1.1782218217849731, "learning_rate": 0.00024520890677291545, "loss": 5.3138, "step": 328500 }, { "epoch": 1.9811024951459824, "grad_norm": 1.2948566675186157, "learning_rate": 0.0002451931814934662, "loss": 5.0881, "step": 328550 }, { "epoch": 1.9814039869273163, "grad_norm": 1.3308680057525635, "learning_rate": 0.0002451774544621223, "loss": 5.1603, "step": 328600 }, { "epoch": 1.9817054787086503, "grad_norm": 2.478628158569336, "learning_rate": 0.0002451617256791732, "loss": 5.3236, "step": 328650 }, { "epoch": 1.9820069704899845, "grad_norm": 0.8048896789550781, "learning_rate": 0.0002451459951449083, "loss": 4.514, "step": 328700 }, { "epoch": 1.9823084622713185, "grad_norm": 1.3090938329696655, "learning_rate": 0.0002451302628596172, "loss": 5.2445, "step": 328750 }, { "epoch": 1.9826099540526525, "grad_norm": 1.238068699836731, "learning_rate": 0.00024511452882358935, "loss": 4.6004, "step": 328800 }, { "epoch": 1.9829114458339867, "grad_norm": 2.244882583618164, "learning_rate": 0.00024509879303711435, "loss": 5.0168, "step": 328850 }, { "epoch": 1.9832129376153205, "grad_norm": 1.7731655836105347, "learning_rate": 0.0002450830555004817, "loss": 5.0185, "step": 328900 }, { "epoch": 1.9835144293966547, "grad_norm": 1.7328234910964966, "learning_rate": 0.0002450673162139811, "loss": 4.9786, "step": 328950 }, { "epoch": 1.9838159211779887, "grad_norm": 0.9473423361778259, "learning_rate": 0.00024505157517790236, "loss": 4.5921, "step": 329000 }, { "epoch": 1.9841174129593226, "grad_norm": 1.725036859512329, "learning_rate": 0.0002450358323925349, "loss": 5.3837, "step": 329050 }, { "epoch": 1.9844189047406569, "grad_norm": 1.3503272533416748, "learning_rate": 0.00024502008785816864, "loss": 5.1933, "step": 329100 }, { "epoch": 1.9847203965219908, "grad_norm": 1.9491267204284668, "learning_rate": 0.0002450043415750932, "loss": 5.0233, "step": 329150 }, { "epoch": 1.9850218883033248, "grad_norm": 1.5669664144515991, "learning_rate": 0.00024498859354359846, "loss": 4.9935, "step": 329200 }, { "epoch": 1.985323380084659, "grad_norm": 1.3805888891220093, "learning_rate": 0.0002449728437639742, "loss": 5.2695, "step": 329250 }, { "epoch": 1.9856248718659928, "grad_norm": 2.0664875507354736, "learning_rate": 0.0002449570922365103, "loss": 5.2353, "step": 329300 }, { "epoch": 1.985926363647327, "grad_norm": 0.7034314274787903, "learning_rate": 0.0002449413389614966, "loss": 5.4351, "step": 329350 }, { "epoch": 1.986227855428661, "grad_norm": 1.4780690670013428, "learning_rate": 0.00024492558393922304, "loss": 4.4792, "step": 329400 }, { "epoch": 1.986529347209995, "grad_norm": 1.9459874629974365, "learning_rate": 0.0002449098271699796, "loss": 4.7724, "step": 329450 }, { "epoch": 1.9868308389913292, "grad_norm": 0.8772138953208923, "learning_rate": 0.00024489406865405624, "loss": 5.3217, "step": 329500 }, { "epoch": 1.9871323307726632, "grad_norm": 1.4958298206329346, "learning_rate": 0.000244878308391743, "loss": 5.2627, "step": 329550 }, { "epoch": 1.9874338225539971, "grad_norm": 1.4726561307907104, "learning_rate": 0.00024486254638332984, "loss": 5.3634, "step": 329600 }, { "epoch": 1.9877353143353313, "grad_norm": 2.0098509788513184, "learning_rate": 0.0002448467826291069, "loss": 5.1291, "step": 329650 }, { "epoch": 1.9880368061166651, "grad_norm": 1.359298586845398, "learning_rate": 0.00024483101712936427, "loss": 5.074, "step": 329700 }, { "epoch": 1.9883382978979993, "grad_norm": 1.5672513246536255, "learning_rate": 0.0002448152498843921, "loss": 4.9899, "step": 329750 }, { "epoch": 1.9886397896793333, "grad_norm": 1.5954530239105225, "learning_rate": 0.0002447994808944806, "loss": 4.4688, "step": 329800 }, { "epoch": 1.9889412814606673, "grad_norm": 0.8863474130630493, "learning_rate": 0.0002447837101599199, "loss": 4.9519, "step": 329850 }, { "epoch": 1.9892427732420015, "grad_norm": 1.7000575065612793, "learning_rate": 0.0002447679376810003, "loss": 5.1411, "step": 329900 }, { "epoch": 1.9895442650233355, "grad_norm": 1.9952645301818848, "learning_rate": 0.0002447521634580121, "loss": 5.1881, "step": 329950 }, { "epoch": 1.9898457568046695, "grad_norm": 1.6158475875854492, "learning_rate": 0.0002447363874912455, "loss": 5.3068, "step": 330000 }, { "epoch": 1.9898457568046695, "eval_loss": 5.504086494445801, "eval_runtime": 38.9902, "eval_samples_per_second": 13.131, "eval_steps_per_second": 6.566, "eval_tts_loss": 7.355676154128805, "step": 330000 }, { "epoch": 1.9901472485860037, "grad_norm": 1.5819543600082397, "learning_rate": 0.00024472060978099087, "loss": 4.748, "step": 330050 }, { "epoch": 1.9904487403673374, "grad_norm": 1.0031746625900269, "learning_rate": 0.00024470483032753854, "loss": 4.7589, "step": 330100 }, { "epoch": 1.9907502321486716, "grad_norm": 1.1144102811813354, "learning_rate": 0.0002446890491311791, "loss": 4.6241, "step": 330150 }, { "epoch": 1.9910517239300056, "grad_norm": 1.2664459943771362, "learning_rate": 0.0002446732661922027, "loss": 5.1826, "step": 330200 }, { "epoch": 1.9913532157113396, "grad_norm": 1.956750750541687, "learning_rate": 0.0002446574815109, "loss": 4.9938, "step": 330250 }, { "epoch": 1.9916547074926738, "grad_norm": 1.9475218057632446, "learning_rate": 0.0002446416950875615, "loss": 5.2235, "step": 330300 }, { "epoch": 1.9919561992740078, "grad_norm": 2.0800089836120605, "learning_rate": 0.00024462590692247754, "loss": 5.0695, "step": 330350 }, { "epoch": 1.9922576910553418, "grad_norm": 0.4825240671634674, "learning_rate": 0.0002446101170159389, "loss": 4.967, "step": 330400 }, { "epoch": 1.992559182836676, "grad_norm": 0.7107057571411133, "learning_rate": 0.000244594325368236, "loss": 4.9201, "step": 330450 }, { "epoch": 1.9928606746180098, "grad_norm": 1.5455878973007202, "learning_rate": 0.00024457853197965956, "loss": 5.1792, "step": 330500 }, { "epoch": 1.993162166399344, "grad_norm": 1.4815753698349, "learning_rate": 0.0002445627368505002, "loss": 4.844, "step": 330550 }, { "epoch": 1.993463658180678, "grad_norm": 0.7781838178634644, "learning_rate": 0.00024454693998104867, "loss": 4.753, "step": 330600 }, { "epoch": 1.993765149962012, "grad_norm": 1.4944474697113037, "learning_rate": 0.0002445311413715956, "loss": 5.259, "step": 330650 }, { "epoch": 1.9940666417433461, "grad_norm": 1.3537507057189941, "learning_rate": 0.00024451534102243176, "loss": 4.9566, "step": 330700 }, { "epoch": 1.9943681335246801, "grad_norm": 1.2680398225784302, "learning_rate": 0.000244499538933848, "loss": 5.1747, "step": 330750 }, { "epoch": 1.994669625306014, "grad_norm": 1.9716559648513794, "learning_rate": 0.000244483735106135, "loss": 5.2575, "step": 330800 }, { "epoch": 1.9949711170873483, "grad_norm": 0.9071578979492188, "learning_rate": 0.00024446792953958376, "loss": 4.5816, "step": 330850 }, { "epoch": 1.995272608868682, "grad_norm": 1.5726099014282227, "learning_rate": 0.00024445212223448505, "loss": 5.5309, "step": 330900 }, { "epoch": 1.9955741006500163, "grad_norm": 1.4358482360839844, "learning_rate": 0.0002444363131911298, "loss": 4.7889, "step": 330950 }, { "epoch": 1.9958755924313505, "grad_norm": 1.3487398624420166, "learning_rate": 0.00024442050240980904, "loss": 5.1681, "step": 331000 }, { "epoch": 1.9961770842126842, "grad_norm": 1.886466145515442, "learning_rate": 0.0002444046898908137, "loss": 5.1083, "step": 331050 }, { "epoch": 1.9964785759940185, "grad_norm": 1.3170462846755981, "learning_rate": 0.0002443888756344347, "loss": 4.8386, "step": 331100 }, { "epoch": 1.9967800677753524, "grad_norm": 1.2522093057632446, "learning_rate": 0.00024437305964096316, "loss": 4.7822, "step": 331150 }, { "epoch": 1.9970815595566864, "grad_norm": 1.3258923292160034, "learning_rate": 0.00024435724191069016, "loss": 5.1445, "step": 331200 }, { "epoch": 1.9973830513380206, "grad_norm": 1.7717766761779785, "learning_rate": 0.0002443414224439067, "loss": 5.1714, "step": 331250 }, { "epoch": 1.9976845431193546, "grad_norm": 0.5128180980682373, "learning_rate": 0.00024432560124090406, "loss": 4.9895, "step": 331300 }, { "epoch": 1.9979860349006886, "grad_norm": 1.8001878261566162, "learning_rate": 0.00024430977830197333, "loss": 4.9349, "step": 331350 }, { "epoch": 1.9982875266820228, "grad_norm": 1.360898494720459, "learning_rate": 0.0002442939536274057, "loss": 5.3935, "step": 331400 }, { "epoch": 1.9985890184633566, "grad_norm": 1.6806880235671997, "learning_rate": 0.00024427812721749246, "loss": 5.1974, "step": 331450 }, { "epoch": 1.9988905102446908, "grad_norm": 1.3954581022262573, "learning_rate": 0.0002442622990725248, "loss": 5.3312, "step": 331500 }, { "epoch": 1.9991920020260248, "grad_norm": 1.3704426288604736, "learning_rate": 0.00024424646919279404, "loss": 4.6599, "step": 331550 }, { "epoch": 1.9994934938073587, "grad_norm": 1.746286392211914, "learning_rate": 0.0002442306375785915, "loss": 5.3665, "step": 331600 }, { "epoch": 1.999794985588693, "grad_norm": 1.414405107498169, "learning_rate": 0.0002442148042302086, "loss": 4.8271, "step": 331650 }, { "epoch": 2.0000964773700267, "grad_norm": 1.6569030284881592, "learning_rate": 0.0002441989691479366, "loss": 5.352, "step": 331700 }, { "epoch": 2.000397969151361, "grad_norm": 0.8414286971092224, "learning_rate": 0.0002441831323320671, "loss": 4.8531, "step": 331750 }, { "epoch": 2.000699460932695, "grad_norm": 1.4081356525421143, "learning_rate": 0.0002441672937828914, "loss": 5.0189, "step": 331800 }, { "epoch": 2.001000952714029, "grad_norm": 0.9003428220748901, "learning_rate": 0.00024415145350070104, "loss": 5.1314, "step": 331850 }, { "epoch": 2.001302444495363, "grad_norm": 1.898199200630188, "learning_rate": 0.00024413561148578755, "loss": 4.8965, "step": 331900 }, { "epoch": 2.0016039362766973, "grad_norm": 2.113023519515991, "learning_rate": 0.00024411976773844245, "loss": 5.1402, "step": 331950 }, { "epoch": 2.001905428058031, "grad_norm": 1.4169294834136963, "learning_rate": 0.00024410392225895737, "loss": 5.2792, "step": 332000 }, { "epoch": 2.0022069198393653, "grad_norm": 1.2290312051773071, "learning_rate": 0.0002440880750476239, "loss": 4.5959, "step": 332050 }, { "epoch": 2.002508411620699, "grad_norm": 1.26749587059021, "learning_rate": 0.0002440722261047336, "loss": 5.1323, "step": 332100 }, { "epoch": 2.0028099034020332, "grad_norm": 2.005586624145508, "learning_rate": 0.00024405637543057834, "loss": 5.2652, "step": 332150 }, { "epoch": 2.0031113951833674, "grad_norm": 1.5847278833389282, "learning_rate": 0.00024404052302544967, "loss": 4.9108, "step": 332200 }, { "epoch": 2.003412886964701, "grad_norm": 1.6386466026306152, "learning_rate": 0.00024402466888963936, "loss": 5.1444, "step": 332250 }, { "epoch": 2.0037143787460354, "grad_norm": 1.782285213470459, "learning_rate": 0.00024400881302343922, "loss": 5.5845, "step": 332300 }, { "epoch": 2.0040158705273696, "grad_norm": 1.0377756357192993, "learning_rate": 0.00024399295542714103, "loss": 4.986, "step": 332350 }, { "epoch": 2.0043173623087034, "grad_norm": 1.9447475671768188, "learning_rate": 0.0002439770961010366, "loss": 5.127, "step": 332400 }, { "epoch": 2.0046188540900376, "grad_norm": 1.5627691745758057, "learning_rate": 0.00024396123504541787, "loss": 5.0553, "step": 332450 }, { "epoch": 2.0049203458713714, "grad_norm": 2.5723843574523926, "learning_rate": 0.0002439453722605767, "loss": 4.6752, "step": 332500 }, { "epoch": 2.0052218376527056, "grad_norm": 1.974170207977295, "learning_rate": 0.00024392950774680496, "loss": 5.2318, "step": 332550 }, { "epoch": 2.0055233294340398, "grad_norm": 1.5254700183868408, "learning_rate": 0.00024391364150439472, "loss": 4.8217, "step": 332600 }, { "epoch": 2.0058248212153735, "grad_norm": 1.4324785470962524, "learning_rate": 0.00024389777353363794, "loss": 4.9474, "step": 332650 }, { "epoch": 2.0061263129967077, "grad_norm": 2.005230188369751, "learning_rate": 0.0002438819038348266, "loss": 5.2739, "step": 332700 }, { "epoch": 2.006427804778042, "grad_norm": 1.238968849182129, "learning_rate": 0.00024386603240825287, "loss": 4.8026, "step": 332750 }, { "epoch": 2.0067292965593757, "grad_norm": 1.3958860635757446, "learning_rate": 0.00024385015925420872, "loss": 4.6606, "step": 332800 }, { "epoch": 2.00703078834071, "grad_norm": 0.951393187046051, "learning_rate": 0.0002438342843729863, "loss": 4.9078, "step": 332850 }, { "epoch": 2.0073322801220437, "grad_norm": 1.3410868644714355, "learning_rate": 0.00024381840776487783, "loss": 5.1894, "step": 332900 }, { "epoch": 2.007633771903378, "grad_norm": 2.1281204223632812, "learning_rate": 0.00024380252943017542, "loss": 4.6413, "step": 332950 }, { "epoch": 2.007935263684712, "grad_norm": 1.2036950588226318, "learning_rate": 0.00024378664936917132, "loss": 5.0972, "step": 333000 }, { "epoch": 2.007935263684712, "eval_loss": 5.528026103973389, "eval_runtime": 39.004, "eval_samples_per_second": 13.127, "eval_steps_per_second": 6.563, "eval_tts_loss": 7.410113661016821, "step": 333000 }, { "epoch": 2.008236755466046, "grad_norm": 1.6456577777862549, "learning_rate": 0.00024377076758215778, "loss": 5.0718, "step": 333050 }, { "epoch": 2.00853824724738, "grad_norm": 0.3984028995037079, "learning_rate": 0.00024375488406942706, "loss": 5.5015, "step": 333100 }, { "epoch": 2.0088397390287143, "grad_norm": 1.7437655925750732, "learning_rate": 0.00024373899883127153, "loss": 5.2633, "step": 333150 }, { "epoch": 2.009141230810048, "grad_norm": 1.291567087173462, "learning_rate": 0.00024372311186798352, "loss": 5.1033, "step": 333200 }, { "epoch": 2.0094427225913822, "grad_norm": 1.6197173595428467, "learning_rate": 0.00024370722317985528, "loss": 5.0914, "step": 333250 }, { "epoch": 2.009744214372716, "grad_norm": 2.687575340270996, "learning_rate": 0.00024369133276717946, "loss": 4.9427, "step": 333300 }, { "epoch": 2.01004570615405, "grad_norm": 1.521823525428772, "learning_rate": 0.0002436754406302483, "loss": 5.2217, "step": 333350 }, { "epoch": 2.0103471979353844, "grad_norm": 3.0102925300598145, "learning_rate": 0.0002436595467693543, "loss": 5.1326, "step": 333400 }, { "epoch": 2.010648689716718, "grad_norm": 1.774237036705017, "learning_rate": 0.00024364365118479008, "loss": 5.1047, "step": 333450 }, { "epoch": 2.0109501814980524, "grad_norm": 2.0167436599731445, "learning_rate": 0.00024362775387684805, "loss": 5.4846, "step": 333500 }, { "epoch": 2.0112516732793866, "grad_norm": 2.6191301345825195, "learning_rate": 0.00024361185484582083, "loss": 5.259, "step": 333550 }, { "epoch": 2.0115531650607203, "grad_norm": 1.3655567169189453, "learning_rate": 0.00024359595409200103, "loss": 4.9515, "step": 333600 }, { "epoch": 2.0118546568420546, "grad_norm": 1.3076293468475342, "learning_rate": 0.00024358005161568126, "loss": 5.0419, "step": 333650 }, { "epoch": 2.0121561486233883, "grad_norm": 1.4695441722869873, "learning_rate": 0.00024356414741715418, "loss": 5.1791, "step": 333700 }, { "epoch": 2.0124576404047225, "grad_norm": 1.7623860836029053, "learning_rate": 0.00024354824149671247, "loss": 4.6837, "step": 333750 }, { "epoch": 2.0127591321860567, "grad_norm": 1.9011040925979614, "learning_rate": 0.00024353233385464892, "loss": 4.8078, "step": 333800 }, { "epoch": 2.0130606239673905, "grad_norm": 1.466750979423523, "learning_rate": 0.00024351642449125618, "loss": 4.9183, "step": 333850 }, { "epoch": 2.0133621157487247, "grad_norm": 0.8389539122581482, "learning_rate": 0.00024350051340682713, "loss": 4.7717, "step": 333900 }, { "epoch": 2.013663607530059, "grad_norm": 1.1686309576034546, "learning_rate": 0.00024348460060165457, "loss": 4.9231, "step": 333950 }, { "epoch": 2.0139650993113927, "grad_norm": 1.0705560445785522, "learning_rate": 0.00024346868607603135, "loss": 5.0373, "step": 334000 }, { "epoch": 2.014266591092727, "grad_norm": 1.9173495769500732, "learning_rate": 0.00024345276983025037, "loss": 4.8048, "step": 334050 }, { "epoch": 2.0145680828740606, "grad_norm": 1.875523567199707, "learning_rate": 0.0002434368518646045, "loss": 4.9787, "step": 334100 }, { "epoch": 2.014869574655395, "grad_norm": 1.5399749279022217, "learning_rate": 0.00024342093217938668, "loss": 4.669, "step": 334150 }, { "epoch": 2.015171066436729, "grad_norm": 1.193884253501892, "learning_rate": 0.00024340501077488996, "loss": 4.9763, "step": 334200 }, { "epoch": 2.015472558218063, "grad_norm": 1.4710352420806885, "learning_rate": 0.00024338908765140734, "loss": 4.9475, "step": 334250 }, { "epoch": 2.015774049999397, "grad_norm": 1.6611073017120361, "learning_rate": 0.0002433731628092318, "loss": 4.8945, "step": 334300 }, { "epoch": 2.0160755417807312, "grad_norm": 0.9451133012771606, "learning_rate": 0.00024335723624865644, "loss": 4.6913, "step": 334350 }, { "epoch": 2.016377033562065, "grad_norm": 1.83916175365448, "learning_rate": 0.0002433413079699744, "loss": 5.2311, "step": 334400 }, { "epoch": 2.016678525343399, "grad_norm": 1.5353672504425049, "learning_rate": 0.00024332537797347876, "loss": 5.0509, "step": 334450 }, { "epoch": 2.016980017124733, "grad_norm": 0.879795491695404, "learning_rate": 0.00024330944625946274, "loss": 5.1533, "step": 334500 }, { "epoch": 2.017281508906067, "grad_norm": 1.9128109216690063, "learning_rate": 0.00024329351282821954, "loss": 4.8999, "step": 334550 }, { "epoch": 2.0175830006874014, "grad_norm": 0.6216639876365662, "learning_rate": 0.00024327757768004233, "loss": 5.1364, "step": 334600 }, { "epoch": 2.017884492468735, "grad_norm": 2.555328845977783, "learning_rate": 0.00024326164081522447, "loss": 5.2441, "step": 334650 }, { "epoch": 2.0181859842500693, "grad_norm": 1.253814697265625, "learning_rate": 0.00024324570223405916, "loss": 4.8815, "step": 334700 }, { "epoch": 2.0184874760314035, "grad_norm": 1.3520307540893555, "learning_rate": 0.0002432297619368398, "loss": 5.264, "step": 334750 }, { "epoch": 2.0187889678127373, "grad_norm": 1.2709214687347412, "learning_rate": 0.0002432138199238597, "loss": 4.9004, "step": 334800 }, { "epoch": 2.0190904595940715, "grad_norm": 1.7352782487869263, "learning_rate": 0.00024319787619541224, "loss": 5.3156, "step": 334850 }, { "epoch": 2.0193919513754057, "grad_norm": 1.4694716930389404, "learning_rate": 0.00024318193075179088, "loss": 4.9313, "step": 334900 }, { "epoch": 2.0196934431567395, "grad_norm": 2.2183237075805664, "learning_rate": 0.00024316598359328909, "loss": 4.8759, "step": 334950 }, { "epoch": 2.0199949349380737, "grad_norm": 2.108157157897949, "learning_rate": 0.00024315003472020032, "loss": 5.1531, "step": 335000 }, { "epoch": 2.0202964267194075, "grad_norm": 2.420886278152466, "learning_rate": 0.0002431340841328181, "loss": 4.8613, "step": 335050 }, { "epoch": 2.0205979185007417, "grad_norm": 2.4299488067626953, "learning_rate": 0.000243118131831436, "loss": 5.1061, "step": 335100 }, { "epoch": 2.020899410282076, "grad_norm": 0.541670560836792, "learning_rate": 0.0002431021778163475, "loss": 4.9921, "step": 335150 }, { "epoch": 2.0212009020634096, "grad_norm": 1.4949275255203247, "learning_rate": 0.00024308622208784635, "loss": 4.878, "step": 335200 }, { "epoch": 2.021502393844744, "grad_norm": 0.7030535340309143, "learning_rate": 0.00024307026464622606, "loss": 5.2664, "step": 335250 }, { "epoch": 2.021803885626078, "grad_norm": 1.635623812675476, "learning_rate": 0.0002430543054917804, "loss": 5.0448, "step": 335300 }, { "epoch": 2.022105377407412, "grad_norm": 1.8187921047210693, "learning_rate": 0.000243038344624803, "loss": 4.6226, "step": 335350 }, { "epoch": 2.022406869188746, "grad_norm": 1.3490924835205078, "learning_rate": 0.0002430223820455877, "loss": 5.2936, "step": 335400 }, { "epoch": 2.0227083609700798, "grad_norm": 1.6587767601013184, "learning_rate": 0.00024300641775442817, "loss": 5.3868, "step": 335450 }, { "epoch": 2.023009852751414, "grad_norm": 2.1085894107818604, "learning_rate": 0.00024299045175161832, "loss": 5.2582, "step": 335500 }, { "epoch": 2.023311344532748, "grad_norm": 1.7304339408874512, "learning_rate": 0.0002429744840374518, "loss": 4.6565, "step": 335550 }, { "epoch": 2.023612836314082, "grad_norm": 1.0755174160003662, "learning_rate": 0.00024295851461222267, "loss": 5.1643, "step": 335600 }, { "epoch": 2.023914328095416, "grad_norm": 1.432179570198059, "learning_rate": 0.0002429425434762247, "loss": 4.9634, "step": 335650 }, { "epoch": 2.0242158198767504, "grad_norm": 1.3000274896621704, "learning_rate": 0.00024292657062975186, "loss": 5.4646, "step": 335700 }, { "epoch": 2.024517311658084, "grad_norm": 1.811636209487915, "learning_rate": 0.0002429105960730981, "loss": 4.792, "step": 335750 }, { "epoch": 2.0248188034394183, "grad_norm": 1.114012360572815, "learning_rate": 0.00024289461980655738, "loss": 5.1003, "step": 335800 }, { "epoch": 2.025120295220752, "grad_norm": 1.7135823965072632, "learning_rate": 0.0002428786418304238, "loss": 4.7013, "step": 335850 }, { "epoch": 2.0254217870020863, "grad_norm": 0.8152114152908325, "learning_rate": 0.00024286266214499134, "loss": 5.0089, "step": 335900 }, { "epoch": 2.0257232787834205, "grad_norm": 1.2827059030532837, "learning_rate": 0.0002428466807505541, "loss": 5.3035, "step": 335950 }, { "epoch": 2.0260247705647543, "grad_norm": 2.214165687561035, "learning_rate": 0.0002428306976474062, "loss": 4.7955, "step": 336000 }, { "epoch": 2.0260247705647543, "eval_loss": 5.4919939041137695, "eval_runtime": 39.2518, "eval_samples_per_second": 13.044, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.3601247797010485, "step": 336000 }, { "epoch": 2.0263262623460885, "grad_norm": 1.9194899797439575, "learning_rate": 0.00024281471283584178, "loss": 4.8784, "step": 336050 }, { "epoch": 2.0266277541274227, "grad_norm": 1.6035815477371216, "learning_rate": 0.0002427987263161551, "loss": 4.8181, "step": 336100 }, { "epoch": 2.0269292459087564, "grad_norm": 1.3755584955215454, "learning_rate": 0.0002427827380886402, "loss": 5.381, "step": 336150 }, { "epoch": 2.0272307376900907, "grad_norm": 1.2260661125183105, "learning_rate": 0.00024276674815359143, "loss": 4.4018, "step": 336200 }, { "epoch": 2.0275322294714244, "grad_norm": 1.9566233158111572, "learning_rate": 0.00024275075651130306, "loss": 5.0091, "step": 336250 }, { "epoch": 2.0278337212527586, "grad_norm": 1.301400065422058, "learning_rate": 0.00024273476316206938, "loss": 5.1639, "step": 336300 }, { "epoch": 2.028135213034093, "grad_norm": 1.8428221940994263, "learning_rate": 0.00024271876810618472, "loss": 5.0666, "step": 336350 }, { "epoch": 2.0284367048154266, "grad_norm": 2.13749361038208, "learning_rate": 0.00024270277134394346, "loss": 4.7776, "step": 336400 }, { "epoch": 2.028738196596761, "grad_norm": 2.112407922744751, "learning_rate": 0.00024268677287564, "loss": 5.1628, "step": 336450 }, { "epoch": 2.029039688378095, "grad_norm": 1.5653398036956787, "learning_rate": 0.0002426707727015687, "loss": 5.2013, "step": 336500 }, { "epoch": 2.0293411801594288, "grad_norm": 1.4763736724853516, "learning_rate": 0.00024265477082202413, "loss": 5.2974, "step": 336550 }, { "epoch": 2.029642671940763, "grad_norm": 0.36466941237449646, "learning_rate": 0.0002426387672373007, "loss": 5.169, "step": 336600 }, { "epoch": 2.0299441637220967, "grad_norm": 2.896078586578369, "learning_rate": 0.000242622761947693, "loss": 5.11, "step": 336650 }, { "epoch": 2.030245655503431, "grad_norm": 1.6663657426834106, "learning_rate": 0.0002426067549534955, "loss": 5.0947, "step": 336700 }, { "epoch": 2.030547147284765, "grad_norm": 1.2640280723571777, "learning_rate": 0.00024259074625500283, "loss": 5.1256, "step": 336750 }, { "epoch": 2.030848639066099, "grad_norm": 1.463811993598938, "learning_rate": 0.00024257473585250966, "loss": 5.1806, "step": 336800 }, { "epoch": 2.031150130847433, "grad_norm": 1.3453940153121948, "learning_rate": 0.00024255872374631055, "loss": 5.0263, "step": 336850 }, { "epoch": 2.0314516226287673, "grad_norm": 0.8998779654502869, "learning_rate": 0.00024254270993670022, "loss": 5.1046, "step": 336900 }, { "epoch": 2.031753114410101, "grad_norm": 1.4282037019729614, "learning_rate": 0.0002425266944239734, "loss": 5.0119, "step": 336950 }, { "epoch": 2.0320546061914353, "grad_norm": 1.8954482078552246, "learning_rate": 0.0002425106772084248, "loss": 4.6996, "step": 337000 }, { "epoch": 2.032356097972769, "grad_norm": 1.4013738632202148, "learning_rate": 0.00024249465829034918, "loss": 5.0824, "step": 337050 }, { "epoch": 2.0326575897541033, "grad_norm": 1.2706763744354248, "learning_rate": 0.0002424786376700414, "loss": 4.7002, "step": 337100 }, { "epoch": 2.0329590815354375, "grad_norm": 2.781315803527832, "learning_rate": 0.00024246261534779625, "loss": 4.9809, "step": 337150 }, { "epoch": 2.0332605733167712, "grad_norm": 1.103316307067871, "learning_rate": 0.0002424465913239086, "loss": 4.7771, "step": 337200 }, { "epoch": 2.0335620650981054, "grad_norm": 2.341132402420044, "learning_rate": 0.00024243056559867341, "loss": 4.797, "step": 337250 }, { "epoch": 2.0338635568794396, "grad_norm": 0.6615859866142273, "learning_rate": 0.00024241453817238558, "loss": 4.8485, "step": 337300 }, { "epoch": 2.0341650486607734, "grad_norm": 1.3061866760253906, "learning_rate": 0.00024239850904533998, "loss": 5.288, "step": 337350 }, { "epoch": 2.0344665404421076, "grad_norm": 1.1687562465667725, "learning_rate": 0.00024238247821783173, "loss": 4.9465, "step": 337400 }, { "epoch": 2.0347680322234414, "grad_norm": 1.0054051876068115, "learning_rate": 0.00024236644569015583, "loss": 4.8447, "step": 337450 }, { "epoch": 2.0350695240047756, "grad_norm": 2.1306564807891846, "learning_rate": 0.00024235041146260724, "loss": 5.0836, "step": 337500 }, { "epoch": 2.03537101578611, "grad_norm": 1.8811393976211548, "learning_rate": 0.0002423343755354812, "loss": 4.6569, "step": 337550 }, { "epoch": 2.0356725075674436, "grad_norm": 1.5238585472106934, "learning_rate": 0.0002423183379090727, "loss": 4.961, "step": 337600 }, { "epoch": 2.0359739993487778, "grad_norm": 1.4898931980133057, "learning_rate": 0.00024230229858367695, "loss": 4.8181, "step": 337650 }, { "epoch": 2.036275491130112, "grad_norm": 1.758607029914856, "learning_rate": 0.0002422862575595891, "loss": 5.1634, "step": 337700 }, { "epoch": 2.0365769829114457, "grad_norm": 1.2400190830230713, "learning_rate": 0.00024227021483710437, "loss": 5.041, "step": 337750 }, { "epoch": 2.03687847469278, "grad_norm": 2.4299190044403076, "learning_rate": 0.00024225417041651807, "loss": 5.4228, "step": 337800 }, { "epoch": 2.0371799664741137, "grad_norm": 0.8738823533058167, "learning_rate": 0.00024223812429812535, "loss": 4.9989, "step": 337850 }, { "epoch": 2.037481458255448, "grad_norm": 1.5127989053726196, "learning_rate": 0.00024222207648222165, "loss": 4.9455, "step": 337900 }, { "epoch": 2.037782950036782, "grad_norm": 2.890824556350708, "learning_rate": 0.00024220602696910213, "loss": 4.8617, "step": 337950 }, { "epoch": 2.038084441818116, "grad_norm": 1.3592301607131958, "learning_rate": 0.00024218997575906238, "loss": 4.5959, "step": 338000 }, { "epoch": 2.03838593359945, "grad_norm": 1.2957932949066162, "learning_rate": 0.00024217392285239768, "loss": 5.044, "step": 338050 }, { "epoch": 2.0386874253807843, "grad_norm": 0.4316710829734802, "learning_rate": 0.0002421578682494034, "loss": 4.8371, "step": 338100 }, { "epoch": 2.038988917162118, "grad_norm": 0.7674546241760254, "learning_rate": 0.0002421418119503751, "loss": 5.284, "step": 338150 }, { "epoch": 2.0392904089434523, "grad_norm": 1.049466609954834, "learning_rate": 0.00024212575395560828, "loss": 4.8378, "step": 338200 }, { "epoch": 2.039591900724786, "grad_norm": 1.6817651987075806, "learning_rate": 0.00024210969426539838, "loss": 5.3342, "step": 338250 }, { "epoch": 2.0398933925061202, "grad_norm": 1.466780662536621, "learning_rate": 0.00024209363288004105, "loss": 5.3492, "step": 338300 }, { "epoch": 2.0401948842874544, "grad_norm": 1.234505534172058, "learning_rate": 0.0002420775697998318, "loss": 5.4131, "step": 338350 }, { "epoch": 2.040496376068788, "grad_norm": 1.8259859085083008, "learning_rate": 0.00024206150502506627, "loss": 5.0292, "step": 338400 }, { "epoch": 2.0407978678501224, "grad_norm": 1.1903996467590332, "learning_rate": 0.00024204543855604015, "loss": 5.1462, "step": 338450 }, { "epoch": 2.0410993596314566, "grad_norm": 2.0119662284851074, "learning_rate": 0.00024202937039304907, "loss": 5.1697, "step": 338500 }, { "epoch": 2.0414008514127904, "grad_norm": 1.2040048837661743, "learning_rate": 0.00024201330053638876, "loss": 4.9399, "step": 338550 }, { "epoch": 2.0417023431941246, "grad_norm": 1.7193377017974854, "learning_rate": 0.00024199722898635503, "loss": 4.9465, "step": 338600 }, { "epoch": 2.042003834975459, "grad_norm": 1.4101430177688599, "learning_rate": 0.00024198115574324347, "loss": 4.8977, "step": 338650 }, { "epoch": 2.0423053267567925, "grad_norm": 1.8116579055786133, "learning_rate": 0.0002419650808073501, "loss": 5.0283, "step": 338700 }, { "epoch": 2.0426068185381268, "grad_norm": 2.3936116695404053, "learning_rate": 0.0002419490041789706, "loss": 5.1465, "step": 338750 }, { "epoch": 2.0429083103194605, "grad_norm": 2.166867256164551, "learning_rate": 0.00024193292585840095, "loss": 5.1934, "step": 338800 }, { "epoch": 2.0432098021007947, "grad_norm": 1.308333158493042, "learning_rate": 0.00024191684584593692, "loss": 4.7144, "step": 338850 }, { "epoch": 2.043511293882129, "grad_norm": 0.4594177007675171, "learning_rate": 0.00024190076414187458, "loss": 4.9003, "step": 338900 }, { "epoch": 2.0438127856634627, "grad_norm": 1.7457382678985596, "learning_rate": 0.00024188468074650977, "loss": 5.3823, "step": 338950 }, { "epoch": 2.044114277444797, "grad_norm": 1.7317363023757935, "learning_rate": 0.00024186859566013857, "loss": 5.5262, "step": 339000 }, { "epoch": 2.044114277444797, "eval_loss": 5.484551429748535, "eval_runtime": 39.0756, "eval_samples_per_second": 13.103, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.322540773946743, "step": 339000 }, { "epoch": 2.044415769226131, "grad_norm": 2.368875026702881, "learning_rate": 0.000241852508883057, "loss": 4.862, "step": 339050 }, { "epoch": 2.044717261007465, "grad_norm": 1.8905293941497803, "learning_rate": 0.00024183642041556105, "loss": 5.1176, "step": 339100 }, { "epoch": 2.045018752788799, "grad_norm": 2.123636484146118, "learning_rate": 0.00024182033025794684, "loss": 5.2379, "step": 339150 }, { "epoch": 2.045320244570133, "grad_norm": 1.599198579788208, "learning_rate": 0.0002418042384105105, "loss": 5.4265, "step": 339200 }, { "epoch": 2.045621736351467, "grad_norm": 1.3515197038650513, "learning_rate": 0.00024178814487354813, "loss": 4.9662, "step": 339250 }, { "epoch": 2.0459232281328013, "grad_norm": 1.5276049375534058, "learning_rate": 0.00024177204964735596, "loss": 4.798, "step": 339300 }, { "epoch": 2.046224719914135, "grad_norm": 0.6578608751296997, "learning_rate": 0.0002417559527322302, "loss": 5.1959, "step": 339350 }, { "epoch": 2.046526211695469, "grad_norm": 0.9256620407104492, "learning_rate": 0.00024173985412846707, "loss": 5.2393, "step": 339400 }, { "epoch": 2.0468277034768034, "grad_norm": 1.4360705614089966, "learning_rate": 0.00024172375383636283, "loss": 4.7462, "step": 339450 }, { "epoch": 2.047129195258137, "grad_norm": 2.571544647216797, "learning_rate": 0.00024170765185621383, "loss": 4.9885, "step": 339500 }, { "epoch": 2.0474306870394714, "grad_norm": 1.633650302886963, "learning_rate": 0.00024169154818831631, "loss": 5.2223, "step": 339550 }, { "epoch": 2.047732178820805, "grad_norm": 1.1593573093414307, "learning_rate": 0.00024167544283296673, "loss": 5.1707, "step": 339600 }, { "epoch": 2.0480336706021394, "grad_norm": 1.6555681228637695, "learning_rate": 0.00024165933579046146, "loss": 5.122, "step": 339650 }, { "epoch": 2.0483351623834736, "grad_norm": 2.259857416152954, "learning_rate": 0.0002416432270610969, "loss": 5.2188, "step": 339700 }, { "epoch": 2.0486366541648073, "grad_norm": 1.377581238746643, "learning_rate": 0.00024162711664516958, "loss": 5.0613, "step": 339750 }, { "epoch": 2.0489381459461415, "grad_norm": 0.9765885472297668, "learning_rate": 0.0002416110045429759, "loss": 4.7421, "step": 339800 }, { "epoch": 2.0492396377274757, "grad_norm": 1.4030851125717163, "learning_rate": 0.0002415948907548124, "loss": 5.0444, "step": 339850 }, { "epoch": 2.0495411295088095, "grad_norm": 1.5230673551559448, "learning_rate": 0.0002415787752809757, "loss": 5.1757, "step": 339900 }, { "epoch": 2.0498426212901437, "grad_norm": 0.8878608345985413, "learning_rate": 0.00024156265812176226, "loss": 5.0658, "step": 339950 }, { "epoch": 2.0501441130714775, "grad_norm": 1.5977190732955933, "learning_rate": 0.0002415465392774688, "loss": 4.9675, "step": 340000 }, { "epoch": 2.0504456048528117, "grad_norm": 1.7713067531585693, "learning_rate": 0.00024153041874839192, "loss": 4.9662, "step": 340050 }, { "epoch": 2.050747096634146, "grad_norm": 2.390695810317993, "learning_rate": 0.00024151429653482831, "loss": 5.2744, "step": 340100 }, { "epoch": 2.0510485884154797, "grad_norm": 2.480381965637207, "learning_rate": 0.00024149817263707465, "loss": 5.1347, "step": 340150 }, { "epoch": 2.051350080196814, "grad_norm": 1.6490719318389893, "learning_rate": 0.00024148204705542775, "loss": 4.3729, "step": 340200 }, { "epoch": 2.051651571978148, "grad_norm": 2.1725752353668213, "learning_rate": 0.0002414659197901843, "loss": 4.9374, "step": 340250 }, { "epoch": 2.051953063759482, "grad_norm": 1.3430888652801514, "learning_rate": 0.00024144979084164114, "loss": 5.2875, "step": 340300 }, { "epoch": 2.052254555540816, "grad_norm": 1.3985099792480469, "learning_rate": 0.00024143366021009505, "loss": 4.982, "step": 340350 }, { "epoch": 2.05255604732215, "grad_norm": 1.5002256631851196, "learning_rate": 0.00024141752789584293, "loss": 4.4601, "step": 340400 }, { "epoch": 2.052857539103484, "grad_norm": 1.5235862731933594, "learning_rate": 0.0002414013938991817, "loss": 4.9099, "step": 340450 }, { "epoch": 2.053159030884818, "grad_norm": 1.1212533712387085, "learning_rate": 0.00024138525822040824, "loss": 5.3142, "step": 340500 }, { "epoch": 2.053460522666152, "grad_norm": 1.5030529499053955, "learning_rate": 0.00024136912085981948, "loss": 4.9829, "step": 340550 }, { "epoch": 2.053762014447486, "grad_norm": 1.3150782585144043, "learning_rate": 0.00024135298181771246, "loss": 4.8484, "step": 340600 }, { "epoch": 2.0540635062288204, "grad_norm": 1.1893599033355713, "learning_rate": 0.00024133684109438423, "loss": 5.2673, "step": 340650 }, { "epoch": 2.054364998010154, "grad_norm": 1.6792101860046387, "learning_rate": 0.00024132069869013174, "loss": 5.0735, "step": 340700 }, { "epoch": 2.0546664897914884, "grad_norm": 1.3411933183670044, "learning_rate": 0.00024130455460525207, "loss": 4.8554, "step": 340750 }, { "epoch": 2.054967981572822, "grad_norm": 1.3783215284347534, "learning_rate": 0.00024128840884004245, "loss": 5.1907, "step": 340800 }, { "epoch": 2.0552694733541563, "grad_norm": 1.3241782188415527, "learning_rate": 0.00024127226139479988, "loss": 4.998, "step": 340850 }, { "epoch": 2.0555709651354905, "grad_norm": 1.3825637102127075, "learning_rate": 0.00024125611226982164, "loss": 5.0518, "step": 340900 }, { "epoch": 2.0558724569168243, "grad_norm": 1.5708775520324707, "learning_rate": 0.00024123996146540488, "loss": 4.9289, "step": 340950 }, { "epoch": 2.0561739486981585, "grad_norm": 1.9204723834991455, "learning_rate": 0.00024122380898184678, "loss": 4.9868, "step": 341000 }, { "epoch": 2.0564754404794927, "grad_norm": 1.1937377452850342, "learning_rate": 0.00024120765481944472, "loss": 4.3812, "step": 341050 }, { "epoch": 2.0567769322608265, "grad_norm": 1.3214423656463623, "learning_rate": 0.00024119149897849592, "loss": 4.6198, "step": 341100 }, { "epoch": 2.0570784240421607, "grad_norm": 0.9333351254463196, "learning_rate": 0.00024117534145929765, "loss": 5.4553, "step": 341150 }, { "epoch": 2.0573799158234944, "grad_norm": 0.9740355610847473, "learning_rate": 0.00024115918226214744, "loss": 5.2365, "step": 341200 }, { "epoch": 2.0576814076048286, "grad_norm": 2.0740432739257812, "learning_rate": 0.0002411430213873425, "loss": 4.8991, "step": 341250 }, { "epoch": 2.057982899386163, "grad_norm": 1.3224209547042847, "learning_rate": 0.00024112685883518034, "loss": 4.9463, "step": 341300 }, { "epoch": 2.0582843911674966, "grad_norm": 1.9334617853164673, "learning_rate": 0.00024111069460595842, "loss": 4.938, "step": 341350 }, { "epoch": 2.058585882948831, "grad_norm": 1.360449194908142, "learning_rate": 0.00024109452869997417, "loss": 5.2817, "step": 341400 }, { "epoch": 2.058887374730165, "grad_norm": 1.3831433057785034, "learning_rate": 0.00024107836111752508, "loss": 5.0119, "step": 341450 }, { "epoch": 2.059188866511499, "grad_norm": 2.1516189575195312, "learning_rate": 0.00024106219185890878, "loss": 5.2094, "step": 341500 }, { "epoch": 2.059490358292833, "grad_norm": 1.5526535511016846, "learning_rate": 0.00024104602092442275, "loss": 4.9318, "step": 341550 }, { "epoch": 2.0597918500741668, "grad_norm": 2.1309316158294678, "learning_rate": 0.00024102984831436464, "loss": 5.4242, "step": 341600 }, { "epoch": 2.060093341855501, "grad_norm": 1.368143916130066, "learning_rate": 0.0002410136740290321, "loss": 4.7402, "step": 341650 }, { "epoch": 2.060394833636835, "grad_norm": 0.7229447364807129, "learning_rate": 0.00024099749806872275, "loss": 4.8952, "step": 341700 }, { "epoch": 2.060696325418169, "grad_norm": 1.4331094026565552, "learning_rate": 0.00024098132043373434, "loss": 5.0083, "step": 341750 }, { "epoch": 2.060997817199503, "grad_norm": 1.390188217163086, "learning_rate": 0.00024096514112436455, "loss": 4.8072, "step": 341800 }, { "epoch": 2.0612993089808374, "grad_norm": 1.2908754348754883, "learning_rate": 0.0002409489601409111, "loss": 5.3588, "step": 341850 }, { "epoch": 2.061600800762171, "grad_norm": 2.1596834659576416, "learning_rate": 0.00024093277748367194, "loss": 4.9919, "step": 341900 }, { "epoch": 2.0619022925435053, "grad_norm": 1.2273002862930298, "learning_rate": 0.00024091659315294468, "loss": 5.1341, "step": 341950 }, { "epoch": 2.062203784324839, "grad_norm": 1.2018749713897705, "learning_rate": 0.00024090040714902733, "loss": 4.809, "step": 342000 }, { "epoch": 2.062203784324839, "eval_loss": 5.490420818328857, "eval_runtime": 39.1439, "eval_samples_per_second": 13.08, "eval_steps_per_second": 6.54, "eval_tts_loss": 7.384163642487763, "step": 342000 }, { "epoch": 2.0625052761061733, "grad_norm": 0.5476087331771851, "learning_rate": 0.00024088421947221765, "loss": 5.1874, "step": 342050 }, { "epoch": 2.0628067678875075, "grad_norm": 1.647473692893982, "learning_rate": 0.00024086803012281365, "loss": 5.0303, "step": 342100 }, { "epoch": 2.0631082596688413, "grad_norm": 1.918879747390747, "learning_rate": 0.00024085183910111323, "loss": 5.0773, "step": 342150 }, { "epoch": 2.0634097514501755, "grad_norm": 1.3601131439208984, "learning_rate": 0.0002408356464074144, "loss": 5.2645, "step": 342200 }, { "epoch": 2.0637112432315097, "grad_norm": 1.2774375677108765, "learning_rate": 0.00024081945204201504, "loss": 5.0443, "step": 342250 }, { "epoch": 2.0640127350128434, "grad_norm": 1.6412625312805176, "learning_rate": 0.00024080325600521336, "loss": 4.9642, "step": 342300 }, { "epoch": 2.0643142267941776, "grad_norm": 1.8976428508758545, "learning_rate": 0.00024078705829730726, "loss": 4.8916, "step": 342350 }, { "epoch": 2.064615718575512, "grad_norm": 1.4015052318572998, "learning_rate": 0.00024077085891859495, "loss": 4.9011, "step": 342400 }, { "epoch": 2.0649172103568456, "grad_norm": 1.7930493354797363, "learning_rate": 0.00024075465786937452, "loss": 5.0502, "step": 342450 }, { "epoch": 2.06521870213818, "grad_norm": 0.3295939266681671, "learning_rate": 0.00024073845514994413, "loss": 4.5561, "step": 342500 }, { "epoch": 2.0655201939195136, "grad_norm": 2.515552043914795, "learning_rate": 0.00024072225076060194, "loss": 5.0769, "step": 342550 }, { "epoch": 2.065821685700848, "grad_norm": 1.602497935295105, "learning_rate": 0.0002407060447016462, "loss": 5.0472, "step": 342600 }, { "epoch": 2.066123177482182, "grad_norm": 2.0400655269622803, "learning_rate": 0.0002406898369733752, "loss": 5.1176, "step": 342650 }, { "epoch": 2.0664246692635158, "grad_norm": 1.69877028465271, "learning_rate": 0.0002406736275760871, "loss": 5.1361, "step": 342700 }, { "epoch": 2.06672616104485, "grad_norm": 1.7480158805847168, "learning_rate": 0.00024065741651008033, "loss": 5.0132, "step": 342750 }, { "epoch": 2.067027652826184, "grad_norm": 1.4088311195373535, "learning_rate": 0.00024064120377565316, "loss": 5.1766, "step": 342800 }, { "epoch": 2.067329144607518, "grad_norm": 2.145326614379883, "learning_rate": 0.000240624989373104, "loss": 4.8538, "step": 342850 }, { "epoch": 2.067630636388852, "grad_norm": 2.1512022018432617, "learning_rate": 0.00024060877330273124, "loss": 4.6887, "step": 342900 }, { "epoch": 2.067932128170186, "grad_norm": 1.5396537780761719, "learning_rate": 0.0002405925555648333, "loss": 4.9771, "step": 342950 }, { "epoch": 2.06823361995152, "grad_norm": 1.4203848838806152, "learning_rate": 0.00024057633615970863, "loss": 4.8301, "step": 343000 }, { "epoch": 2.0685351117328543, "grad_norm": 1.7109780311584473, "learning_rate": 0.0002405601150876558, "loss": 5.0837, "step": 343050 }, { "epoch": 2.068836603514188, "grad_norm": 0.9553400874137878, "learning_rate": 0.00024054389234897327, "loss": 4.7556, "step": 343100 }, { "epoch": 2.0691380952955223, "grad_norm": 1.5644668340682983, "learning_rate": 0.00024052766794395962, "loss": 4.8856, "step": 343150 }, { "epoch": 2.0694395870768565, "grad_norm": 1.3635233640670776, "learning_rate": 0.0002405114418729134, "loss": 4.5554, "step": 343200 }, { "epoch": 2.0697410788581903, "grad_norm": 1.6574430465698242, "learning_rate": 0.00024049521413613333, "loss": 5.0559, "step": 343250 }, { "epoch": 2.0700425706395245, "grad_norm": 1.9256527423858643, "learning_rate": 0.00024047898473391794, "loss": 4.8576, "step": 343300 }, { "epoch": 2.0703440624208582, "grad_norm": 1.5824919939041138, "learning_rate": 0.00024046275366656597, "loss": 5.3556, "step": 343350 }, { "epoch": 2.0706455542021924, "grad_norm": 1.568924069404602, "learning_rate": 0.00024044652093437607, "loss": 5.1256, "step": 343400 }, { "epoch": 2.0709470459835266, "grad_norm": 1.7376257181167603, "learning_rate": 0.0002404302865376471, "loss": 5.0877, "step": 343450 }, { "epoch": 2.0712485377648604, "grad_norm": 1.1198816299438477, "learning_rate": 0.0002404140504766777, "loss": 4.8661, "step": 343500 }, { "epoch": 2.0715500295461946, "grad_norm": 1.5994782447814941, "learning_rate": 0.00024039781275176676, "loss": 5.0964, "step": 343550 }, { "epoch": 2.071851521327529, "grad_norm": 0.4293056130409241, "learning_rate": 0.00024038157336321307, "loss": 4.8013, "step": 343600 }, { "epoch": 2.0721530131088626, "grad_norm": 1.848304271697998, "learning_rate": 0.00024036533231131552, "loss": 4.8486, "step": 343650 }, { "epoch": 2.072454504890197, "grad_norm": 1.684349536895752, "learning_rate": 0.00024034908959637296, "loss": 4.6982, "step": 343700 }, { "epoch": 2.0727559966715305, "grad_norm": 1.2535368204116821, "learning_rate": 0.00024033284521868435, "loss": 5.0315, "step": 343750 }, { "epoch": 2.0730574884528647, "grad_norm": 1.4179004430770874, "learning_rate": 0.00024031659917854865, "loss": 4.5111, "step": 343800 }, { "epoch": 2.073358980234199, "grad_norm": 1.271213412284851, "learning_rate": 0.00024030035147626477, "loss": 4.9321, "step": 343850 }, { "epoch": 2.0736604720155327, "grad_norm": 0.2782977819442749, "learning_rate": 0.00024028410211213188, "loss": 4.5395, "step": 343900 }, { "epoch": 2.073961963796867, "grad_norm": 1.4401723146438599, "learning_rate": 0.00024026785108644886, "loss": 5.0821, "step": 343950 }, { "epoch": 2.074263455578201, "grad_norm": 0.7402086853981018, "learning_rate": 0.00024025159839951492, "loss": 4.5577, "step": 344000 }, { "epoch": 2.074564947359535, "grad_norm": 1.8472695350646973, "learning_rate": 0.00024023534405162906, "loss": 4.8335, "step": 344050 }, { "epoch": 2.074866439140869, "grad_norm": 1.3529765605926514, "learning_rate": 0.00024021908804309052, "loss": 5.1021, "step": 344100 }, { "epoch": 2.075167930922203, "grad_norm": 1.7606970071792603, "learning_rate": 0.00024020283037419835, "loss": 5.2102, "step": 344150 }, { "epoch": 2.075469422703537, "grad_norm": 1.6473772525787354, "learning_rate": 0.00024018657104525184, "loss": 5.293, "step": 344200 }, { "epoch": 2.0757709144848713, "grad_norm": 1.3622710704803467, "learning_rate": 0.00024017031005655018, "loss": 5.3109, "step": 344250 }, { "epoch": 2.076072406266205, "grad_norm": 1.3241182565689087, "learning_rate": 0.0002401540474083927, "loss": 5.1531, "step": 344300 }, { "epoch": 2.0763738980475392, "grad_norm": 2.463862180709839, "learning_rate": 0.00024013778310107855, "loss": 5.3223, "step": 344350 }, { "epoch": 2.0766753898288735, "grad_norm": 1.9017609357833862, "learning_rate": 0.00024012151713490718, "loss": 4.9975, "step": 344400 }, { "epoch": 2.076976881610207, "grad_norm": 1.3623417615890503, "learning_rate": 0.0002401052495101779, "loss": 4.5688, "step": 344450 }, { "epoch": 2.0772783733915414, "grad_norm": 1.9483040571212769, "learning_rate": 0.00024008898022719006, "loss": 5.1694, "step": 344500 }, { "epoch": 2.077579865172875, "grad_norm": 1.649497628211975, "learning_rate": 0.00024007270928624313, "loss": 5.3815, "step": 344550 }, { "epoch": 2.0778813569542094, "grad_norm": 1.547579050064087, "learning_rate": 0.00024005643668763654, "loss": 4.7346, "step": 344600 }, { "epoch": 2.0781828487355436, "grad_norm": 1.3204736709594727, "learning_rate": 0.0002400401624316697, "loss": 5.1803, "step": 344650 }, { "epoch": 2.0784843405168774, "grad_norm": 2.2097132205963135, "learning_rate": 0.00024002388651864223, "loss": 4.963, "step": 344700 }, { "epoch": 2.0787858322982116, "grad_norm": 2.3185572624206543, "learning_rate": 0.0002400076089488536, "loss": 5.3052, "step": 344750 }, { "epoch": 2.0790873240795458, "grad_norm": 1.3222354650497437, "learning_rate": 0.00023999132972260328, "loss": 5.1446, "step": 344800 }, { "epoch": 2.0793888158608795, "grad_norm": 1.2985196113586426, "learning_rate": 0.00023997504884019105, "loss": 5.3846, "step": 344850 }, { "epoch": 2.0796903076422137, "grad_norm": 1.4251117706298828, "learning_rate": 0.00023995876630191645, "loss": 4.695, "step": 344900 }, { "epoch": 2.0799917994235475, "grad_norm": 3.3376340866088867, "learning_rate": 0.00023994248210807906, "loss": 5.0555, "step": 344950 }, { "epoch": 2.0802932912048817, "grad_norm": 1.2514233589172363, "learning_rate": 0.00023992619625897872, "loss": 4.994, "step": 345000 }, { "epoch": 2.0802932912048817, "eval_loss": 5.500593662261963, "eval_runtime": 38.9751, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.568, "eval_tts_loss": 7.344838511139265, "step": 345000 }, { "epoch": 2.080594782986216, "grad_norm": 1.4748079776763916, "learning_rate": 0.00023990990875491497, "loss": 4.8627, "step": 345050 }, { "epoch": 2.0808962747675497, "grad_norm": 0.8230248689651489, "learning_rate": 0.00023989361959618775, "loss": 4.8884, "step": 345100 }, { "epoch": 2.081197766548884, "grad_norm": 1.8125331401824951, "learning_rate": 0.00023987732878309674, "loss": 4.7876, "step": 345150 }, { "epoch": 2.081499258330218, "grad_norm": 2.9090824127197266, "learning_rate": 0.0002398610363159417, "loss": 5.2263, "step": 345200 }, { "epoch": 2.081800750111552, "grad_norm": 1.4903500080108643, "learning_rate": 0.00023984474219502258, "loss": 5.1738, "step": 345250 }, { "epoch": 2.082102241892886, "grad_norm": 1.338962197303772, "learning_rate": 0.00023982844642063918, "loss": 4.8262, "step": 345300 }, { "epoch": 2.0824037336742203, "grad_norm": 2.7821695804595947, "learning_rate": 0.00023981214899309138, "loss": 5.2457, "step": 345350 }, { "epoch": 2.082705225455554, "grad_norm": 0.6844075918197632, "learning_rate": 0.00023979584991267922, "loss": 5.4522, "step": 345400 }, { "epoch": 2.0830067172368882, "grad_norm": 1.5712294578552246, "learning_rate": 0.00023977954917970252, "loss": 4.8078, "step": 345450 }, { "epoch": 2.083308209018222, "grad_norm": 1.6887949705123901, "learning_rate": 0.00023976324679446137, "loss": 5.0345, "step": 345500 }, { "epoch": 2.083609700799556, "grad_norm": 1.4800479412078857, "learning_rate": 0.00023974694275725574, "loss": 5.0987, "step": 345550 }, { "epoch": 2.0839111925808904, "grad_norm": 1.2945135831832886, "learning_rate": 0.00023973063706838574, "loss": 4.9297, "step": 345600 }, { "epoch": 2.084212684362224, "grad_norm": 1.375853180885315, "learning_rate": 0.00023971432972815136, "loss": 5.0427, "step": 345650 }, { "epoch": 2.0845141761435584, "grad_norm": 1.2068860530853271, "learning_rate": 0.00023969802073685285, "loss": 5.3496, "step": 345700 }, { "epoch": 2.084815667924892, "grad_norm": 2.1997549533843994, "learning_rate": 0.0002396817100947902, "loss": 5.2489, "step": 345750 }, { "epoch": 2.0851171597062264, "grad_norm": 1.6094400882720947, "learning_rate": 0.00023966539780226372, "loss": 5.0407, "step": 345800 }, { "epoch": 2.0854186514875606, "grad_norm": 2.1292786598205566, "learning_rate": 0.0002396490838595735, "loss": 5.402, "step": 345850 }, { "epoch": 2.0857201432688943, "grad_norm": 2.13678240776062, "learning_rate": 0.0002396327682670199, "loss": 5.0278, "step": 345900 }, { "epoch": 2.0860216350502285, "grad_norm": 1.306435465812683, "learning_rate": 0.00023961645102490303, "loss": 5.1732, "step": 345950 }, { "epoch": 2.0863231268315627, "grad_norm": 1.3249770402908325, "learning_rate": 0.0002396001321335233, "loss": 5.3586, "step": 346000 }, { "epoch": 2.0866246186128965, "grad_norm": 2.136225461959839, "learning_rate": 0.000239583811593181, "loss": 4.9233, "step": 346050 }, { "epoch": 2.0869261103942307, "grad_norm": 0.9340272545814514, "learning_rate": 0.0002395674894041765, "loss": 4.8839, "step": 346100 }, { "epoch": 2.087227602175565, "grad_norm": 1.350788950920105, "learning_rate": 0.00023955116556681016, "loss": 4.9866, "step": 346150 }, { "epoch": 2.0875290939568987, "grad_norm": 1.5455620288848877, "learning_rate": 0.00023953484008138246, "loss": 4.9419, "step": 346200 }, { "epoch": 2.087830585738233, "grad_norm": 1.138250708580017, "learning_rate": 0.00023951851294819375, "loss": 5.1577, "step": 346250 }, { "epoch": 2.0881320775195666, "grad_norm": 0.9186525344848633, "learning_rate": 0.0002395021841675446, "loss": 4.9278, "step": 346300 }, { "epoch": 2.088433569300901, "grad_norm": 2.3575713634490967, "learning_rate": 0.00023948585373973546, "loss": 5.1246, "step": 346350 }, { "epoch": 2.088735061082235, "grad_norm": 1.5348230600357056, "learning_rate": 0.00023946952166506685, "loss": 5.0617, "step": 346400 }, { "epoch": 2.089036552863569, "grad_norm": 1.3049554824829102, "learning_rate": 0.00023945318794383943, "loss": 4.7986, "step": 346450 }, { "epoch": 2.089338044644903, "grad_norm": 1.0119627714157104, "learning_rate": 0.00023943685257635374, "loss": 4.6549, "step": 346500 }, { "epoch": 2.0896395364262372, "grad_norm": 0.8980438113212585, "learning_rate": 0.00023942051556291035, "loss": 5.0145, "step": 346550 }, { "epoch": 2.089941028207571, "grad_norm": 1.8420298099517822, "learning_rate": 0.00023940417690381004, "loss": 4.8097, "step": 346600 }, { "epoch": 2.090242519988905, "grad_norm": 1.3740957975387573, "learning_rate": 0.00023938783659935344, "loss": 4.9952, "step": 346650 }, { "epoch": 2.090544011770239, "grad_norm": 2.3211519718170166, "learning_rate": 0.0002393714946498412, "loss": 5.2982, "step": 346700 }, { "epoch": 2.090845503551573, "grad_norm": 2.0093905925750732, "learning_rate": 0.0002393551510555742, "loss": 4.898, "step": 346750 }, { "epoch": 2.0911469953329074, "grad_norm": 1.1295137405395508, "learning_rate": 0.00023933880581685315, "loss": 5.0901, "step": 346800 }, { "epoch": 2.091448487114241, "grad_norm": 1.4973881244659424, "learning_rate": 0.0002393224589339789, "loss": 5.2532, "step": 346850 }, { "epoch": 2.0917499788955753, "grad_norm": 1.4657634496688843, "learning_rate": 0.00023930611040725224, "loss": 4.7759, "step": 346900 }, { "epoch": 2.0920514706769096, "grad_norm": 1.969988465309143, "learning_rate": 0.00023928976023697407, "loss": 4.9088, "step": 346950 }, { "epoch": 2.0923529624582433, "grad_norm": 1.469736099243164, "learning_rate": 0.0002392734084234453, "loss": 5.6082, "step": 347000 }, { "epoch": 2.0926544542395775, "grad_norm": 1.2503300905227661, "learning_rate": 0.00023925705496696684, "loss": 4.7415, "step": 347050 }, { "epoch": 2.0929559460209113, "grad_norm": 2.4167582988739014, "learning_rate": 0.00023924069986783962, "loss": 4.9953, "step": 347100 }, { "epoch": 2.0932574378022455, "grad_norm": 1.5189305543899536, "learning_rate": 0.00023922434312636468, "loss": 5.1649, "step": 347150 }, { "epoch": 2.0935589295835797, "grad_norm": 1.3662625551223755, "learning_rate": 0.00023920798474284306, "loss": 4.6278, "step": 347200 }, { "epoch": 2.0938604213649135, "grad_norm": 1.5683282613754272, "learning_rate": 0.0002391916247175758, "loss": 4.9901, "step": 347250 }, { "epoch": 2.0941619131462477, "grad_norm": 1.8266394138336182, "learning_rate": 0.00023917526305086394, "loss": 5.0523, "step": 347300 }, { "epoch": 2.094463404927582, "grad_norm": 1.3766546249389648, "learning_rate": 0.00023915889974300864, "loss": 4.9425, "step": 347350 }, { "epoch": 2.0947648967089156, "grad_norm": 1.7081294059753418, "learning_rate": 0.000239142534794311, "loss": 4.976, "step": 347400 }, { "epoch": 2.09506638849025, "grad_norm": 1.217564344406128, "learning_rate": 0.00023912616820507225, "loss": 4.624, "step": 347450 }, { "epoch": 2.0953678802715836, "grad_norm": 2.052583932876587, "learning_rate": 0.0002391097999755936, "loss": 5.2118, "step": 347500 }, { "epoch": 2.095669372052918, "grad_norm": 1.9127397537231445, "learning_rate": 0.0002390934301061762, "loss": 4.8645, "step": 347550 }, { "epoch": 2.095970863834252, "grad_norm": 1.9122239351272583, "learning_rate": 0.00023907705859712135, "loss": 5.4309, "step": 347600 }, { "epoch": 2.096272355615586, "grad_norm": 1.3229680061340332, "learning_rate": 0.00023906068544873041, "loss": 4.6683, "step": 347650 }, { "epoch": 2.09657384739692, "grad_norm": 1.5956919193267822, "learning_rate": 0.00023904431066130457, "loss": 4.9555, "step": 347700 }, { "epoch": 2.096875339178254, "grad_norm": 0.749758780002594, "learning_rate": 0.00023902793423514533, "loss": 5.2788, "step": 347750 }, { "epoch": 2.097176830959588, "grad_norm": 0.831181526184082, "learning_rate": 0.000239011556170554, "loss": 5.2357, "step": 347800 }, { "epoch": 2.097478322740922, "grad_norm": 2.0669045448303223, "learning_rate": 0.00023899517646783198, "loss": 4.9876, "step": 347850 }, { "epoch": 2.097779814522256, "grad_norm": 1.4380364418029785, "learning_rate": 0.00023897879512728076, "loss": 5.1945, "step": 347900 }, { "epoch": 2.09808130630359, "grad_norm": 1.4203132390975952, "learning_rate": 0.0002389624121492018, "loss": 4.686, "step": 347950 }, { "epoch": 2.0983827980849243, "grad_norm": 1.9679478406906128, "learning_rate": 0.00023894602753389658, "loss": 4.7299, "step": 348000 }, { "epoch": 2.0983827980849243, "eval_loss": 5.49969482421875, "eval_runtime": 39.0368, "eval_samples_per_second": 13.116, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.349054386911075, "step": 348000 }, { "epoch": 2.098684289866258, "grad_norm": 1.5645172595977783, "learning_rate": 0.00023892964128166666, "loss": 4.7389, "step": 348050 }, { "epoch": 2.0989857816475923, "grad_norm": 1.6531450748443604, "learning_rate": 0.00023891325339281364, "loss": 4.8922, "step": 348100 }, { "epoch": 2.0992872734289265, "grad_norm": 1.4367849826812744, "learning_rate": 0.00023889686386763904, "loss": 5.549, "step": 348150 }, { "epoch": 2.0995887652102603, "grad_norm": 3.4692776203155518, "learning_rate": 0.0002388804727064445, "loss": 4.9983, "step": 348200 }, { "epoch": 2.0998902569915945, "grad_norm": 1.1634835004806519, "learning_rate": 0.00023886407990953175, "loss": 4.6841, "step": 348250 }, { "epoch": 2.1001917487729282, "grad_norm": 1.565075397491455, "learning_rate": 0.00023884768547720236, "loss": 5.0795, "step": 348300 }, { "epoch": 2.1004932405542625, "grad_norm": 1.3609249591827393, "learning_rate": 0.0002388312894097582, "loss": 4.7547, "step": 348350 }, { "epoch": 2.1007947323355967, "grad_norm": 3.2397403717041016, "learning_rate": 0.00023881489170750084, "loss": 5.0368, "step": 348400 }, { "epoch": 2.1010962241169304, "grad_norm": 2.038282871246338, "learning_rate": 0.00023879849237073224, "loss": 4.8749, "step": 348450 }, { "epoch": 2.1013977158982646, "grad_norm": 1.9932349920272827, "learning_rate": 0.00023878209139975405, "loss": 5.2779, "step": 348500 }, { "epoch": 2.101699207679599, "grad_norm": 1.8379701375961304, "learning_rate": 0.00023876568879486817, "loss": 5.1968, "step": 348550 }, { "epoch": 2.1020006994609326, "grad_norm": 1.0640755891799927, "learning_rate": 0.00023874928455637644, "loss": 4.8365, "step": 348600 }, { "epoch": 2.102302191242267, "grad_norm": 0.43969234824180603, "learning_rate": 0.00023873287868458084, "loss": 5.054, "step": 348650 }, { "epoch": 2.1026036830236006, "grad_norm": 2.324681520462036, "learning_rate": 0.00023871647117978316, "loss": 5.1814, "step": 348700 }, { "epoch": 2.1029051748049348, "grad_norm": 1.7813951969146729, "learning_rate": 0.0002387000620422855, "loss": 4.6744, "step": 348750 }, { "epoch": 2.103206666586269, "grad_norm": 2.1613945960998535, "learning_rate": 0.00023868365127238978, "loss": 4.9484, "step": 348800 }, { "epoch": 2.1035081583676027, "grad_norm": 1.6057883501052856, "learning_rate": 0.00023866723887039796, "loss": 4.8223, "step": 348850 }, { "epoch": 2.103809650148937, "grad_norm": 0.8109497427940369, "learning_rate": 0.00023865082483661215, "loss": 4.9032, "step": 348900 }, { "epoch": 2.104111141930271, "grad_norm": 1.4879207611083984, "learning_rate": 0.00023863440917133445, "loss": 5.1413, "step": 348950 }, { "epoch": 2.104412633711605, "grad_norm": 1.3100950717926025, "learning_rate": 0.00023861799187486696, "loss": 5.3908, "step": 349000 }, { "epoch": 2.104714125492939, "grad_norm": 1.253089189529419, "learning_rate": 0.0002386015729475118, "loss": 4.9514, "step": 349050 }, { "epoch": 2.1050156172742733, "grad_norm": 0.9886460304260254, "learning_rate": 0.00023858515238957108, "loss": 4.9537, "step": 349100 }, { "epoch": 2.105317109055607, "grad_norm": 0.8448439836502075, "learning_rate": 0.00023856873020134707, "loss": 5.2702, "step": 349150 }, { "epoch": 2.1056186008369413, "grad_norm": 1.1570404767990112, "learning_rate": 0.00023855230638314195, "loss": 4.9811, "step": 349200 }, { "epoch": 2.105920092618275, "grad_norm": 2.069633960723877, "learning_rate": 0.00023853588093525801, "loss": 4.8313, "step": 349250 }, { "epoch": 2.1062215843996093, "grad_norm": 1.1824392080307007, "learning_rate": 0.00023851945385799754, "loss": 4.8785, "step": 349300 }, { "epoch": 2.1065230761809435, "grad_norm": 1.909995675086975, "learning_rate": 0.00023850302515166286, "loss": 4.9411, "step": 349350 }, { "epoch": 2.1068245679622772, "grad_norm": 1.8553781509399414, "learning_rate": 0.00023848659481655627, "loss": 5.3602, "step": 349400 }, { "epoch": 2.1071260597436114, "grad_norm": 2.181842088699341, "learning_rate": 0.0002384701628529802, "loss": 5.04, "step": 349450 }, { "epoch": 2.107427551524945, "grad_norm": 1.4854520559310913, "learning_rate": 0.00023845372926123705, "loss": 5.0606, "step": 349500 }, { "epoch": 2.1077290433062794, "grad_norm": 2.361543655395508, "learning_rate": 0.00023843729404162927, "loss": 5.0394, "step": 349550 }, { "epoch": 2.1080305350876136, "grad_norm": 2.1072001457214355, "learning_rate": 0.00023842085719445922, "loss": 4.7875, "step": 349600 }, { "epoch": 2.1083320268689474, "grad_norm": 1.7886433601379395, "learning_rate": 0.00023840441872002952, "loss": 5.2456, "step": 349650 }, { "epoch": 2.1086335186502816, "grad_norm": 1.6521990299224854, "learning_rate": 0.00023838797861864267, "loss": 5.2545, "step": 349700 }, { "epoch": 2.108935010431616, "grad_norm": 1.9960274696350098, "learning_rate": 0.00023837153689060115, "loss": 5.0116, "step": 349750 }, { "epoch": 2.1092365022129496, "grad_norm": 2.6015801429748535, "learning_rate": 0.00023835509353620766, "loss": 5.0483, "step": 349800 }, { "epoch": 2.1095379939942838, "grad_norm": 2.2318978309631348, "learning_rate": 0.00023833864855576478, "loss": 4.9783, "step": 349850 }, { "epoch": 2.109839485775618, "grad_norm": 1.39707350730896, "learning_rate": 0.00023832220194957508, "loss": 5.4306, "step": 349900 }, { "epoch": 2.1101409775569517, "grad_norm": 3.1446921825408936, "learning_rate": 0.00023830575371794135, "loss": 4.9588, "step": 349950 }, { "epoch": 2.110442469338286, "grad_norm": 1.9532839059829712, "learning_rate": 0.0002382893038611662, "loss": 5.4814, "step": 350000 }, { "epoch": 2.1107439611196197, "grad_norm": 2.1122148036956787, "learning_rate": 0.0002382728523795524, "loss": 4.8616, "step": 350050 }, { "epoch": 2.111045452900954, "grad_norm": 1.310932993888855, "learning_rate": 0.00023825639927340276, "loss": 5.0877, "step": 350100 }, { "epoch": 2.111346944682288, "grad_norm": 1.6922802925109863, "learning_rate": 0.00023823994454302003, "loss": 5.0493, "step": 350150 }, { "epoch": 2.111648436463622, "grad_norm": 1.8220001459121704, "learning_rate": 0.00023822348818870707, "loss": 4.9877, "step": 350200 }, { "epoch": 2.111949928244956, "grad_norm": 0.8062180876731873, "learning_rate": 0.00023820703021076666, "loss": 5.0253, "step": 350250 }, { "epoch": 2.1122514200262903, "grad_norm": 1.685318112373352, "learning_rate": 0.00023819057060950178, "loss": 5.1265, "step": 350300 }, { "epoch": 2.112552911807624, "grad_norm": 1.3541555404663086, "learning_rate": 0.0002381741093852153, "loss": 4.8567, "step": 350350 }, { "epoch": 2.1128544035889583, "grad_norm": 0.8520934581756592, "learning_rate": 0.00023815764653821014, "loss": 5.2523, "step": 350400 }, { "epoch": 2.113155895370292, "grad_norm": 1.301801085472107, "learning_rate": 0.0002381411820687893, "loss": 5.1649, "step": 350450 }, { "epoch": 2.1134573871516262, "grad_norm": 1.4423127174377441, "learning_rate": 0.00023812471597725582, "loss": 5.0298, "step": 350500 }, { "epoch": 2.1137588789329604, "grad_norm": 0.8082759380340576, "learning_rate": 0.0002381082482639127, "loss": 4.6495, "step": 350550 }, { "epoch": 2.114060370714294, "grad_norm": 1.7657803297042847, "learning_rate": 0.00023809177892906297, "loss": 4.9892, "step": 350600 }, { "epoch": 2.1143618624956284, "grad_norm": 1.5334287881851196, "learning_rate": 0.0002380753079730098, "loss": 5.3536, "step": 350650 }, { "epoch": 2.1146633542769626, "grad_norm": 1.3644369840621948, "learning_rate": 0.00023805883539605626, "loss": 4.9501, "step": 350700 }, { "epoch": 2.1149648460582964, "grad_norm": 1.1498746871948242, "learning_rate": 0.00023804236119850554, "loss": 5.0052, "step": 350750 }, { "epoch": 2.1152663378396306, "grad_norm": 2.390435218811035, "learning_rate": 0.00023802588538066077, "loss": 4.8441, "step": 350800 }, { "epoch": 2.1155678296209643, "grad_norm": 1.4155938625335693, "learning_rate": 0.00023800940794282526, "loss": 5.0774, "step": 350850 }, { "epoch": 2.1158693214022986, "grad_norm": 1.6656830310821533, "learning_rate": 0.0002379929288853021, "loss": 5.3995, "step": 350900 }, { "epoch": 2.1161708131836328, "grad_norm": 1.6433736085891724, "learning_rate": 0.00023797644820839473, "loss": 4.9137, "step": 350950 }, { "epoch": 2.1164723049649665, "grad_norm": 1.2974942922592163, "learning_rate": 0.00023795996591240636, "loss": 4.9762, "step": 351000 }, { "epoch": 2.1164723049649665, "eval_loss": 5.493133068084717, "eval_runtime": 39.1486, "eval_samples_per_second": 13.078, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.367378085194515, "step": 351000 }, { "epoch": 2.1167737967463007, "grad_norm": 1.812204122543335, "learning_rate": 0.00023794348199764038, "loss": 4.9741, "step": 351050 }, { "epoch": 2.117075288527635, "grad_norm": 1.2726013660430908, "learning_rate": 0.00023792699646440005, "loss": 5.2747, "step": 351100 }, { "epoch": 2.1173767803089687, "grad_norm": 2.7772185802459717, "learning_rate": 0.00023791050931298887, "loss": 5.0972, "step": 351150 }, { "epoch": 2.117678272090303, "grad_norm": 2.2069549560546875, "learning_rate": 0.00023789402054371018, "loss": 4.8875, "step": 351200 }, { "epoch": 2.1179797638716367, "grad_norm": 2.9188244342803955, "learning_rate": 0.00023787753015686754, "loss": 5.1453, "step": 351250 }, { "epoch": 2.118281255652971, "grad_norm": 1.7986136674880981, "learning_rate": 0.0002378610381527643, "loss": 5.1887, "step": 351300 }, { "epoch": 2.118582747434305, "grad_norm": 1.0054244995117188, "learning_rate": 0.00023784454453170404, "loss": 5.2317, "step": 351350 }, { "epoch": 2.118884239215639, "grad_norm": 1.835063099861145, "learning_rate": 0.00023782804929399037, "loss": 4.9729, "step": 351400 }, { "epoch": 2.119185730996973, "grad_norm": 1.9761985540390015, "learning_rate": 0.00023781155243992675, "loss": 5.272, "step": 351450 }, { "epoch": 2.1194872227783073, "grad_norm": 2.3305134773254395, "learning_rate": 0.0002377950539698168, "loss": 4.9604, "step": 351500 }, { "epoch": 2.119788714559641, "grad_norm": 1.2689961194992065, "learning_rate": 0.00023777855388396418, "loss": 5.2943, "step": 351550 }, { "epoch": 2.1200902063409752, "grad_norm": 2.1641929149627686, "learning_rate": 0.00023776205218267258, "loss": 5.1437, "step": 351600 }, { "epoch": 2.120391698122309, "grad_norm": 1.810424566268921, "learning_rate": 0.0002377455488662456, "loss": 4.7731, "step": 351650 }, { "epoch": 2.120693189903643, "grad_norm": 2.075652599334717, "learning_rate": 0.00023772904393498702, "loss": 4.8922, "step": 351700 }, { "epoch": 2.1209946816849774, "grad_norm": 2.1921372413635254, "learning_rate": 0.0002377125373892006, "loss": 5.2379, "step": 351750 }, { "epoch": 2.121296173466311, "grad_norm": 0.5176324844360352, "learning_rate": 0.00023769602922919008, "loss": 5.1807, "step": 351800 }, { "epoch": 2.1215976652476454, "grad_norm": 2.600656032562256, "learning_rate": 0.00023767951945525935, "loss": 4.9267, "step": 351850 }, { "epoch": 2.1218991570289796, "grad_norm": 1.4376792907714844, "learning_rate": 0.00023766300806771213, "loss": 5.0824, "step": 351900 }, { "epoch": 2.1222006488103133, "grad_norm": 1.3541091680526733, "learning_rate": 0.00023764649506685242, "loss": 4.7928, "step": 351950 }, { "epoch": 2.1225021405916475, "grad_norm": 1.8345011472702026, "learning_rate": 0.000237629980452984, "loss": 5.139, "step": 352000 }, { "epoch": 2.1228036323729813, "grad_norm": 1.871681809425354, "learning_rate": 0.0002376134642264108, "loss": 5.4439, "step": 352050 }, { "epoch": 2.1231051241543155, "grad_norm": 1.9089404344558716, "learning_rate": 0.00023759694638743684, "loss": 5.2388, "step": 352100 }, { "epoch": 2.1234066159356497, "grad_norm": 1.9456546306610107, "learning_rate": 0.00023758042693636613, "loss": 5.029, "step": 352150 }, { "epoch": 2.1237081077169835, "grad_norm": 1.7004691362380981, "learning_rate": 0.00023756390587350264, "loss": 5.3298, "step": 352200 }, { "epoch": 2.1240095994983177, "grad_norm": 1.673795461654663, "learning_rate": 0.0002375473831991504, "loss": 5.0069, "step": 352250 }, { "epoch": 2.124311091279652, "grad_norm": 0.8306607604026794, "learning_rate": 0.00023753085891361353, "loss": 4.9615, "step": 352300 }, { "epoch": 2.1246125830609857, "grad_norm": 1.383116364479065, "learning_rate": 0.00023751433301719607, "loss": 4.9221, "step": 352350 }, { "epoch": 2.12491407484232, "grad_norm": 2.6424381732940674, "learning_rate": 0.00023749780551020225, "loss": 5.4801, "step": 352400 }, { "epoch": 2.1252155666236536, "grad_norm": 1.272430419921875, "learning_rate": 0.00023748127639293618, "loss": 5.2488, "step": 352450 }, { "epoch": 2.125517058404988, "grad_norm": 1.7676327228546143, "learning_rate": 0.000237464745665702, "loss": 5.0683, "step": 352500 }, { "epoch": 2.125818550186322, "grad_norm": 1.0760722160339355, "learning_rate": 0.00023744821332880408, "loss": 4.9284, "step": 352550 }, { "epoch": 2.126120041967656, "grad_norm": 1.852011799812317, "learning_rate": 0.00023743167938254658, "loss": 5.038, "step": 352600 }, { "epoch": 2.12642153374899, "grad_norm": 1.5201940536499023, "learning_rate": 0.00023741514382723378, "loss": 4.6965, "step": 352650 }, { "epoch": 2.126723025530324, "grad_norm": 1.7058260440826416, "learning_rate": 0.00023739860666316997, "loss": 4.9859, "step": 352700 }, { "epoch": 2.127024517311658, "grad_norm": 1.2517311573028564, "learning_rate": 0.00023738206789065958, "loss": 5.3606, "step": 352750 }, { "epoch": 2.127326009092992, "grad_norm": 2.2170374393463135, "learning_rate": 0.00023736552751000686, "loss": 4.7808, "step": 352800 }, { "epoch": 2.1276275008743264, "grad_norm": 1.5809245109558105, "learning_rate": 0.00023734898552151633, "loss": 5.1346, "step": 352850 }, { "epoch": 2.12792899265566, "grad_norm": 1.8945618867874146, "learning_rate": 0.0002373324419254924, "loss": 5.2462, "step": 352900 }, { "epoch": 2.1282304844369944, "grad_norm": 0.8785726428031921, "learning_rate": 0.00023731589672223945, "loss": 5.1962, "step": 352950 }, { "epoch": 2.128531976218328, "grad_norm": 1.3456499576568604, "learning_rate": 0.00023729934991206209, "loss": 5.0768, "step": 353000 }, { "epoch": 2.1288334679996623, "grad_norm": 2.828864812850952, "learning_rate": 0.00023728280149526473, "loss": 4.7764, "step": 353050 }, { "epoch": 2.1291349597809965, "grad_norm": 1.5524208545684814, "learning_rate": 0.00023726625147215197, "loss": 5.0045, "step": 353100 }, { "epoch": 2.1294364515623303, "grad_norm": 1.033977746963501, "learning_rate": 0.00023724969984302842, "loss": 4.8717, "step": 353150 }, { "epoch": 2.1297379433436645, "grad_norm": 2.5432817935943604, "learning_rate": 0.00023723314660819864, "loss": 4.781, "step": 353200 }, { "epoch": 2.1300394351249983, "grad_norm": 1.5055949687957764, "learning_rate": 0.00023721659176796723, "loss": 5.1774, "step": 353250 }, { "epoch": 2.1303409269063325, "grad_norm": 1.689841628074646, "learning_rate": 0.00023720003532263898, "loss": 5.169, "step": 353300 }, { "epoch": 2.1306424186876667, "grad_norm": 1.5695592164993286, "learning_rate": 0.00023718347727251849, "loss": 4.9155, "step": 353350 }, { "epoch": 2.1309439104690004, "grad_norm": 1.3340730667114258, "learning_rate": 0.00023716691761791053, "loss": 4.8896, "step": 353400 }, { "epoch": 2.1312454022503347, "grad_norm": 2.0187153816223145, "learning_rate": 0.00023715035635911988, "loss": 4.8515, "step": 353450 }, { "epoch": 2.131546894031669, "grad_norm": 1.4540696144104004, "learning_rate": 0.00023713379349645124, "loss": 5.3993, "step": 353500 }, { "epoch": 2.1318483858130026, "grad_norm": 1.4089250564575195, "learning_rate": 0.0002371172290302095, "loss": 5.2667, "step": 353550 }, { "epoch": 2.132149877594337, "grad_norm": 2.7140212059020996, "learning_rate": 0.0002371006629606995, "loss": 5.1814, "step": 353600 }, { "epoch": 2.132451369375671, "grad_norm": 1.8618381023406982, "learning_rate": 0.00023708409528822605, "loss": 4.5207, "step": 353650 }, { "epoch": 2.132752861157005, "grad_norm": 1.5345760583877563, "learning_rate": 0.00023706752601309415, "loss": 4.7049, "step": 353700 }, { "epoch": 2.133054352938339, "grad_norm": 1.364035964012146, "learning_rate": 0.00023705095513560867, "loss": 4.9357, "step": 353750 }, { "epoch": 2.1333558447196728, "grad_norm": 2.0954713821411133, "learning_rate": 0.00023703438265607459, "loss": 5.208, "step": 353800 }, { "epoch": 2.133657336501007, "grad_norm": 1.5852835178375244, "learning_rate": 0.0002370178085747969, "loss": 5.2049, "step": 353850 }, { "epoch": 2.133958828282341, "grad_norm": 0.505648136138916, "learning_rate": 0.00023700123289208062, "loss": 4.8323, "step": 353900 }, { "epoch": 2.134260320063675, "grad_norm": 1.5238564014434814, "learning_rate": 0.0002369846556082308, "loss": 5.0588, "step": 353950 }, { "epoch": 2.134561811845009, "grad_norm": 1.3487062454223633, "learning_rate": 0.0002369680767235525, "loss": 4.5468, "step": 354000 }, { "epoch": 2.134561811845009, "eval_loss": 5.485201835632324, "eval_runtime": 39.055, "eval_samples_per_second": 13.11, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.274131117994054, "step": 354000 }, { "epoch": 2.134863303626343, "grad_norm": 1.5973035097122192, "learning_rate": 0.00023695149623835095, "loss": 4.9282, "step": 354050 }, { "epoch": 2.135164795407677, "grad_norm": 2.08272123336792, "learning_rate": 0.00023693491415293112, "loss": 5.0983, "step": 354100 }, { "epoch": 2.1354662871890113, "grad_norm": 0.7807849049568176, "learning_rate": 0.00023691833046759827, "loss": 4.6917, "step": 354150 }, { "epoch": 2.135767778970345, "grad_norm": 1.965896725654602, "learning_rate": 0.00023690174518265758, "loss": 4.6504, "step": 354200 }, { "epoch": 2.1360692707516793, "grad_norm": 2.185537815093994, "learning_rate": 0.00023688515829841427, "loss": 5.4058, "step": 354250 }, { "epoch": 2.1363707625330135, "grad_norm": 0.4006425142288208, "learning_rate": 0.00023686856981517367, "loss": 4.7155, "step": 354300 }, { "epoch": 2.1366722543143473, "grad_norm": 1.7833222150802612, "learning_rate": 0.00023685197973324101, "loss": 5.4418, "step": 354350 }, { "epoch": 2.1369737460956815, "grad_norm": 1.7473355531692505, "learning_rate": 0.00023683538805292155, "loss": 5.0125, "step": 354400 }, { "epoch": 2.1372752378770157, "grad_norm": 0.9667778015136719, "learning_rate": 0.0002368187947745207, "loss": 5.0064, "step": 354450 }, { "epoch": 2.1375767296583494, "grad_norm": 1.4989691972732544, "learning_rate": 0.0002368021998983439, "loss": 5.0991, "step": 354500 }, { "epoch": 2.1378782214396836, "grad_norm": 1.5562145709991455, "learning_rate": 0.00023678560342469643, "loss": 4.9606, "step": 354550 }, { "epoch": 2.1381797132210174, "grad_norm": 1.433308720588684, "learning_rate": 0.00023676900535388377, "loss": 4.9482, "step": 354600 }, { "epoch": 2.1384812050023516, "grad_norm": 2.259385347366333, "learning_rate": 0.00023675240568621144, "loss": 4.5384, "step": 354650 }, { "epoch": 2.138782696783686, "grad_norm": 1.4463998079299927, "learning_rate": 0.0002367358044219848, "loss": 5.0201, "step": 354700 }, { "epoch": 2.1390841885650196, "grad_norm": 2.13822340965271, "learning_rate": 0.00023671920156150957, "loss": 4.886, "step": 354750 }, { "epoch": 2.139385680346354, "grad_norm": 1.8611863851547241, "learning_rate": 0.00023670259710509113, "loss": 5.056, "step": 354800 }, { "epoch": 2.139687172127688, "grad_norm": 1.3324700593948364, "learning_rate": 0.0002366859910530351, "loss": 5.1017, "step": 354850 }, { "epoch": 2.1399886639090218, "grad_norm": 1.714543104171753, "learning_rate": 0.0002366693834056471, "loss": 4.726, "step": 354900 }, { "epoch": 2.140290155690356, "grad_norm": 1.635402798652649, "learning_rate": 0.0002366527741632329, "loss": 5.1386, "step": 354950 }, { "epoch": 2.1405916474716897, "grad_norm": 1.9927849769592285, "learning_rate": 0.00023663616332609793, "loss": 5.0864, "step": 355000 }, { "epoch": 2.140893139253024, "grad_norm": 1.4678620100021362, "learning_rate": 0.00023661955089454808, "loss": 5.0783, "step": 355050 }, { "epoch": 2.141194631034358, "grad_norm": 2.687244176864624, "learning_rate": 0.00023660293686888895, "loss": 4.895, "step": 355100 }, { "epoch": 2.141496122815692, "grad_norm": 1.4325799942016602, "learning_rate": 0.0002365863212494264, "loss": 4.8695, "step": 355150 }, { "epoch": 2.141797614597026, "grad_norm": 0.6198286414146423, "learning_rate": 0.00023656970403646615, "loss": 4.5351, "step": 355200 }, { "epoch": 2.1420991063783603, "grad_norm": 1.0027360916137695, "learning_rate": 0.00023655308523031404, "loss": 4.8399, "step": 355250 }, { "epoch": 2.142400598159694, "grad_norm": 1.5033818483352661, "learning_rate": 0.00023653646483127593, "loss": 5.0969, "step": 355300 }, { "epoch": 2.1427020899410283, "grad_norm": 2.0325276851654053, "learning_rate": 0.00023651984283965766, "loss": 4.9638, "step": 355350 }, { "epoch": 2.143003581722362, "grad_norm": 1.915210485458374, "learning_rate": 0.00023650321925576515, "loss": 4.9398, "step": 355400 }, { "epoch": 2.1433050735036963, "grad_norm": 1.684273600578308, "learning_rate": 0.00023648659407990436, "loss": 4.5682, "step": 355450 }, { "epoch": 2.1436065652850305, "grad_norm": 1.5083168745040894, "learning_rate": 0.00023646996731238124, "loss": 5.0959, "step": 355500 }, { "epoch": 2.1439080570663642, "grad_norm": 2.2713377475738525, "learning_rate": 0.00023645333895350173, "loss": 5.1414, "step": 355550 }, { "epoch": 2.1442095488476984, "grad_norm": 1.8775745630264282, "learning_rate": 0.00023643670900357191, "loss": 5.0887, "step": 355600 }, { "epoch": 2.1445110406290326, "grad_norm": 1.3203953504562378, "learning_rate": 0.00023642007746289782, "loss": 5.0966, "step": 355650 }, { "epoch": 2.1448125324103664, "grad_norm": 1.0600841045379639, "learning_rate": 0.0002364034443317855, "loss": 5.3466, "step": 355700 }, { "epoch": 2.1451140241917006, "grad_norm": 2.2958123683929443, "learning_rate": 0.00023638680961054117, "loss": 4.7981, "step": 355750 }, { "epoch": 2.1454155159730344, "grad_norm": 1.3718390464782715, "learning_rate": 0.00023637017329947084, "loss": 4.9946, "step": 355800 }, { "epoch": 2.1457170077543686, "grad_norm": 0.9301065802574158, "learning_rate": 0.0002363535353988807, "loss": 5.1834, "step": 355850 }, { "epoch": 2.146018499535703, "grad_norm": 1.2523903846740723, "learning_rate": 0.00023633689590907698, "loss": 4.5747, "step": 355900 }, { "epoch": 2.1463199913170365, "grad_norm": 1.1202999353408813, "learning_rate": 0.00023632025483036594, "loss": 4.5031, "step": 355950 }, { "epoch": 2.1466214830983708, "grad_norm": 1.18617844581604, "learning_rate": 0.00023630361216305377, "loss": 5.224, "step": 356000 }, { "epoch": 2.146922974879705, "grad_norm": 1.3758790493011475, "learning_rate": 0.00023628696790744676, "loss": 4.9073, "step": 356050 }, { "epoch": 2.1472244666610387, "grad_norm": 1.5088459253311157, "learning_rate": 0.00023627032206385132, "loss": 4.77, "step": 356100 }, { "epoch": 2.147525958442373, "grad_norm": 1.602639079093933, "learning_rate": 0.00023625367463257364, "loss": 4.635, "step": 356150 }, { "epoch": 2.1478274502237067, "grad_norm": 2.772603750228882, "learning_rate": 0.00023623702561392017, "loss": 4.9303, "step": 356200 }, { "epoch": 2.148128942005041, "grad_norm": 1.7904086112976074, "learning_rate": 0.00023622037500819736, "loss": 5.1463, "step": 356250 }, { "epoch": 2.148430433786375, "grad_norm": 1.5882681608200073, "learning_rate": 0.00023620372281571154, "loss": 4.9525, "step": 356300 }, { "epoch": 2.148731925567709, "grad_norm": 2.0975327491760254, "learning_rate": 0.00023618706903676928, "loss": 5.185, "step": 356350 }, { "epoch": 2.149033417349043, "grad_norm": 1.6759189367294312, "learning_rate": 0.000236170413671677, "loss": 5.0033, "step": 356400 }, { "epoch": 2.1493349091303773, "grad_norm": 1.8101494312286377, "learning_rate": 0.00023615375672074118, "loss": 4.6609, "step": 356450 }, { "epoch": 2.149636400911711, "grad_norm": 1.461408257484436, "learning_rate": 0.00023613709818426845, "loss": 4.8412, "step": 356500 }, { "epoch": 2.1499378926930452, "grad_norm": 1.4895219802856445, "learning_rate": 0.00023612043806256534, "loss": 4.9717, "step": 356550 }, { "epoch": 2.1502393844743795, "grad_norm": 1.190940022468567, "learning_rate": 0.00023610377635593845, "loss": 5.0915, "step": 356600 }, { "epoch": 2.150540876255713, "grad_norm": 1.2588834762573242, "learning_rate": 0.00023608711306469447, "loss": 5.0321, "step": 356650 }, { "epoch": 2.1508423680370474, "grad_norm": 1.225446343421936, "learning_rate": 0.00023607044818914002, "loss": 5.2411, "step": 356700 }, { "epoch": 2.151143859818381, "grad_norm": 2.560309886932373, "learning_rate": 0.0002360537817295818, "loss": 5.1849, "step": 356750 }, { "epoch": 2.1514453515997154, "grad_norm": 0.6776605248451233, "learning_rate": 0.0002360371136863265, "loss": 5.0297, "step": 356800 }, { "epoch": 2.1517468433810496, "grad_norm": 2.0663187503814697, "learning_rate": 0.00023602044405968093, "loss": 4.6815, "step": 356850 }, { "epoch": 2.1520483351623834, "grad_norm": 1.4217742681503296, "learning_rate": 0.00023600377284995182, "loss": 5.3164, "step": 356900 }, { "epoch": 2.1523498269437176, "grad_norm": 2.181976318359375, "learning_rate": 0.00023598710005744605, "loss": 4.9731, "step": 356950 }, { "epoch": 2.1526513187250513, "grad_norm": 1.5047893524169922, "learning_rate": 0.00023597042568247037, "loss": 4.9615, "step": 357000 }, { "epoch": 2.1526513187250513, "eval_loss": 5.490405559539795, "eval_runtime": 39.1814, "eval_samples_per_second": 13.067, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.334113418755581, "step": 357000 }, { "epoch": 2.1529528105063855, "grad_norm": 1.4509316682815552, "learning_rate": 0.00023595374972533174, "loss": 5.2938, "step": 357050 }, { "epoch": 2.1532543022877197, "grad_norm": 1.6795964241027832, "learning_rate": 0.00023593707218633696, "loss": 4.9207, "step": 357100 }, { "epoch": 2.1535557940690535, "grad_norm": 1.1888511180877686, "learning_rate": 0.00023592039306579303, "loss": 4.9423, "step": 357150 }, { "epoch": 2.1538572858503877, "grad_norm": 1.2763694524765015, "learning_rate": 0.00023590371236400685, "loss": 5.0824, "step": 357200 }, { "epoch": 2.154158777631722, "grad_norm": 2.1168322563171387, "learning_rate": 0.0002358870300812855, "loss": 5.2844, "step": 357250 }, { "epoch": 2.1544602694130557, "grad_norm": 2.4129717350006104, "learning_rate": 0.0002358703462179359, "loss": 4.8406, "step": 357300 }, { "epoch": 2.15476176119439, "grad_norm": 1.3531564474105835, "learning_rate": 0.00023585366077426512, "loss": 4.9719, "step": 357350 }, { "epoch": 2.155063252975724, "grad_norm": 0.9028947949409485, "learning_rate": 0.0002358369737505802, "loss": 4.8739, "step": 357400 }, { "epoch": 2.155364744757058, "grad_norm": 2.111502170562744, "learning_rate": 0.00023582028514718828, "loss": 4.6747, "step": 357450 }, { "epoch": 2.155666236538392, "grad_norm": 1.9210641384124756, "learning_rate": 0.0002358035949643965, "loss": 4.7136, "step": 357500 }, { "epoch": 2.155967728319726, "grad_norm": 1.3103984594345093, "learning_rate": 0.000235786903202512, "loss": 4.9363, "step": 357550 }, { "epoch": 2.15626922010106, "grad_norm": 1.9691712856292725, "learning_rate": 0.000235770209861842, "loss": 4.785, "step": 357600 }, { "epoch": 2.1565707118823942, "grad_norm": 1.82803475856781, "learning_rate": 0.00023575351494269366, "loss": 4.7558, "step": 357650 }, { "epoch": 2.156872203663728, "grad_norm": 2.435480833053589, "learning_rate": 0.00023573681844537428, "loss": 4.7805, "step": 357700 }, { "epoch": 2.157173695445062, "grad_norm": 1.621532678604126, "learning_rate": 0.00023572012037019108, "loss": 5.1736, "step": 357750 }, { "epoch": 2.157475187226396, "grad_norm": 2.40805721282959, "learning_rate": 0.00023570342071745139, "loss": 5.2077, "step": 357800 }, { "epoch": 2.15777667900773, "grad_norm": 1.3671730756759644, "learning_rate": 0.0002356867194874626, "loss": 4.6138, "step": 357850 }, { "epoch": 2.1580781707890644, "grad_norm": 1.4636471271514893, "learning_rate": 0.00023567001668053195, "loss": 4.9982, "step": 357900 }, { "epoch": 2.158379662570398, "grad_norm": 0.5186167359352112, "learning_rate": 0.0002356533122969669, "loss": 5.0147, "step": 357950 }, { "epoch": 2.1586811543517324, "grad_norm": 1.495434284210205, "learning_rate": 0.00023563660633707493, "loss": 4.6884, "step": 358000 }, { "epoch": 2.1589826461330666, "grad_norm": 1.993231177330017, "learning_rate": 0.0002356198988011634, "loss": 5.1789, "step": 358050 }, { "epoch": 2.1592841379144003, "grad_norm": 1.2438993453979492, "learning_rate": 0.00023560318968953985, "loss": 4.5768, "step": 358100 }, { "epoch": 2.1595856296957345, "grad_norm": 1.5999810695648193, "learning_rate": 0.0002355864790025117, "loss": 5.2756, "step": 358150 }, { "epoch": 2.1598871214770687, "grad_norm": 2.157334089279175, "learning_rate": 0.00023556976674038655, "loss": 5.1911, "step": 358200 }, { "epoch": 2.1601886132584025, "grad_norm": 1.7346665859222412, "learning_rate": 0.00023555305290347197, "loss": 5.0972, "step": 358250 }, { "epoch": 2.1604901050397367, "grad_norm": 1.1412951946258545, "learning_rate": 0.00023553633749207558, "loss": 5.1593, "step": 358300 }, { "epoch": 2.1607915968210705, "grad_norm": 1.874810814857483, "learning_rate": 0.00023551962050650494, "loss": 5.4169, "step": 358350 }, { "epoch": 2.1610930886024047, "grad_norm": 1.848753571510315, "learning_rate": 0.0002355029019470677, "loss": 5.0486, "step": 358400 }, { "epoch": 2.161394580383739, "grad_norm": 1.5450457334518433, "learning_rate": 0.00023548618181407159, "loss": 4.6283, "step": 358450 }, { "epoch": 2.1616960721650726, "grad_norm": 0.6100249886512756, "learning_rate": 0.0002354694601078243, "loss": 4.4666, "step": 358500 }, { "epoch": 2.161997563946407, "grad_norm": 1.707689881324768, "learning_rate": 0.00023545273682863357, "loss": 5.2132, "step": 358550 }, { "epoch": 2.162299055727741, "grad_norm": 1.7177993059158325, "learning_rate": 0.00023543601197680716, "loss": 4.9002, "step": 358600 }, { "epoch": 2.162600547509075, "grad_norm": 1.488269567489624, "learning_rate": 0.00023541928555265287, "loss": 5.2995, "step": 358650 }, { "epoch": 2.162902039290409, "grad_norm": 1.3556183576583862, "learning_rate": 0.00023540255755647856, "loss": 4.9068, "step": 358700 }, { "epoch": 2.163203531071743, "grad_norm": 1.0610865354537964, "learning_rate": 0.00023538582798859203, "loss": 4.8353, "step": 358750 }, { "epoch": 2.163505022853077, "grad_norm": 1.3060022592544556, "learning_rate": 0.00023536909684930118, "loss": 5.06, "step": 358800 }, { "epoch": 2.163806514634411, "grad_norm": 1.4806352853775024, "learning_rate": 0.00023535236413891393, "loss": 4.9044, "step": 358850 }, { "epoch": 2.164108006415745, "grad_norm": 1.2930933237075806, "learning_rate": 0.00023533562985773824, "loss": 5.0204, "step": 358900 }, { "epoch": 2.164409498197079, "grad_norm": 2.0487353801727295, "learning_rate": 0.00023531889400608207, "loss": 4.8256, "step": 358950 }, { "epoch": 2.1647109899784134, "grad_norm": 1.923093557357788, "learning_rate": 0.00023530215658425338, "loss": 5.2505, "step": 359000 }, { "epoch": 2.165012481759747, "grad_norm": 1.9461649656295776, "learning_rate": 0.00023528541759256026, "loss": 5.1887, "step": 359050 }, { "epoch": 2.1653139735410813, "grad_norm": 1.9651479721069336, "learning_rate": 0.00023526867703131072, "loss": 4.7689, "step": 359100 }, { "epoch": 2.165615465322415, "grad_norm": 0.967905580997467, "learning_rate": 0.0002352519349008129, "loss": 4.9082, "step": 359150 }, { "epoch": 2.1659169571037493, "grad_norm": 1.838297963142395, "learning_rate": 0.0002352351912013748, "loss": 4.6222, "step": 359200 }, { "epoch": 2.1662184488850835, "grad_norm": 1.7240220308303833, "learning_rate": 0.00023521844593330472, "loss": 5.2099, "step": 359250 }, { "epoch": 2.1665199406664173, "grad_norm": 1.953192114830017, "learning_rate": 0.00023520169909691072, "loss": 4.9738, "step": 359300 }, { "epoch": 2.1668214324477515, "grad_norm": 1.4154109954833984, "learning_rate": 0.00023518495069250103, "loss": 5.0617, "step": 359350 }, { "epoch": 2.1671229242290857, "grad_norm": 2.1366896629333496, "learning_rate": 0.0002351682007203839, "loss": 5.0865, "step": 359400 }, { "epoch": 2.1674244160104195, "grad_norm": 2.8829915523529053, "learning_rate": 0.0002351514491808676, "loss": 5.4244, "step": 359450 }, { "epoch": 2.1677259077917537, "grad_norm": 2.027991771697998, "learning_rate": 0.0002351346960742604, "loss": 5.1141, "step": 359500 }, { "epoch": 2.1680273995730874, "grad_norm": 0.5093584656715393, "learning_rate": 0.00023511794140087058, "loss": 4.7093, "step": 359550 }, { "epoch": 2.1683288913544216, "grad_norm": 1.6163122653961182, "learning_rate": 0.00023510118516100652, "loss": 5.0113, "step": 359600 }, { "epoch": 2.168630383135756, "grad_norm": 2.32906436920166, "learning_rate": 0.00023508442735497656, "loss": 5.4597, "step": 359650 }, { "epoch": 2.1689318749170896, "grad_norm": 1.1142467260360718, "learning_rate": 0.00023506766798308922, "loss": 5.0785, "step": 359700 }, { "epoch": 2.169233366698424, "grad_norm": 2.2976205348968506, "learning_rate": 0.0002350509070456528, "loss": 4.889, "step": 359750 }, { "epoch": 2.169534858479758, "grad_norm": 2.503711700439453, "learning_rate": 0.00023503414454297583, "loss": 5.0642, "step": 359800 }, { "epoch": 2.169836350261092, "grad_norm": 2.027477741241455, "learning_rate": 0.00023501738047536672, "loss": 5.0903, "step": 359850 }, { "epoch": 2.170137842042426, "grad_norm": 2.424328327178955, "learning_rate": 0.0002350006148431341, "loss": 5.2787, "step": 359900 }, { "epoch": 2.1704393338237598, "grad_norm": 2.397448778152466, "learning_rate": 0.0002349838476465865, "loss": 4.8979, "step": 359950 }, { "epoch": 2.170740825605094, "grad_norm": 1.4855562448501587, "learning_rate": 0.0002349670788860324, "loss": 5.2215, "step": 360000 }, { "epoch": 2.170740825605094, "eval_loss": 5.48117208480835, "eval_runtime": 39.2429, "eval_samples_per_second": 13.047, "eval_steps_per_second": 6.523, "eval_tts_loss": 7.368436406147708, "step": 360000 }, { "epoch": 2.171042317386428, "grad_norm": 2.146392583847046, "learning_rate": 0.0002349503085617805, "loss": 5.4122, "step": 360050 }, { "epoch": 2.171343809167762, "grad_norm": 2.02713942527771, "learning_rate": 0.00023493353667413937, "loss": 5.2972, "step": 360100 }, { "epoch": 2.171645300949096, "grad_norm": 1.8242506980895996, "learning_rate": 0.00023491676322341773, "loss": 5.2287, "step": 360150 }, { "epoch": 2.1719467927304303, "grad_norm": 1.4977481365203857, "learning_rate": 0.00023489998820992424, "loss": 4.6435, "step": 360200 }, { "epoch": 2.172248284511764, "grad_norm": 1.851745843887329, "learning_rate": 0.00023488321163396761, "loss": 4.8269, "step": 360250 }, { "epoch": 2.1725497762930983, "grad_norm": 1.351804256439209, "learning_rate": 0.00023486643349585664, "loss": 4.8474, "step": 360300 }, { "epoch": 2.1728512680744325, "grad_norm": 2.9883127212524414, "learning_rate": 0.00023484965379590004, "loss": 5.4393, "step": 360350 }, { "epoch": 2.1731527598557663, "grad_norm": 2.618375778198242, "learning_rate": 0.0002348328725344067, "loss": 4.8514, "step": 360400 }, { "epoch": 2.1734542516371005, "grad_norm": 1.4671138525009155, "learning_rate": 0.00023481608971168534, "loss": 5.3028, "step": 360450 }, { "epoch": 2.1737557434184342, "grad_norm": 1.8103410005569458, "learning_rate": 0.00023479930532804494, "loss": 4.9505, "step": 360500 }, { "epoch": 2.1740572351997685, "grad_norm": 0.576184093952179, "learning_rate": 0.00023478251938379428, "loss": 5.176, "step": 360550 }, { "epoch": 2.1743587269811027, "grad_norm": 1.3194146156311035, "learning_rate": 0.0002347657318792424, "loss": 5.4645, "step": 360600 }, { "epoch": 2.1746602187624364, "grad_norm": 2.897505521774292, "learning_rate": 0.00023474894281469815, "loss": 4.6483, "step": 360650 }, { "epoch": 2.1749617105437706, "grad_norm": 3.5889124870300293, "learning_rate": 0.00023473215219047057, "loss": 4.8095, "step": 360700 }, { "epoch": 2.1752632023251044, "grad_norm": 1.4788411855697632, "learning_rate": 0.00023471536000686867, "loss": 4.886, "step": 360750 }, { "epoch": 2.1755646941064386, "grad_norm": 0.6294626593589783, "learning_rate": 0.00023469856626420142, "loss": 4.5679, "step": 360800 }, { "epoch": 2.175866185887773, "grad_norm": 1.3382189273834229, "learning_rate": 0.00023468177096277798, "loss": 4.7411, "step": 360850 }, { "epoch": 2.1761676776691066, "grad_norm": 1.2409656047821045, "learning_rate": 0.00023466497410290735, "loss": 5.0928, "step": 360900 }, { "epoch": 2.1764691694504408, "grad_norm": 1.5604532957077026, "learning_rate": 0.00023464817568489874, "loss": 5.5552, "step": 360950 }, { "epoch": 2.176770661231775, "grad_norm": 1.8239063024520874, "learning_rate": 0.00023463137570906118, "loss": 4.4343, "step": 361000 }, { "epoch": 2.1770721530131087, "grad_norm": 1.7936159372329712, "learning_rate": 0.00023461457417570399, "loss": 5.0954, "step": 361050 }, { "epoch": 2.177373644794443, "grad_norm": 1.9462615251541138, "learning_rate": 0.0002345977710851363, "loss": 5.0344, "step": 361100 }, { "epoch": 2.177675136575777, "grad_norm": 1.689653754234314, "learning_rate": 0.00023458096643766735, "loss": 5.0518, "step": 361150 }, { "epoch": 2.177976628357111, "grad_norm": 0.7657845616340637, "learning_rate": 0.00023456416023360643, "loss": 4.8677, "step": 361200 }, { "epoch": 2.178278120138445, "grad_norm": 2.1701881885528564, "learning_rate": 0.0002345473524732628, "loss": 5.4284, "step": 361250 }, { "epoch": 2.178579611919779, "grad_norm": 2.306117057800293, "learning_rate": 0.00023453054315694582, "loss": 5.2478, "step": 361300 }, { "epoch": 2.178881103701113, "grad_norm": 1.3138651847839355, "learning_rate": 0.00023451373228496484, "loss": 4.912, "step": 361350 }, { "epoch": 2.1791825954824473, "grad_norm": 1.8378722667694092, "learning_rate": 0.0002344969198576292, "loss": 4.7199, "step": 361400 }, { "epoch": 2.179484087263781, "grad_norm": 1.6884523630142212, "learning_rate": 0.00023448010587524834, "loss": 5.0102, "step": 361450 }, { "epoch": 2.1797855790451153, "grad_norm": 1.5216184854507446, "learning_rate": 0.0002344632903381317, "loss": 4.9372, "step": 361500 }, { "epoch": 2.180087070826449, "grad_norm": 1.706298589706421, "learning_rate": 0.0002344464732465887, "loss": 4.988, "step": 361550 }, { "epoch": 2.1803885626077832, "grad_norm": 1.4513505697250366, "learning_rate": 0.00023442965460092894, "loss": 5.2161, "step": 361600 }, { "epoch": 2.1806900543891174, "grad_norm": 1.310868740081787, "learning_rate": 0.0002344128344014618, "loss": 5.0212, "step": 361650 }, { "epoch": 2.180991546170451, "grad_norm": 1.659927487373352, "learning_rate": 0.00023439601264849697, "loss": 5.0132, "step": 361700 }, { "epoch": 2.1812930379517854, "grad_norm": 1.753757357597351, "learning_rate": 0.00023437918934234394, "loss": 4.6426, "step": 361750 }, { "epoch": 2.1815945297331196, "grad_norm": 0.7928467988967896, "learning_rate": 0.0002343623644833123, "loss": 4.8399, "step": 361800 }, { "epoch": 2.1818960215144534, "grad_norm": 1.9026477336883545, "learning_rate": 0.00023434553807171176, "loss": 4.2571, "step": 361850 }, { "epoch": 2.1821975132957876, "grad_norm": 1.424299955368042, "learning_rate": 0.00023432871010785198, "loss": 4.9019, "step": 361900 }, { "epoch": 2.182499005077122, "grad_norm": 1.1132073402404785, "learning_rate": 0.00023431188059204266, "loss": 4.791, "step": 361950 }, { "epoch": 2.1828004968584556, "grad_norm": 1.7412751913070679, "learning_rate": 0.00023429504952459347, "loss": 5.5677, "step": 362000 }, { "epoch": 2.1831019886397898, "grad_norm": 2.1314165592193604, "learning_rate": 0.00023427821690581417, "loss": 5.2724, "step": 362050 }, { "epoch": 2.1834034804211235, "grad_norm": 1.4082081317901611, "learning_rate": 0.00023426138273601458, "loss": 5.4154, "step": 362100 }, { "epoch": 2.1837049722024577, "grad_norm": 1.9299144744873047, "learning_rate": 0.00023424454701550447, "loss": 4.999, "step": 362150 }, { "epoch": 2.184006463983792, "grad_norm": 1.871943712234497, "learning_rate": 0.0002342277097445937, "loss": 5.1246, "step": 362200 }, { "epoch": 2.1843079557651257, "grad_norm": 2.149111747741699, "learning_rate": 0.00023421087092359214, "loss": 5.0161, "step": 362250 }, { "epoch": 2.18460944754646, "grad_norm": 1.2303000688552856, "learning_rate": 0.0002341940305528097, "loss": 4.9818, "step": 362300 }, { "epoch": 2.184910939327794, "grad_norm": 1.3766030073165894, "learning_rate": 0.00023417718863255626, "loss": 5.3801, "step": 362350 }, { "epoch": 2.185212431109128, "grad_norm": 1.2161273956298828, "learning_rate": 0.00023416034516314177, "loss": 5.0043, "step": 362400 }, { "epoch": 2.185513922890462, "grad_norm": 2.1573069095611572, "learning_rate": 0.00023414350014487625, "loss": 5.0267, "step": 362450 }, { "epoch": 2.185815414671796, "grad_norm": 1.197798490524292, "learning_rate": 0.00023412665357806968, "loss": 4.9911, "step": 362500 }, { "epoch": 2.18611690645313, "grad_norm": 1.960366129875183, "learning_rate": 0.0002341098054630321, "loss": 4.7723, "step": 362550 }, { "epoch": 2.1864183982344643, "grad_norm": 0.7724676728248596, "learning_rate": 0.00023409295580007362, "loss": 5.0257, "step": 362600 }, { "epoch": 2.186719890015798, "grad_norm": 1.3768411874771118, "learning_rate": 0.00023407610458950427, "loss": 5.4422, "step": 362650 }, { "epoch": 2.1870213817971322, "grad_norm": 1.6819757223129272, "learning_rate": 0.0002340592518316342, "loss": 4.7784, "step": 362700 }, { "epoch": 2.1873228735784664, "grad_norm": 1.0843089818954468, "learning_rate": 0.00023404239752677355, "loss": 5.1995, "step": 362750 }, { "epoch": 2.1876243653598, "grad_norm": 1.3959383964538574, "learning_rate": 0.00023402554167523248, "loss": 4.9761, "step": 362800 }, { "epoch": 2.1879258571411344, "grad_norm": 1.5226668119430542, "learning_rate": 0.0002340086842773213, "loss": 5.4732, "step": 362850 }, { "epoch": 2.188227348922468, "grad_norm": 1.815210223197937, "learning_rate": 0.00023399182533335013, "loss": 4.8825, "step": 362900 }, { "epoch": 2.1885288407038024, "grad_norm": 1.9594085216522217, "learning_rate": 0.00023397496484362927, "loss": 4.8683, "step": 362950 }, { "epoch": 2.1888303324851366, "grad_norm": 2.06872820854187, "learning_rate": 0.00023395810280846905, "loss": 5.2856, "step": 363000 }, { "epoch": 2.1888303324851366, "eval_loss": 5.500622749328613, "eval_runtime": 39.079, "eval_samples_per_second": 13.102, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.34398955882029, "step": 363000 }, { "epoch": 2.1891318242664703, "grad_norm": 1.6483278274536133, "learning_rate": 0.00023394123922817973, "loss": 5.0361, "step": 363050 }, { "epoch": 2.1894333160478046, "grad_norm": 1.4084163904190063, "learning_rate": 0.0002339243741030717, "loss": 4.7386, "step": 363100 }, { "epoch": 2.1897348078291388, "grad_norm": 1.3794647455215454, "learning_rate": 0.00023390750743345533, "loss": 4.8842, "step": 363150 }, { "epoch": 2.1900362996104725, "grad_norm": 1.6809171438217163, "learning_rate": 0.00023389063921964106, "loss": 5.0455, "step": 363200 }, { "epoch": 2.1903377913918067, "grad_norm": 2.4049386978149414, "learning_rate": 0.00023387376946193926, "loss": 4.661, "step": 363250 }, { "epoch": 2.1906392831731405, "grad_norm": 1.6223524808883667, "learning_rate": 0.00023385689816066043, "loss": 4.9459, "step": 363300 }, { "epoch": 2.1909407749544747, "grad_norm": 1.5105130672454834, "learning_rate": 0.0002338400253161151, "loss": 4.9127, "step": 363350 }, { "epoch": 2.191242266735809, "grad_norm": 1.7729846239089966, "learning_rate": 0.0002338231509286137, "loss": 4.9776, "step": 363400 }, { "epoch": 2.1915437585171427, "grad_norm": 1.617348074913025, "learning_rate": 0.00023380627499846682, "loss": 5.226, "step": 363450 }, { "epoch": 2.191845250298477, "grad_norm": 1.7842713594436646, "learning_rate": 0.00023378939752598506, "loss": 5.3775, "step": 363500 }, { "epoch": 2.192146742079811, "grad_norm": 1.8779044151306152, "learning_rate": 0.00023377251851147904, "loss": 5.019, "step": 363550 }, { "epoch": 2.192448233861145, "grad_norm": 2.4826550483703613, "learning_rate": 0.00023375563795525933, "loss": 5.2232, "step": 363600 }, { "epoch": 2.192749725642479, "grad_norm": 1.4880361557006836, "learning_rate": 0.00023373875585763667, "loss": 5.2013, "step": 363650 }, { "epoch": 2.193051217423813, "grad_norm": 1.2902781963348389, "learning_rate": 0.00023372187221892168, "loss": 5.0029, "step": 363700 }, { "epoch": 2.193352709205147, "grad_norm": 1.4348505735397339, "learning_rate": 0.0002337049870394251, "loss": 5.3348, "step": 363750 }, { "epoch": 2.1936542009864812, "grad_norm": 1.5667804479599, "learning_rate": 0.00023368810031945775, "loss": 4.9652, "step": 363800 }, { "epoch": 2.193955692767815, "grad_norm": 2.3806095123291016, "learning_rate": 0.00023367121205933027, "loss": 5.2914, "step": 363850 }, { "epoch": 2.194257184549149, "grad_norm": 2.1315646171569824, "learning_rate": 0.00023365432225935358, "loss": 5.0528, "step": 363900 }, { "epoch": 2.1945586763304834, "grad_norm": 1.370103359222412, "learning_rate": 0.0002336374309198385, "loss": 4.9012, "step": 363950 }, { "epoch": 2.194860168111817, "grad_norm": 1.527489423751831, "learning_rate": 0.00023362053804109578, "loss": 4.6421, "step": 364000 }, { "epoch": 2.1951616598931514, "grad_norm": 2.6855432987213135, "learning_rate": 0.0002336036436234364, "loss": 4.7715, "step": 364050 }, { "epoch": 2.1954631516744856, "grad_norm": 1.5261310338974, "learning_rate": 0.00023358674766717132, "loss": 5.2019, "step": 364100 }, { "epoch": 2.1957646434558193, "grad_norm": 1.5307552814483643, "learning_rate": 0.00023356985017261141, "loss": 4.9162, "step": 364150 }, { "epoch": 2.1960661352371535, "grad_norm": 1.3188307285308838, "learning_rate": 0.00023355295114006768, "loss": 5.1826, "step": 364200 }, { "epoch": 2.1963676270184873, "grad_norm": 1.2123832702636719, "learning_rate": 0.00023353605056985113, "loss": 5.085, "step": 364250 }, { "epoch": 2.1966691187998215, "grad_norm": 1.385267972946167, "learning_rate": 0.0002335191484622727, "loss": 5.1491, "step": 364300 }, { "epoch": 2.1969706105811557, "grad_norm": 1.0772637128829956, "learning_rate": 0.0002335022448176436, "loss": 4.2765, "step": 364350 }, { "epoch": 2.1972721023624895, "grad_norm": 2.056598663330078, "learning_rate": 0.00023348533963627485, "loss": 5.2553, "step": 364400 }, { "epoch": 2.1975735941438237, "grad_norm": 1.1572209596633911, "learning_rate": 0.00023346843291847756, "loss": 5.2303, "step": 364450 }, { "epoch": 2.1978750859251575, "grad_norm": 2.1923272609710693, "learning_rate": 0.00023345152466456288, "loss": 4.8233, "step": 364500 }, { "epoch": 2.1981765777064917, "grad_norm": 2.1367592811584473, "learning_rate": 0.000233434614874842, "loss": 5.0249, "step": 364550 }, { "epoch": 2.198478069487826, "grad_norm": 1.9015021324157715, "learning_rate": 0.00023341770354962606, "loss": 5.1381, "step": 364600 }, { "epoch": 2.1987795612691596, "grad_norm": 1.9928615093231201, "learning_rate": 0.00023340079068922631, "loss": 5.2861, "step": 364650 }, { "epoch": 2.199081053050494, "grad_norm": 1.8555134534835815, "learning_rate": 0.00023338387629395408, "loss": 5.3624, "step": 364700 }, { "epoch": 2.199382544831828, "grad_norm": 2.187966823577881, "learning_rate": 0.00023336696036412055, "loss": 5.2543, "step": 364750 }, { "epoch": 2.199684036613162, "grad_norm": 1.441055178642273, "learning_rate": 0.00023335004290003713, "loss": 4.9607, "step": 364800 }, { "epoch": 2.199985528394496, "grad_norm": 2.4181582927703857, "learning_rate": 0.00023333312390201512, "loss": 5.09, "step": 364850 }, { "epoch": 2.20028702017583, "grad_norm": 1.4884631633758545, "learning_rate": 0.00023331620337036586, "loss": 4.8366, "step": 364900 }, { "epoch": 2.200588511957164, "grad_norm": 3.1516547203063965, "learning_rate": 0.00023329928130540076, "loss": 4.9893, "step": 364950 }, { "epoch": 2.200890003738498, "grad_norm": 3.0672383308410645, "learning_rate": 0.0002332823577074313, "loss": 4.8811, "step": 365000 }, { "epoch": 2.201191495519832, "grad_norm": 1.1805092096328735, "learning_rate": 0.00023326543257676892, "loss": 4.6948, "step": 365050 }, { "epoch": 2.201492987301166, "grad_norm": 0.6130797863006592, "learning_rate": 0.00023324850591372503, "loss": 4.8743, "step": 365100 }, { "epoch": 2.2017944790825004, "grad_norm": 1.3080024719238281, "learning_rate": 0.00023323157771861123, "loss": 4.543, "step": 365150 }, { "epoch": 2.202095970863834, "grad_norm": 1.6837722063064575, "learning_rate": 0.00023321464799173897, "loss": 4.9474, "step": 365200 }, { "epoch": 2.2023974626451683, "grad_norm": 1.433898687362671, "learning_rate": 0.00023319771673341986, "loss": 5.0237, "step": 365250 }, { "epoch": 2.202698954426502, "grad_norm": 3.14622163772583, "learning_rate": 0.00023318078394396556, "loss": 5.229, "step": 365300 }, { "epoch": 2.2030004462078363, "grad_norm": 1.7449626922607422, "learning_rate": 0.0002331638496236876, "loss": 5.438, "step": 365350 }, { "epoch": 2.2033019379891705, "grad_norm": 2.1181554794311523, "learning_rate": 0.0002331469137728977, "loss": 4.8727, "step": 365400 }, { "epoch": 2.2036034297705043, "grad_norm": 2.451768159866333, "learning_rate": 0.00023312997639190748, "loss": 5.4359, "step": 365450 }, { "epoch": 2.2039049215518385, "grad_norm": 1.2707699537277222, "learning_rate": 0.00023311303748102869, "loss": 5.163, "step": 365500 }, { "epoch": 2.2042064133331727, "grad_norm": 1.8162384033203125, "learning_rate": 0.00023309609704057302, "loss": 5.0485, "step": 365550 }, { "epoch": 2.2045079051145064, "grad_norm": 2.244022846221924, "learning_rate": 0.0002330791550708523, "loss": 4.9878, "step": 365600 }, { "epoch": 2.2048093968958407, "grad_norm": 2.324296712875366, "learning_rate": 0.00023306221157217831, "loss": 5.0867, "step": 365650 }, { "epoch": 2.205110888677175, "grad_norm": 1.2874693870544434, "learning_rate": 0.00023304526654486284, "loss": 4.7085, "step": 365700 }, { "epoch": 2.2054123804585086, "grad_norm": 1.9728920459747314, "learning_rate": 0.00023302831998921775, "loss": 5.3121, "step": 365750 }, { "epoch": 2.205713872239843, "grad_norm": 2.4226644039154053, "learning_rate": 0.00023301137190555492, "loss": 5.0781, "step": 365800 }, { "epoch": 2.2060153640211766, "grad_norm": 1.9639838933944702, "learning_rate": 0.00023299442229418623, "loss": 4.8547, "step": 365850 }, { "epoch": 2.206316855802511, "grad_norm": 1.7315504550933838, "learning_rate": 0.00023297747115542367, "loss": 5.13, "step": 365900 }, { "epoch": 2.206618347583845, "grad_norm": 1.694995641708374, "learning_rate": 0.00023296051848957914, "loss": 5.134, "step": 365950 }, { "epoch": 2.2069198393651788, "grad_norm": 1.2929607629776, "learning_rate": 0.00023294356429696467, "loss": 5.2932, "step": 366000 }, { "epoch": 2.2069198393651788, "eval_loss": 5.467554569244385, "eval_runtime": 39.1649, "eval_samples_per_second": 13.073, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.320334097312016, "step": 366000 }, { "epoch": 2.207221331146513, "grad_norm": 1.4235475063323975, "learning_rate": 0.00023292660857789226, "loss": 5.1794, "step": 366050 }, { "epoch": 2.207522822927847, "grad_norm": 2.8292319774627686, "learning_rate": 0.00023290965133267397, "loss": 5.0081, "step": 366100 }, { "epoch": 2.207824314709181, "grad_norm": 1.8978221416473389, "learning_rate": 0.00023289269256162188, "loss": 4.9206, "step": 366150 }, { "epoch": 2.208125806490515, "grad_norm": 2.1662213802337646, "learning_rate": 0.00023287573226504803, "loss": 4.8171, "step": 366200 }, { "epoch": 2.208427298271849, "grad_norm": 1.60488760471344, "learning_rate": 0.00023285877044326465, "loss": 4.8593, "step": 366250 }, { "epoch": 2.208728790053183, "grad_norm": 2.369823694229126, "learning_rate": 0.0002328418070965838, "loss": 4.524, "step": 366300 }, { "epoch": 2.2090302818345173, "grad_norm": 1.9393819570541382, "learning_rate": 0.00023282484222531777, "loss": 5.0486, "step": 366350 }, { "epoch": 2.209331773615851, "grad_norm": 0.40544798970222473, "learning_rate": 0.00023280787582977868, "loss": 4.9378, "step": 366400 }, { "epoch": 2.2096332653971853, "grad_norm": 2.5437445640563965, "learning_rate": 0.0002327909079102788, "loss": 5.0356, "step": 366450 }, { "epoch": 2.2099347571785195, "grad_norm": 1.3147473335266113, "learning_rate": 0.00023277393846713043, "loss": 4.9209, "step": 366500 }, { "epoch": 2.2102362489598533, "grad_norm": 1.3816112279891968, "learning_rate": 0.00023275696750064582, "loss": 5.2525, "step": 366550 }, { "epoch": 2.2105377407411875, "grad_norm": 1.3034782409667969, "learning_rate": 0.00023273999501113735, "loss": 5.3558, "step": 366600 }, { "epoch": 2.2108392325225212, "grad_norm": 1.7079001665115356, "learning_rate": 0.00023272302099891734, "loss": 4.8865, "step": 366650 }, { "epoch": 2.2111407243038554, "grad_norm": 1.4570074081420898, "learning_rate": 0.00023270604546429818, "loss": 4.9696, "step": 366700 }, { "epoch": 2.2114422160851896, "grad_norm": 1.337539553642273, "learning_rate": 0.00023268906840759232, "loss": 5.2328, "step": 366750 }, { "epoch": 2.2117437078665234, "grad_norm": 1.2101855278015137, "learning_rate": 0.0002326720898291121, "loss": 5.2164, "step": 366800 }, { "epoch": 2.2120451996478576, "grad_norm": 1.763978123664856, "learning_rate": 0.00023265510972917005, "loss": 5.3448, "step": 366850 }, { "epoch": 2.212346691429192, "grad_norm": 0.4170485734939575, "learning_rate": 0.00023263812810807866, "loss": 4.8422, "step": 366900 }, { "epoch": 2.2126481832105256, "grad_norm": 2.463179111480713, "learning_rate": 0.0002326211449661505, "loss": 4.933, "step": 366950 }, { "epoch": 2.21294967499186, "grad_norm": 1.481269121170044, "learning_rate": 0.00023260416030369805, "loss": 5.0645, "step": 367000 }, { "epoch": 2.2132511667731936, "grad_norm": 1.3228386640548706, "learning_rate": 0.0002325871741210339, "loss": 5.1966, "step": 367050 }, { "epoch": 2.2135526585545278, "grad_norm": 1.7933412790298462, "learning_rate": 0.00023257018641847067, "loss": 5.053, "step": 367100 }, { "epoch": 2.213854150335862, "grad_norm": 1.460490107536316, "learning_rate": 0.000232553197196321, "loss": 5.311, "step": 367150 }, { "epoch": 2.2141556421171957, "grad_norm": 0.5343098640441895, "learning_rate": 0.00023253620645489754, "loss": 4.4459, "step": 367200 }, { "epoch": 2.21445713389853, "grad_norm": 0.9823309779167175, "learning_rate": 0.000232519214194513, "loss": 5.4581, "step": 367250 }, { "epoch": 2.214758625679864, "grad_norm": 1.2287297248840332, "learning_rate": 0.00023250222041548008, "loss": 5.2577, "step": 367300 }, { "epoch": 2.215060117461198, "grad_norm": 1.5692847967147827, "learning_rate": 0.00023248522511811148, "loss": 4.7242, "step": 367350 }, { "epoch": 2.215361609242532, "grad_norm": 1.231514573097229, "learning_rate": 0.00023246822830272007, "loss": 5.0051, "step": 367400 }, { "epoch": 2.215663101023866, "grad_norm": 1.9410039186477661, "learning_rate": 0.00023245122996961865, "loss": 4.7178, "step": 367450 }, { "epoch": 2.2159645928052, "grad_norm": 1.9373217821121216, "learning_rate": 0.00023243423011911994, "loss": 5.0379, "step": 367500 }, { "epoch": 2.2162660845865343, "grad_norm": 1.3920196294784546, "learning_rate": 0.00023241722875153684, "loss": 5.1225, "step": 367550 }, { "epoch": 2.216567576367868, "grad_norm": 1.4491792917251587, "learning_rate": 0.00023240022586718236, "loss": 5.0494, "step": 367600 }, { "epoch": 2.2168690681492023, "grad_norm": 2.046279191970825, "learning_rate": 0.00023238322146636924, "loss": 5.2256, "step": 367650 }, { "epoch": 2.2171705599305365, "grad_norm": 1.7255593538284302, "learning_rate": 0.00023236621554941053, "loss": 5.3093, "step": 367700 }, { "epoch": 2.2174720517118702, "grad_norm": 2.0333251953125, "learning_rate": 0.00023234920811661917, "loss": 5.2848, "step": 367750 }, { "epoch": 2.2177735434932044, "grad_norm": 2.235109806060791, "learning_rate": 0.0002323321991683081, "loss": 5.2484, "step": 367800 }, { "epoch": 2.2180750352745386, "grad_norm": 1.8084259033203125, "learning_rate": 0.00023231518870479043, "loss": 5.3875, "step": 367850 }, { "epoch": 2.2183765270558724, "grad_norm": 2.0217905044555664, "learning_rate": 0.00023229817672637922, "loss": 4.89, "step": 367900 }, { "epoch": 2.2186780188372066, "grad_norm": 1.3749279975891113, "learning_rate": 0.00023228116323338745, "loss": 4.9681, "step": 367950 }, { "epoch": 2.2189795106185404, "grad_norm": 1.3139586448669434, "learning_rate": 0.00023226414822612835, "loss": 5.3038, "step": 368000 }, { "epoch": 2.2192810023998746, "grad_norm": 2.536442756652832, "learning_rate": 0.00023224713170491495, "loss": 5.1528, "step": 368050 }, { "epoch": 2.219582494181209, "grad_norm": 2.8088810443878174, "learning_rate": 0.00023223011367006047, "loss": 5.049, "step": 368100 }, { "epoch": 2.2198839859625425, "grad_norm": 1.5082322359085083, "learning_rate": 0.00023221309412187812, "loss": 5.1546, "step": 368150 }, { "epoch": 2.2201854777438768, "grad_norm": 1.521449089050293, "learning_rate": 0.00023219607306068112, "loss": 4.6483, "step": 368200 }, { "epoch": 2.2204869695252105, "grad_norm": 1.7469232082366943, "learning_rate": 0.00023217905048678262, "loss": 5.1714, "step": 368250 }, { "epoch": 2.2207884613065447, "grad_norm": 1.6509912014007568, "learning_rate": 0.00023216202640049605, "loss": 4.9608, "step": 368300 }, { "epoch": 2.221089953087879, "grad_norm": 1.5829507112503052, "learning_rate": 0.00023214500080213457, "loss": 5.4671, "step": 368350 }, { "epoch": 2.2213914448692127, "grad_norm": 1.312660574913025, "learning_rate": 0.00023212797369201162, "loss": 4.9411, "step": 368400 }, { "epoch": 2.221692936650547, "grad_norm": 1.1276296377182007, "learning_rate": 0.0002321109450704405, "loss": 5.0895, "step": 368450 }, { "epoch": 2.221994428431881, "grad_norm": 1.2075295448303223, "learning_rate": 0.00023209391493773465, "loss": 5.0671, "step": 368500 }, { "epoch": 2.222295920213215, "grad_norm": 1.6542305946350098, "learning_rate": 0.0002320768832942074, "loss": 5.017, "step": 368550 }, { "epoch": 2.222597411994549, "grad_norm": 1.7560713291168213, "learning_rate": 0.00023205985014017224, "loss": 5.2697, "step": 368600 }, { "epoch": 2.2228989037758833, "grad_norm": 1.6801178455352783, "learning_rate": 0.00023204281547594264, "loss": 5.4222, "step": 368650 }, { "epoch": 2.223200395557217, "grad_norm": 1.3605215549468994, "learning_rate": 0.00023202577930183217, "loss": 4.8341, "step": 368700 }, { "epoch": 2.2235018873385513, "grad_norm": 0.8987089395523071, "learning_rate": 0.00023200874161815425, "loss": 4.7787, "step": 368750 }, { "epoch": 2.223803379119885, "grad_norm": 2.4649276733398438, "learning_rate": 0.00023199170242522247, "loss": 4.7151, "step": 368800 }, { "epoch": 2.224104870901219, "grad_norm": 2.2453222274780273, "learning_rate": 0.00023197466172335043, "loss": 4.9119, "step": 368850 }, { "epoch": 2.2244063626825534, "grad_norm": 1.6944475173950195, "learning_rate": 0.00023195761951285172, "loss": 5.0546, "step": 368900 }, { "epoch": 2.224707854463887, "grad_norm": 0.801544189453125, "learning_rate": 0.00023194057579403992, "loss": 5.0154, "step": 368950 }, { "epoch": 2.2250093462452214, "grad_norm": 1.4473950862884521, "learning_rate": 0.00023192353056722886, "loss": 5.1885, "step": 369000 }, { "epoch": 2.2250093462452214, "eval_loss": 5.4897356033325195, "eval_runtime": 39.1622, "eval_samples_per_second": 13.074, "eval_steps_per_second": 6.537, "eval_tts_loss": 7.305650014571633, "step": 369000 }, { "epoch": 2.225310838026555, "grad_norm": 0.2999698221683502, "learning_rate": 0.00023190648383273208, "loss": 5.1933, "step": 369050 }, { "epoch": 2.2256123298078894, "grad_norm": 0.8405106663703918, "learning_rate": 0.00023188943559086333, "loss": 4.8361, "step": 369100 }, { "epoch": 2.2259138215892236, "grad_norm": 1.7806825637817383, "learning_rate": 0.00023187238584193644, "loss": 5.2148, "step": 369150 }, { "epoch": 2.2262153133705573, "grad_norm": 2.08113956451416, "learning_rate": 0.00023185533458626507, "loss": 5.5143, "step": 369200 }, { "epoch": 2.2265168051518915, "grad_norm": 1.3685336112976074, "learning_rate": 0.0002318382818241631, "loss": 4.9776, "step": 369250 }, { "epoch": 2.2268182969332257, "grad_norm": 1.6902269124984741, "learning_rate": 0.00023182122755594436, "loss": 4.9673, "step": 369300 }, { "epoch": 2.2271197887145595, "grad_norm": 0.8615409135818481, "learning_rate": 0.00023180417178192268, "loss": 5.2486, "step": 369350 }, { "epoch": 2.2274212804958937, "grad_norm": 1.5358864068984985, "learning_rate": 0.00023178711450241196, "loss": 4.9602, "step": 369400 }, { "epoch": 2.227722772277228, "grad_norm": 2.264754295349121, "learning_rate": 0.0002317700557177261, "loss": 4.7393, "step": 369450 }, { "epoch": 2.2280242640585617, "grad_norm": 0.5941570997238159, "learning_rate": 0.00023175299542817907, "loss": 5.162, "step": 369500 }, { "epoch": 2.228325755839896, "grad_norm": 2.0684654712677, "learning_rate": 0.00023173593363408481, "loss": 5.106, "step": 369550 }, { "epoch": 2.2286272476212297, "grad_norm": 1.0901967287063599, "learning_rate": 0.00023171887033575735, "loss": 5.4077, "step": 369600 }, { "epoch": 2.228928739402564, "grad_norm": 2.205622673034668, "learning_rate": 0.0002317018055335107, "loss": 4.6993, "step": 369650 }, { "epoch": 2.229230231183898, "grad_norm": 1.6395814418792725, "learning_rate": 0.00023168473922765892, "loss": 5.0907, "step": 369700 }, { "epoch": 2.229531722965232, "grad_norm": 2.2599120140075684, "learning_rate": 0.00023166767141851607, "loss": 5.3749, "step": 369750 }, { "epoch": 2.229833214746566, "grad_norm": 1.3290753364562988, "learning_rate": 0.0002316506021063963, "loss": 4.7934, "step": 369800 }, { "epoch": 2.2301347065279002, "grad_norm": 1.5648866891860962, "learning_rate": 0.0002316335312916137, "loss": 4.8487, "step": 369850 }, { "epoch": 2.230436198309234, "grad_norm": 1.8857125043869019, "learning_rate": 0.00023161645897448244, "loss": 4.8889, "step": 369900 }, { "epoch": 2.230737690090568, "grad_norm": 2.0739424228668213, "learning_rate": 0.00023159938515531674, "loss": 4.9445, "step": 369950 }, { "epoch": 2.231039181871902, "grad_norm": 1.5771440267562866, "learning_rate": 0.00023158230983443082, "loss": 5.1513, "step": 370000 }, { "epoch": 2.231340673653236, "grad_norm": 1.3060270547866821, "learning_rate": 0.00023156523301213886, "loss": 4.7174, "step": 370050 }, { "epoch": 2.2316421654345704, "grad_norm": 1.363120436668396, "learning_rate": 0.00023154815468875528, "loss": 5.0051, "step": 370100 }, { "epoch": 2.231943657215904, "grad_norm": 1.5106806755065918, "learning_rate": 0.0002315310748645942, "loss": 5.1372, "step": 370150 }, { "epoch": 2.2322451489972384, "grad_norm": 2.356579065322876, "learning_rate": 0.0002315139935399701, "loss": 5.2775, "step": 370200 }, { "epoch": 2.2325466407785726, "grad_norm": 1.1896814107894897, "learning_rate": 0.00023149691071519722, "loss": 5.0231, "step": 370250 }, { "epoch": 2.2328481325599063, "grad_norm": 1.5162303447723389, "learning_rate": 0.00023147982639059003, "loss": 5.1783, "step": 370300 }, { "epoch": 2.2331496243412405, "grad_norm": 1.4047366380691528, "learning_rate": 0.0002314627405664629, "loss": 4.585, "step": 370350 }, { "epoch": 2.2334511161225743, "grad_norm": 0.5184338688850403, "learning_rate": 0.0002314456532431303, "loss": 5.3224, "step": 370400 }, { "epoch": 2.2337526079039085, "grad_norm": 0.8764426708221436, "learning_rate": 0.00023142856442090668, "loss": 5.025, "step": 370450 }, { "epoch": 2.2340540996852427, "grad_norm": 1.3108813762664795, "learning_rate": 0.00023141147410010656, "loss": 5.0965, "step": 370500 }, { "epoch": 2.2343555914665765, "grad_norm": 1.3759855031967163, "learning_rate": 0.00023139438228104439, "loss": 5.4, "step": 370550 }, { "epoch": 2.2346570832479107, "grad_norm": 1.8979573249816895, "learning_rate": 0.00023137728896403478, "loss": 4.6327, "step": 370600 }, { "epoch": 2.234958575029245, "grad_norm": 1.9439263343811035, "learning_rate": 0.00023136019414939232, "loss": 5.1296, "step": 370650 }, { "epoch": 2.2352600668105786, "grad_norm": 1.9807525873184204, "learning_rate": 0.00023134309783743158, "loss": 4.8872, "step": 370700 }, { "epoch": 2.235561558591913, "grad_norm": 1.5884846448898315, "learning_rate": 0.0002313260000284672, "loss": 5.1102, "step": 370750 }, { "epoch": 2.2358630503732466, "grad_norm": 1.2675096988677979, "learning_rate": 0.00023130890072281388, "loss": 5.1577, "step": 370800 }, { "epoch": 2.236164542154581, "grad_norm": 1.5022988319396973, "learning_rate": 0.00023129179992078624, "loss": 4.8254, "step": 370850 }, { "epoch": 2.236466033935915, "grad_norm": 2.1286189556121826, "learning_rate": 0.00023127469762269906, "loss": 4.9749, "step": 370900 }, { "epoch": 2.236767525717249, "grad_norm": 1.9727933406829834, "learning_rate": 0.00023125759382886706, "loss": 5.3077, "step": 370950 }, { "epoch": 2.237069017498583, "grad_norm": 1.5632096529006958, "learning_rate": 0.00023124048853960497, "loss": 4.9343, "step": 371000 }, { "epoch": 2.237370509279917, "grad_norm": 1.7488605976104736, "learning_rate": 0.00023122338175522765, "loss": 4.9938, "step": 371050 }, { "epoch": 2.237672001061251, "grad_norm": 1.4790657758712769, "learning_rate": 0.0002312062734760499, "loss": 4.9221, "step": 371100 }, { "epoch": 2.237973492842585, "grad_norm": 2.01053786277771, "learning_rate": 0.00023118916370238652, "loss": 5.061, "step": 371150 }, { "epoch": 2.238274984623919, "grad_norm": 1.6269526481628418, "learning_rate": 0.00023117205243455255, "loss": 5.003, "step": 371200 }, { "epoch": 2.238576476405253, "grad_norm": 1.3228932619094849, "learning_rate": 0.0002311549396728627, "loss": 4.7628, "step": 371250 }, { "epoch": 2.2388779681865874, "grad_norm": 1.9150304794311523, "learning_rate": 0.00023113782541763205, "loss": 5.107, "step": 371300 }, { "epoch": 2.239179459967921, "grad_norm": 2.437839984893799, "learning_rate": 0.0002311207096691755, "loss": 5.2539, "step": 371350 }, { "epoch": 2.2394809517492553, "grad_norm": 1.4322892427444458, "learning_rate": 0.00023110359242780806, "loss": 4.8249, "step": 371400 }, { "epoch": 2.2397824435305895, "grad_norm": 1.6794545650482178, "learning_rate": 0.00023108647369384473, "loss": 5.3518, "step": 371450 }, { "epoch": 2.2400839353119233, "grad_norm": 1.7661176919937134, "learning_rate": 0.0002310693534676006, "loss": 5.1142, "step": 371500 }, { "epoch": 2.2403854270932575, "grad_norm": 1.4620351791381836, "learning_rate": 0.00023105223174939068, "loss": 5.4754, "step": 371550 }, { "epoch": 2.2406869188745917, "grad_norm": 1.4026199579238892, "learning_rate": 0.00023103510853953008, "loss": 5.0011, "step": 371600 }, { "epoch": 2.2409884106559255, "grad_norm": 0.97698974609375, "learning_rate": 0.00023101798383833404, "loss": 4.6316, "step": 371650 }, { "epoch": 2.2412899024372597, "grad_norm": 1.469478964805603, "learning_rate": 0.00023100085764611754, "loss": 4.9457, "step": 371700 }, { "epoch": 2.2415913942185934, "grad_norm": 1.7798900604248047, "learning_rate": 0.00023098372996319583, "loss": 4.8721, "step": 371750 }, { "epoch": 2.2418928859999276, "grad_norm": 1.2553881406784058, "learning_rate": 0.00023096660078988422, "loss": 4.9055, "step": 371800 }, { "epoch": 2.242194377781262, "grad_norm": 1.3327488899230957, "learning_rate": 0.00023094947012649786, "loss": 5.06, "step": 371850 }, { "epoch": 2.2424958695625956, "grad_norm": 1.1830133199691772, "learning_rate": 0.00023093233797335199, "loss": 5.1894, "step": 371900 }, { "epoch": 2.24279736134393, "grad_norm": 2.9531943798065186, "learning_rate": 0.00023091520433076194, "loss": 4.9907, "step": 371950 }, { "epoch": 2.2430988531252636, "grad_norm": 1.7719162702560425, "learning_rate": 0.000230898069199043, "loss": 5.2451, "step": 372000 }, { "epoch": 2.2430988531252636, "eval_loss": 5.468642711639404, "eval_runtime": 39.0539, "eval_samples_per_second": 13.11, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.3293066059154715, "step": 372000 }, { "epoch": 2.243400344906598, "grad_norm": 1.6416304111480713, "learning_rate": 0.0002308809325785106, "loss": 4.9178, "step": 372050 }, { "epoch": 2.243701836687932, "grad_norm": 1.0915031433105469, "learning_rate": 0.00023086379446948002, "loss": 5.0396, "step": 372100 }, { "epoch": 2.2440033284692658, "grad_norm": 1.9712566137313843, "learning_rate": 0.00023084665487226668, "loss": 4.9856, "step": 372150 }, { "epoch": 2.2443048202506, "grad_norm": 1.6633009910583496, "learning_rate": 0.00023082951378718605, "loss": 5.1296, "step": 372200 }, { "epoch": 2.244606312031934, "grad_norm": 1.9096101522445679, "learning_rate": 0.00023081237121455364, "loss": 4.9344, "step": 372250 }, { "epoch": 2.244907803813268, "grad_norm": 2.161292552947998, "learning_rate": 0.00023079522715468477, "loss": 4.9767, "step": 372300 }, { "epoch": 2.245209295594602, "grad_norm": 1.8705167770385742, "learning_rate": 0.00023077808160789512, "loss": 5.052, "step": 372350 }, { "epoch": 2.2455107873759363, "grad_norm": 2.1857287883758545, "learning_rate": 0.0002307609345745001, "loss": 5.205, "step": 372400 }, { "epoch": 2.24581227915727, "grad_norm": 1.461155652999878, "learning_rate": 0.00023074378605481534, "loss": 5.2292, "step": 372450 }, { "epoch": 2.2461137709386043, "grad_norm": 1.5278050899505615, "learning_rate": 0.00023072663604915646, "loss": 4.9785, "step": 372500 }, { "epoch": 2.246415262719938, "grad_norm": 2.1842446327209473, "learning_rate": 0.000230709484557839, "loss": 4.9613, "step": 372550 }, { "epoch": 2.2467167545012723, "grad_norm": 1.6686828136444092, "learning_rate": 0.00023069233158117867, "loss": 4.9097, "step": 372600 }, { "epoch": 2.2470182462826065, "grad_norm": 1.4897608757019043, "learning_rate": 0.00023067517711949115, "loss": 4.9991, "step": 372650 }, { "epoch": 2.2473197380639403, "grad_norm": 1.77566397190094, "learning_rate": 0.00023065802117309216, "loss": 5.1261, "step": 372700 }, { "epoch": 2.2476212298452745, "grad_norm": 1.703369379043579, "learning_rate": 0.0002306408637422973, "loss": 4.9617, "step": 372750 }, { "epoch": 2.2479227216266087, "grad_norm": 1.5678452253341675, "learning_rate": 0.00023062370482742248, "loss": 4.7186, "step": 372800 }, { "epoch": 2.2482242134079424, "grad_norm": 2.12127947807312, "learning_rate": 0.00023060654442878345, "loss": 5.4681, "step": 372850 }, { "epoch": 2.2485257051892766, "grad_norm": 1.3695220947265625, "learning_rate": 0.00023058938254669594, "loss": 4.9367, "step": 372900 }, { "epoch": 2.2488271969706104, "grad_norm": 1.8577362298965454, "learning_rate": 0.00023057221918147588, "loss": 5.1282, "step": 372950 }, { "epoch": 2.2491286887519446, "grad_norm": 1.323050856590271, "learning_rate": 0.0002305550543334391, "loss": 5.4527, "step": 373000 }, { "epoch": 2.249430180533279, "grad_norm": 1.8550374507904053, "learning_rate": 0.00023053788800290152, "loss": 4.7052, "step": 373050 }, { "epoch": 2.2497316723146126, "grad_norm": 1.297593355178833, "learning_rate": 0.00023052072019017907, "loss": 5.1781, "step": 373100 }, { "epoch": 2.250033164095947, "grad_norm": 0.4587338864803314, "learning_rate": 0.0002305035508955876, "loss": 5.0646, "step": 373150 }, { "epoch": 2.250334655877281, "grad_norm": 0.7963524460792542, "learning_rate": 0.0002304863801194432, "loss": 4.8402, "step": 373200 }, { "epoch": 2.2506361476586147, "grad_norm": 2.1234476566314697, "learning_rate": 0.00023046920786206186, "loss": 5.3931, "step": 373250 }, { "epoch": 2.250937639439949, "grad_norm": 2.4067203998565674, "learning_rate": 0.0002304520341237596, "loss": 4.935, "step": 373300 }, { "epoch": 2.2512391312212827, "grad_norm": 1.9181175231933594, "learning_rate": 0.0002304348589048524, "loss": 5.1673, "step": 373350 }, { "epoch": 2.251540623002617, "grad_norm": 1.4382990598678589, "learning_rate": 0.00023041768220565643, "loss": 4.9602, "step": 373400 }, { "epoch": 2.251842114783951, "grad_norm": 1.712393879890442, "learning_rate": 0.0002304005040264878, "loss": 5.1376, "step": 373450 }, { "epoch": 2.252143606565285, "grad_norm": 1.4028478860855103, "learning_rate": 0.0002303833243676626, "loss": 5.1859, "step": 373500 }, { "epoch": 2.252445098346619, "grad_norm": 2.0285584926605225, "learning_rate": 0.00023036614322949705, "loss": 5.0573, "step": 373550 }, { "epoch": 2.252746590127953, "grad_norm": 1.8892557621002197, "learning_rate": 0.00023034896061230735, "loss": 4.9891, "step": 373600 }, { "epoch": 2.253048081909287, "grad_norm": 2.900996208190918, "learning_rate": 0.00023033177651640965, "loss": 5.0268, "step": 373650 }, { "epoch": 2.2533495736906213, "grad_norm": 1.1921203136444092, "learning_rate": 0.00023031459094212026, "loss": 5.1043, "step": 373700 }, { "epoch": 2.2536510654719555, "grad_norm": 1.7642040252685547, "learning_rate": 0.00023029740388975543, "loss": 5.2066, "step": 373750 }, { "epoch": 2.2539525572532892, "grad_norm": 0.9471394419670105, "learning_rate": 0.00023028021535963147, "loss": 5.3608, "step": 373800 }, { "epoch": 2.2542540490346235, "grad_norm": 1.1375837326049805, "learning_rate": 0.00023026302535206474, "loss": 4.7912, "step": 373850 }, { "epoch": 2.254555540815957, "grad_norm": 1.036869764328003, "learning_rate": 0.00023024583386737152, "loss": 5.0483, "step": 373900 }, { "epoch": 2.2548570325972914, "grad_norm": 1.4095361232757568, "learning_rate": 0.00023022864090586827, "loss": 4.8733, "step": 373950 }, { "epoch": 2.2551585243786256, "grad_norm": 2.15440034866333, "learning_rate": 0.00023021144646787135, "loss": 4.6686, "step": 374000 }, { "epoch": 2.2554600161599594, "grad_norm": 1.3382378816604614, "learning_rate": 0.00023019425055369728, "loss": 4.8567, "step": 374050 }, { "epoch": 2.2557615079412936, "grad_norm": 1.8499186038970947, "learning_rate": 0.00023017705316366247, "loss": 5.3034, "step": 374100 }, { "epoch": 2.2560629997226274, "grad_norm": 1.8221927881240845, "learning_rate": 0.00023015985429808338, "loss": 5.028, "step": 374150 }, { "epoch": 2.2563644915039616, "grad_norm": 1.8819751739501953, "learning_rate": 0.00023014265395727656, "loss": 4.8225, "step": 374200 }, { "epoch": 2.2566659832852958, "grad_norm": 0.6755362749099731, "learning_rate": 0.00023012545214155862, "loss": 5.3099, "step": 374250 }, { "epoch": 2.2569674750666295, "grad_norm": 2.5917627811431885, "learning_rate": 0.000230108248851246, "loss": 5.0526, "step": 374300 }, { "epoch": 2.2572689668479637, "grad_norm": 1.8511402606964111, "learning_rate": 0.00023009104408665542, "loss": 5.1912, "step": 374350 }, { "epoch": 2.257570458629298, "grad_norm": 1.552215576171875, "learning_rate": 0.00023007383784810347, "loss": 5.1308, "step": 374400 }, { "epoch": 2.2578719504106317, "grad_norm": 1.7380503416061401, "learning_rate": 0.00023005663013590685, "loss": 5.0856, "step": 374450 }, { "epoch": 2.258173442191966, "grad_norm": 2.2557590007781982, "learning_rate": 0.0002300394209503821, "loss": 4.6773, "step": 374500 }, { "epoch": 2.2584749339733, "grad_norm": 1.4049394130706787, "learning_rate": 0.0002300222102918461, "loss": 5.4517, "step": 374550 }, { "epoch": 2.258776425754634, "grad_norm": 0.6209449768066406, "learning_rate": 0.0002300049981606155, "loss": 5.0686, "step": 374600 }, { "epoch": 2.259077917535968, "grad_norm": 1.3973045349121094, "learning_rate": 0.0002299877845570071, "loss": 5.1873, "step": 374650 }, { "epoch": 2.259379409317302, "grad_norm": 1.8414249420166016, "learning_rate": 0.00022997056948133765, "loss": 4.5819, "step": 374700 }, { "epoch": 2.259680901098636, "grad_norm": 1.7334054708480835, "learning_rate": 0.000229953352933924, "loss": 5.0557, "step": 374750 }, { "epoch": 2.2599823928799703, "grad_norm": 2.2195887565612793, "learning_rate": 0.00022993613491508301, "loss": 5.166, "step": 374800 }, { "epoch": 2.260283884661304, "grad_norm": 1.731540322303772, "learning_rate": 0.00022991891542513152, "loss": 5.5443, "step": 374850 }, { "epoch": 2.2605853764426382, "grad_norm": 1.7794809341430664, "learning_rate": 0.00022990169446438644, "loss": 4.4637, "step": 374900 }, { "epoch": 2.260886868223972, "grad_norm": 1.5573879480361938, "learning_rate": 0.00022988447203316467, "loss": 5.0016, "step": 374950 }, { "epoch": 2.261188360005306, "grad_norm": 1.538473129272461, "learning_rate": 0.00022986724813178321, "loss": 4.4229, "step": 375000 }, { "epoch": 2.261188360005306, "eval_loss": 5.4601240158081055, "eval_runtime": 38.9419, "eval_samples_per_second": 13.148, "eval_steps_per_second": 6.574, "eval_tts_loss": 7.34110348052355, "step": 375000 }, { "epoch": 2.2614898517866404, "grad_norm": 1.4616340398788452, "learning_rate": 0.0002298500227605591, "loss": 4.7596, "step": 375050 }, { "epoch": 2.261791343567974, "grad_norm": 0.8524361252784729, "learning_rate": 0.0002298327959198091, "loss": 5.1244, "step": 375100 }, { "epoch": 2.2620928353493084, "grad_norm": 1.312134027481079, "learning_rate": 0.00022981556760985058, "loss": 5.1987, "step": 375150 }, { "epoch": 2.2623943271306426, "grad_norm": 0.7268648147583008, "learning_rate": 0.00022979833783100037, "loss": 4.5664, "step": 375200 }, { "epoch": 2.2626958189119764, "grad_norm": 2.235886573791504, "learning_rate": 0.00022978110658357565, "loss": 4.8226, "step": 375250 }, { "epoch": 2.2629973106933106, "grad_norm": 1.297741174697876, "learning_rate": 0.0002297638738678935, "loss": 5.2135, "step": 375300 }, { "epoch": 2.2632988024746448, "grad_norm": 0.9171215295791626, "learning_rate": 0.0002297466396842711, "loss": 4.6838, "step": 375350 }, { "epoch": 2.2636002942559785, "grad_norm": 1.3238636255264282, "learning_rate": 0.0002297294040330256, "loss": 4.6646, "step": 375400 }, { "epoch": 2.2639017860373127, "grad_norm": 1.3399871587753296, "learning_rate": 0.00022971216691447418, "loss": 4.5686, "step": 375450 }, { "epoch": 2.2642032778186465, "grad_norm": 1.454264521598816, "learning_rate": 0.00022969492832893408, "loss": 5.2558, "step": 375500 }, { "epoch": 2.2645047695999807, "grad_norm": 1.4319590330123901, "learning_rate": 0.0002296776882767226, "loss": 4.7962, "step": 375550 }, { "epoch": 2.264806261381315, "grad_norm": 2.1928091049194336, "learning_rate": 0.00022966044675815694, "loss": 5.1003, "step": 375600 }, { "epoch": 2.2651077531626487, "grad_norm": 1.5411471128463745, "learning_rate": 0.00022964320377355445, "loss": 5.0505, "step": 375650 }, { "epoch": 2.265409244943983, "grad_norm": 1.9919911623001099, "learning_rate": 0.0002296259593232324, "loss": 4.7852, "step": 375700 }, { "epoch": 2.2657107367253166, "grad_norm": 3.055171489715576, "learning_rate": 0.0002296087134075083, "loss": 4.9049, "step": 375750 }, { "epoch": 2.266012228506651, "grad_norm": 1.5948867797851562, "learning_rate": 0.00022959146602669937, "loss": 4.8923, "step": 375800 }, { "epoch": 2.266313720287985, "grad_norm": 1.2724887132644653, "learning_rate": 0.00022957421718112312, "loss": 4.9256, "step": 375850 }, { "epoch": 2.266615212069319, "grad_norm": 2.480879545211792, "learning_rate": 0.00022955696687109695, "loss": 5.1958, "step": 375900 }, { "epoch": 2.266916703850653, "grad_norm": 1.993470311164856, "learning_rate": 0.0002295397150969383, "loss": 4.6453, "step": 375950 }, { "epoch": 2.2672181956319872, "grad_norm": 1.5992623567581177, "learning_rate": 0.00022952246185896474, "loss": 4.863, "step": 376000 }, { "epoch": 2.267519687413321, "grad_norm": 1.9073699712753296, "learning_rate": 0.0002295052071574938, "loss": 5.1434, "step": 376050 }, { "epoch": 2.267821179194655, "grad_norm": 2.644749164581299, "learning_rate": 0.00022948795099284288, "loss": 5.3932, "step": 376100 }, { "epoch": 2.2681226709759894, "grad_norm": 1.3315736055374146, "learning_rate": 0.0002294706933653297, "loss": 5.2449, "step": 376150 }, { "epoch": 2.268424162757323, "grad_norm": 1.8184144496917725, "learning_rate": 0.00022945343427527184, "loss": 5.0867, "step": 376200 }, { "epoch": 2.2687256545386574, "grad_norm": 2.3973748683929443, "learning_rate": 0.00022943617372298691, "loss": 5.0788, "step": 376250 }, { "epoch": 2.269027146319991, "grad_norm": 1.279107689857483, "learning_rate": 0.00022941891170879253, "loss": 5.2012, "step": 376300 }, { "epoch": 2.2693286381013253, "grad_norm": 1.6257474422454834, "learning_rate": 0.00022940164823300648, "loss": 4.9311, "step": 376350 }, { "epoch": 2.2696301298826596, "grad_norm": 1.2722105979919434, "learning_rate": 0.00022938438329594637, "loss": 5.3624, "step": 376400 }, { "epoch": 2.2699316216639933, "grad_norm": 1.4243887662887573, "learning_rate": 0.00022936711689792994, "loss": 4.8028, "step": 376450 }, { "epoch": 2.2702331134453275, "grad_norm": 1.5374045372009277, "learning_rate": 0.00022934984903927506, "loss": 5.291, "step": 376500 }, { "epoch": 2.2705346052266613, "grad_norm": 2.397887945175171, "learning_rate": 0.00022933257972029936, "loss": 5.2557, "step": 376550 }, { "epoch": 2.2708360970079955, "grad_norm": 2.2682034969329834, "learning_rate": 0.0002293153089413208, "loss": 5.2014, "step": 376600 }, { "epoch": 2.2711375887893297, "grad_norm": 1.37668776512146, "learning_rate": 0.00022929803670265718, "loss": 4.9187, "step": 376650 }, { "epoch": 2.2714390805706635, "grad_norm": 1.3608359098434448, "learning_rate": 0.00022928076300462633, "loss": 5.3949, "step": 376700 }, { "epoch": 2.2717405723519977, "grad_norm": 2.6159822940826416, "learning_rate": 0.00022926348784754616, "loss": 5.1864, "step": 376750 }, { "epoch": 2.272042064133332, "grad_norm": 1.8923958539962769, "learning_rate": 0.0002292462112317346, "loss": 5.0281, "step": 376800 }, { "epoch": 2.2723435559146656, "grad_norm": 2.573577642440796, "learning_rate": 0.0002292289331575097, "loss": 4.9038, "step": 376850 }, { "epoch": 2.272645047696, "grad_norm": 2.3802504539489746, "learning_rate": 0.0002292116536251893, "loss": 4.8962, "step": 376900 }, { "epoch": 2.272946539477334, "grad_norm": 1.3071328401565552, "learning_rate": 0.00022919437263509143, "loss": 5.2475, "step": 376950 }, { "epoch": 2.273248031258668, "grad_norm": 0.6621105670928955, "learning_rate": 0.00022917709018753418, "loss": 4.7413, "step": 377000 }, { "epoch": 2.273549523040002, "grad_norm": 1.8168185949325562, "learning_rate": 0.00022915980628283555, "loss": 5.3028, "step": 377050 }, { "epoch": 2.273851014821336, "grad_norm": 2.0765576362609863, "learning_rate": 0.00022914252092131365, "loss": 4.5556, "step": 377100 }, { "epoch": 2.27415250660267, "grad_norm": 1.9741816520690918, "learning_rate": 0.00022912523410328661, "loss": 4.9498, "step": 377150 }, { "epoch": 2.274453998384004, "grad_norm": 0.9563419818878174, "learning_rate": 0.00022910794582907257, "loss": 4.577, "step": 377200 }, { "epoch": 2.274755490165338, "grad_norm": 1.6480600833892822, "learning_rate": 0.00022909065609898966, "loss": 4.8328, "step": 377250 }, { "epoch": 2.275056981946672, "grad_norm": 1.6111620664596558, "learning_rate": 0.0002290733649133561, "loss": 5.0503, "step": 377300 }, { "epoch": 2.275358473728006, "grad_norm": 1.2724562883377075, "learning_rate": 0.0002290560722724901, "loss": 5.147, "step": 377350 }, { "epoch": 2.27565996550934, "grad_norm": 1.6794462203979492, "learning_rate": 0.00022903877817670993, "loss": 5.1528, "step": 377400 }, { "epoch": 2.2759614572906743, "grad_norm": 1.9021284580230713, "learning_rate": 0.00022902148262633385, "loss": 4.8072, "step": 377450 }, { "epoch": 2.2762629490720085, "grad_norm": 2.3140833377838135, "learning_rate": 0.00022900418562168013, "loss": 4.9958, "step": 377500 }, { "epoch": 2.2765644408533423, "grad_norm": 1.2797785997390747, "learning_rate": 0.00022898688716306706, "loss": 4.57, "step": 377550 }, { "epoch": 2.2768659326346765, "grad_norm": 1.3735511302947998, "learning_rate": 0.00022896958725081315, "loss": 4.7449, "step": 377600 }, { "epoch": 2.2771674244160103, "grad_norm": 1.4966816902160645, "learning_rate": 0.00022895228588523664, "loss": 5.2881, "step": 377650 }, { "epoch": 2.2774689161973445, "grad_norm": 1.8265821933746338, "learning_rate": 0.0002289349830666559, "loss": 5.311, "step": 377700 }, { "epoch": 2.2777704079786787, "grad_norm": 2.1518266201019287, "learning_rate": 0.00022891767879538956, "loss": 4.6187, "step": 377750 }, { "epoch": 2.2780718997600125, "grad_norm": 2.2436397075653076, "learning_rate": 0.00022890037307175595, "loss": 5.019, "step": 377800 }, { "epoch": 2.2783733915413467, "grad_norm": 0.4193081259727478, "learning_rate": 0.00022888306589607355, "loss": 4.7316, "step": 377850 }, { "epoch": 2.2786748833226804, "grad_norm": 1.7674351930618286, "learning_rate": 0.00022886575726866086, "loss": 5.2383, "step": 377900 }, { "epoch": 2.2789763751040146, "grad_norm": 1.5062251091003418, "learning_rate": 0.00022884844718983651, "loss": 4.9272, "step": 377950 }, { "epoch": 2.279277866885349, "grad_norm": 1.9072846174240112, "learning_rate": 0.00022883113565991899, "loss": 5.6022, "step": 378000 }, { "epoch": 2.279277866885349, "eval_loss": 5.474540710449219, "eval_runtime": 38.8514, "eval_samples_per_second": 13.178, "eval_steps_per_second": 6.589, "eval_tts_loss": 7.332439714644272, "step": 378000 }, { "epoch": 2.2795793586666826, "grad_norm": 1.681265950202942, "learning_rate": 0.00022881382267922693, "loss": 5.1339, "step": 378050 }, { "epoch": 2.279880850448017, "grad_norm": 1.7187031507492065, "learning_rate": 0.0002287965082480789, "loss": 5.227, "step": 378100 }, { "epoch": 2.280182342229351, "grad_norm": 1.3331811428070068, "learning_rate": 0.00022877919236679358, "loss": 4.7659, "step": 378150 }, { "epoch": 2.2804838340106848, "grad_norm": 2.5745022296905518, "learning_rate": 0.00022876187503568967, "loss": 4.4922, "step": 378200 }, { "epoch": 2.280785325792019, "grad_norm": 1.6079704761505127, "learning_rate": 0.00022874455625508587, "loss": 5.0892, "step": 378250 }, { "epoch": 2.281086817573353, "grad_norm": 1.576027750968933, "learning_rate": 0.00022872723602530086, "loss": 5.0827, "step": 378300 }, { "epoch": 2.281388309354687, "grad_norm": 0.9493857622146606, "learning_rate": 0.0002287099143466534, "loss": 5.2232, "step": 378350 }, { "epoch": 2.281689801136021, "grad_norm": 1.1970932483673096, "learning_rate": 0.00022869259121946237, "loss": 5.0483, "step": 378400 }, { "epoch": 2.281991292917355, "grad_norm": 1.2895119190216064, "learning_rate": 0.00022867526664404646, "loss": 4.793, "step": 378450 }, { "epoch": 2.282292784698689, "grad_norm": 2.0715816020965576, "learning_rate": 0.0002286579406207245, "loss": 5.2372, "step": 378500 }, { "epoch": 2.2825942764800233, "grad_norm": 1.6275333166122437, "learning_rate": 0.00022864061314981543, "loss": 4.6712, "step": 378550 }, { "epoch": 2.282895768261357, "grad_norm": 1.5946532487869263, "learning_rate": 0.0002286232842316381, "loss": 5.0573, "step": 378600 }, { "epoch": 2.2831972600426913, "grad_norm": 1.3747557401657104, "learning_rate": 0.00022860595386651144, "loss": 4.7925, "step": 378650 }, { "epoch": 2.283498751824025, "grad_norm": 0.9435725212097168, "learning_rate": 0.00022858862205475433, "loss": 4.9166, "step": 378700 }, { "epoch": 2.2838002436053593, "grad_norm": 1.339374303817749, "learning_rate": 0.00022857128879668582, "loss": 5.0252, "step": 378750 }, { "epoch": 2.2841017353866935, "grad_norm": 1.9538356065750122, "learning_rate": 0.00022855395409262486, "loss": 5.1024, "step": 378800 }, { "epoch": 2.2844032271680272, "grad_norm": 1.648510217666626, "learning_rate": 0.00022853661794289048, "loss": 4.9742, "step": 378850 }, { "epoch": 2.2847047189493614, "grad_norm": 1.3373067378997803, "learning_rate": 0.0002285192803478017, "loss": 5.0752, "step": 378900 }, { "epoch": 2.2850062107306957, "grad_norm": 1.6765592098236084, "learning_rate": 0.00022850194130767768, "loss": 5.2515, "step": 378950 }, { "epoch": 2.2853077025120294, "grad_norm": 1.7782530784606934, "learning_rate": 0.00022848460082283738, "loss": 5.2193, "step": 379000 }, { "epoch": 2.2856091942933636, "grad_norm": 1.4305062294006348, "learning_rate": 0.00022846725889360007, "loss": 5.1115, "step": 379050 }, { "epoch": 2.285910686074698, "grad_norm": 2.0247535705566406, "learning_rate": 0.00022844991552028478, "loss": 4.9245, "step": 379100 }, { "epoch": 2.2862121778560316, "grad_norm": 0.9407694935798645, "learning_rate": 0.00022843257070321077, "loss": 5.3854, "step": 379150 }, { "epoch": 2.286513669637366, "grad_norm": 0.7721351385116577, "learning_rate": 0.00022841522444269727, "loss": 5.2287, "step": 379200 }, { "epoch": 2.2868151614186996, "grad_norm": 1.1628499031066895, "learning_rate": 0.00022839787673906343, "loss": 5.3243, "step": 379250 }, { "epoch": 2.2871166532000338, "grad_norm": 1.7745988368988037, "learning_rate": 0.00022838052759262853, "loss": 5.1853, "step": 379300 }, { "epoch": 2.287418144981368, "grad_norm": 2.132218837738037, "learning_rate": 0.0002283631770037119, "loss": 4.756, "step": 379350 }, { "epoch": 2.2877196367627017, "grad_norm": 1.6607756614685059, "learning_rate": 0.00022834582497263285, "loss": 5.199, "step": 379400 }, { "epoch": 2.288021128544036, "grad_norm": 1.5723940134048462, "learning_rate": 0.00022832847149971067, "loss": 5.0949, "step": 379450 }, { "epoch": 2.2883226203253697, "grad_norm": 2.395136594772339, "learning_rate": 0.00022831111658526472, "loss": 5.0064, "step": 379500 }, { "epoch": 2.288624112106704, "grad_norm": 1.522002935409546, "learning_rate": 0.0002282937602296145, "loss": 4.7959, "step": 379550 }, { "epoch": 2.288925603888038, "grad_norm": 1.3467274904251099, "learning_rate": 0.00022827640243307931, "loss": 5.1581, "step": 379600 }, { "epoch": 2.289227095669372, "grad_norm": 1.7510703802108765, "learning_rate": 0.00022825904319597864, "loss": 5.0698, "step": 379650 }, { "epoch": 2.289528587450706, "grad_norm": 2.015789270401001, "learning_rate": 0.00022824168251863198, "loss": 4.693, "step": 379700 }, { "epoch": 2.2898300792320403, "grad_norm": 1.8960521221160889, "learning_rate": 0.00022822432040135875, "loss": 4.9241, "step": 379750 }, { "epoch": 2.290131571013374, "grad_norm": 1.2691203355789185, "learning_rate": 0.00022820695684447863, "loss": 5.2302, "step": 379800 }, { "epoch": 2.2904330627947083, "grad_norm": 1.4165239334106445, "learning_rate": 0.000228189591848311, "loss": 5.3271, "step": 379850 }, { "epoch": 2.2907345545760425, "grad_norm": 1.9749979972839355, "learning_rate": 0.00022817222541317555, "loss": 4.6122, "step": 379900 }, { "epoch": 2.2910360463573762, "grad_norm": 2.1014022827148438, "learning_rate": 0.00022815485753939182, "loss": 4.888, "step": 379950 }, { "epoch": 2.2913375381387104, "grad_norm": 1.2117292881011963, "learning_rate": 0.0002281374882272795, "loss": 4.9086, "step": 380000 }, { "epoch": 2.291639029920044, "grad_norm": 1.9808080196380615, "learning_rate": 0.00022812011747715821, "loss": 4.93, "step": 380050 }, { "epoch": 2.2919405217013784, "grad_norm": 1.635260820388794, "learning_rate": 0.00022810274528934762, "loss": 5.3797, "step": 380100 }, { "epoch": 2.2922420134827126, "grad_norm": 1.3995425701141357, "learning_rate": 0.0002280853716641675, "loss": 4.8149, "step": 380150 }, { "epoch": 2.2925435052640464, "grad_norm": 1.1446166038513184, "learning_rate": 0.00022806799660193754, "loss": 5.0323, "step": 380200 }, { "epoch": 2.2928449970453806, "grad_norm": 1.8852794170379639, "learning_rate": 0.0002280506201029775, "loss": 5.2215, "step": 380250 }, { "epoch": 2.2931464888267143, "grad_norm": 2.3831703662872314, "learning_rate": 0.00022803324216760717, "loss": 5.0512, "step": 380300 }, { "epoch": 2.2934479806080486, "grad_norm": 1.772549033164978, "learning_rate": 0.00022801586279614638, "loss": 4.8549, "step": 380350 }, { "epoch": 2.2937494723893828, "grad_norm": 1.9519531726837158, "learning_rate": 0.000227998481988915, "loss": 4.899, "step": 380400 }, { "epoch": 2.2940509641707165, "grad_norm": 3.0600736141204834, "learning_rate": 0.00022798109974623282, "loss": 4.9757, "step": 380450 }, { "epoch": 2.2943524559520507, "grad_norm": 1.1283807754516602, "learning_rate": 0.00022796371606841984, "loss": 4.9864, "step": 380500 }, { "epoch": 2.294653947733385, "grad_norm": 0.9336887001991272, "learning_rate": 0.00022794633095579587, "loss": 5.451, "step": 380550 }, { "epoch": 2.2949554395147187, "grad_norm": 1.4915839433670044, "learning_rate": 0.00022792894440868097, "loss": 5.0174, "step": 380600 }, { "epoch": 2.295256931296053, "grad_norm": 2.1164395809173584, "learning_rate": 0.00022791155642739504, "loss": 4.5302, "step": 380650 }, { "epoch": 2.295558423077387, "grad_norm": 1.4523783922195435, "learning_rate": 0.00022789416701225808, "loss": 5.0454, "step": 380700 }, { "epoch": 2.295859914858721, "grad_norm": 1.8879238367080688, "learning_rate": 0.0002278767761635901, "loss": 5.2863, "step": 380750 }, { "epoch": 2.296161406640055, "grad_norm": 1.5550237894058228, "learning_rate": 0.00022785938388171126, "loss": 4.7876, "step": 380800 }, { "epoch": 2.296462898421389, "grad_norm": 1.67241632938385, "learning_rate": 0.00022784199016694158, "loss": 5.1377, "step": 380850 }, { "epoch": 2.296764390202723, "grad_norm": 0.3672930896282196, "learning_rate": 0.0002278245950196011, "loss": 5.079, "step": 380900 }, { "epoch": 2.2970658819840573, "grad_norm": 1.153046727180481, "learning_rate": 0.00022780719844001004, "loss": 5.2791, "step": 380950 }, { "epoch": 2.297367373765391, "grad_norm": 1.5734608173370361, "learning_rate": 0.00022778980042848854, "loss": 5.0952, "step": 381000 }, { "epoch": 2.297367373765391, "eval_loss": 5.468260765075684, "eval_runtime": 39.1464, "eval_samples_per_second": 13.079, "eval_steps_per_second": 6.54, "eval_tts_loss": 7.354326392551176, "step": 381000 }, { "epoch": 2.2976688655467252, "grad_norm": 1.9277009963989258, "learning_rate": 0.00022777240098535674, "loss": 4.5379, "step": 381050 }, { "epoch": 2.297970357328059, "grad_norm": 1.1206369400024414, "learning_rate": 0.00022775500011093495, "loss": 4.724, "step": 381100 }, { "epoch": 2.298271849109393, "grad_norm": 1.4843305349349976, "learning_rate": 0.00022773759780554329, "loss": 5.2183, "step": 381150 }, { "epoch": 2.2985733408907274, "grad_norm": 1.4133251905441284, "learning_rate": 0.00022772019406950208, "loss": 5.0521, "step": 381200 }, { "epoch": 2.2988748326720616, "grad_norm": 1.3918771743774414, "learning_rate": 0.00022770278890313166, "loss": 5.2401, "step": 381250 }, { "epoch": 2.2991763244533954, "grad_norm": 1.456997275352478, "learning_rate": 0.00022768538230675224, "loss": 4.7004, "step": 381300 }, { "epoch": 2.2994778162347296, "grad_norm": 1.7960971593856812, "learning_rate": 0.00022766797428068423, "loss": 5.172, "step": 381350 }, { "epoch": 2.2997793080160633, "grad_norm": 1.1150504350662231, "learning_rate": 0.00022765056482524802, "loss": 4.9412, "step": 381400 }, { "epoch": 2.3000807997973975, "grad_norm": 1.4418604373931885, "learning_rate": 0.00022763315394076398, "loss": 5.1423, "step": 381450 }, { "epoch": 2.3003822915787318, "grad_norm": 1.8200525045394897, "learning_rate": 0.0002276157416275525, "loss": 5.1706, "step": 381500 }, { "epoch": 2.3006837833600655, "grad_norm": 1.39532470703125, "learning_rate": 0.0002275983278859341, "loss": 4.6649, "step": 381550 }, { "epoch": 2.3009852751413997, "grad_norm": 1.663176417350769, "learning_rate": 0.00022758091271622916, "loss": 4.8595, "step": 381600 }, { "epoch": 2.3012867669227335, "grad_norm": 1.3751016855239868, "learning_rate": 0.00022756349611875826, "loss": 4.7881, "step": 381650 }, { "epoch": 2.3015882587040677, "grad_norm": 1.4746171236038208, "learning_rate": 0.0002275460780938419, "loss": 5.287, "step": 381700 }, { "epoch": 2.301889750485402, "grad_norm": 2.1307263374328613, "learning_rate": 0.00022752865864180067, "loss": 4.9952, "step": 381750 }, { "epoch": 2.3021912422667357, "grad_norm": 1.6823484897613525, "learning_rate": 0.00022751123776295506, "loss": 5.0095, "step": 381800 }, { "epoch": 2.30249273404807, "grad_norm": 1.4107543230056763, "learning_rate": 0.0002274938154576258, "loss": 5.184, "step": 381850 }, { "epoch": 2.302794225829404, "grad_norm": 1.523305058479309, "learning_rate": 0.0002274763917261334, "loss": 4.8785, "step": 381900 }, { "epoch": 2.303095717610738, "grad_norm": 2.122605800628662, "learning_rate": 0.00022745896656879857, "loss": 5.0741, "step": 381950 }, { "epoch": 2.303397209392072, "grad_norm": 1.0900098085403442, "learning_rate": 0.00022744153998594208, "loss": 5.0932, "step": 382000 }, { "epoch": 2.3036987011734062, "grad_norm": 2.0051188468933105, "learning_rate": 0.0002274241119778845, "loss": 4.7752, "step": 382050 }, { "epoch": 2.30400019295474, "grad_norm": 1.808562994003296, "learning_rate": 0.00022740668254494657, "loss": 4.8593, "step": 382100 }, { "epoch": 2.304301684736074, "grad_norm": 1.5388413667678833, "learning_rate": 0.0002273892516874492, "loss": 5.0613, "step": 382150 }, { "epoch": 2.304603176517408, "grad_norm": 1.3165884017944336, "learning_rate": 0.00022737181940571307, "loss": 4.9991, "step": 382200 }, { "epoch": 2.304904668298742, "grad_norm": 0.8441745042800903, "learning_rate": 0.00022735438570005903, "loss": 4.918, "step": 382250 }, { "epoch": 2.3052061600800764, "grad_norm": 2.0328876972198486, "learning_rate": 0.0002273369505708079, "loss": 5.1802, "step": 382300 }, { "epoch": 2.30550765186141, "grad_norm": 1.6723862886428833, "learning_rate": 0.00022731951401828052, "loss": 4.9119, "step": 382350 }, { "epoch": 2.3058091436427444, "grad_norm": 1.4133089780807495, "learning_rate": 0.00022730207604279793, "loss": 5.0526, "step": 382400 }, { "epoch": 2.306110635424078, "grad_norm": 2.212994337081909, "learning_rate": 0.00022728463664468086, "loss": 5.2268, "step": 382450 }, { "epoch": 2.3064121272054123, "grad_norm": 1.1808452606201172, "learning_rate": 0.00022726719582425036, "loss": 5.2416, "step": 382500 }, { "epoch": 2.3067136189867465, "grad_norm": 1.285627007484436, "learning_rate": 0.0002272497535818274, "loss": 5.5899, "step": 382550 }, { "epoch": 2.3070151107680803, "grad_norm": 2.0776851177215576, "learning_rate": 0.00022723230991773298, "loss": 4.6984, "step": 382600 }, { "epoch": 2.3073166025494145, "grad_norm": 2.25101900100708, "learning_rate": 0.00022721486483228808, "loss": 5.4364, "step": 382650 }, { "epoch": 2.3076180943307487, "grad_norm": 2.2253615856170654, "learning_rate": 0.00022719741832581377, "loss": 5.0671, "step": 382700 }, { "epoch": 2.3079195861120825, "grad_norm": 1.1654834747314453, "learning_rate": 0.00022717997039863118, "loss": 4.8493, "step": 382750 }, { "epoch": 2.3082210778934167, "grad_norm": 1.1910480260849, "learning_rate": 0.0002271625210510614, "loss": 4.8793, "step": 382800 }, { "epoch": 2.308522569674751, "grad_norm": 1.2965058088302612, "learning_rate": 0.00022714507028342553, "loss": 5.0942, "step": 382850 }, { "epoch": 2.3088240614560847, "grad_norm": 1.9297722578048706, "learning_rate": 0.00022712761809604468, "loss": 5.1172, "step": 382900 }, { "epoch": 2.309125553237419, "grad_norm": 1.4404512643814087, "learning_rate": 0.0002271101644892401, "loss": 5.1125, "step": 382950 }, { "epoch": 2.3094270450187526, "grad_norm": 2.3794496059417725, "learning_rate": 0.000227092709463333, "loss": 5.0193, "step": 383000 }, { "epoch": 2.309728536800087, "grad_norm": 1.7630923986434937, "learning_rate": 0.00022707525301864461, "loss": 5.0515, "step": 383050 }, { "epoch": 2.310030028581421, "grad_norm": 1.4945002794265747, "learning_rate": 0.00022705779515549615, "loss": 4.9666, "step": 383100 }, { "epoch": 2.310331520362755, "grad_norm": 1.4340473413467407, "learning_rate": 0.00022704033587420895, "loss": 4.9929, "step": 383150 }, { "epoch": 2.310633012144089, "grad_norm": 2.1680004596710205, "learning_rate": 0.00022702287517510433, "loss": 5.074, "step": 383200 }, { "epoch": 2.3109345039254228, "grad_norm": 1.7426700592041016, "learning_rate": 0.0002270054130585036, "loss": 4.6076, "step": 383250 }, { "epoch": 2.311235995706757, "grad_norm": 1.482933521270752, "learning_rate": 0.00022698794952472813, "loss": 5.0707, "step": 383300 }, { "epoch": 2.311537487488091, "grad_norm": 1.7366962432861328, "learning_rate": 0.00022697048457409936, "loss": 4.7393, "step": 383350 }, { "epoch": 2.311838979269425, "grad_norm": 2.1877171993255615, "learning_rate": 0.0002269530182069386, "loss": 4.875, "step": 383400 }, { "epoch": 2.312140471050759, "grad_norm": 1.3771966695785522, "learning_rate": 0.0002269355504235674, "loss": 4.9199, "step": 383450 }, { "epoch": 2.3124419628320934, "grad_norm": 2.825897455215454, "learning_rate": 0.00022691808122430714, "loss": 4.5259, "step": 383500 }, { "epoch": 2.312743454613427, "grad_norm": 2.095332384109497, "learning_rate": 0.00022690061060947936, "loss": 5.0114, "step": 383550 }, { "epoch": 2.3130449463947613, "grad_norm": 1.0812441110610962, "learning_rate": 0.00022688313857940564, "loss": 5.3564, "step": 383600 }, { "epoch": 2.3133464381760955, "grad_norm": 2.135931968688965, "learning_rate": 0.00022686566513440746, "loss": 5.1611, "step": 383650 }, { "epoch": 2.3136479299574293, "grad_norm": 2.0232455730438232, "learning_rate": 0.00022684819027480636, "loss": 4.7621, "step": 383700 }, { "epoch": 2.3139494217387635, "grad_norm": 0.48811835050582886, "learning_rate": 0.00022683071400092403, "loss": 4.9077, "step": 383750 }, { "epoch": 2.3142509135200973, "grad_norm": 1.8093029260635376, "learning_rate": 0.00022681323631308203, "loss": 5.3125, "step": 383800 }, { "epoch": 2.3145524053014315, "grad_norm": 1.6753871440887451, "learning_rate": 0.00022679575721160208, "loss": 4.9755, "step": 383850 }, { "epoch": 2.3148538970827657, "grad_norm": 0.44991588592529297, "learning_rate": 0.00022677827669680578, "loss": 5.028, "step": 383900 }, { "epoch": 2.3151553888640994, "grad_norm": 0.6453588604927063, "learning_rate": 0.00022676079476901485, "loss": 5.1206, "step": 383950 }, { "epoch": 2.3154568806454336, "grad_norm": 1.0193272829055786, "learning_rate": 0.00022674331142855106, "loss": 5.3922, "step": 384000 }, { "epoch": 2.3154568806454336, "eval_loss": 5.455594062805176, "eval_runtime": 39.1565, "eval_samples_per_second": 13.076, "eval_steps_per_second": 6.538, "eval_tts_loss": 7.382100135921239, "step": 384000 }, { "epoch": 2.3157583724267674, "grad_norm": 0.36631664633750916, "learning_rate": 0.00022672582667573616, "loss": 4.8693, "step": 384050 }, { "epoch": 2.3160598642081016, "grad_norm": 2.6150591373443604, "learning_rate": 0.00022670834051089183, "loss": 5.2578, "step": 384100 }, { "epoch": 2.316361355989436, "grad_norm": 2.1984424591064453, "learning_rate": 0.00022669085293434, "loss": 5.0606, "step": 384150 }, { "epoch": 2.3166628477707696, "grad_norm": 3.162649154663086, "learning_rate": 0.00022667336394640257, "loss": 5.6042, "step": 384200 }, { "epoch": 2.316964339552104, "grad_norm": 1.765419840812683, "learning_rate": 0.00022665587354740114, "loss": 4.7167, "step": 384250 }, { "epoch": 2.317265831333438, "grad_norm": 2.7795538902282715, "learning_rate": 0.00022663838173765785, "loss": 5.084, "step": 384300 }, { "epoch": 2.3175673231147718, "grad_norm": 1.7729766368865967, "learning_rate": 0.00022662088851749444, "loss": 5.1799, "step": 384350 }, { "epoch": 2.317868814896106, "grad_norm": 1.3657193183898926, "learning_rate": 0.000226603393887233, "loss": 5.0266, "step": 384400 }, { "epoch": 2.31817030667744, "grad_norm": 1.1937580108642578, "learning_rate": 0.00022658589784719537, "loss": 5.0048, "step": 384450 }, { "epoch": 2.318471798458774, "grad_norm": 2.451646089553833, "learning_rate": 0.0002265684003977036, "loss": 4.8508, "step": 384500 }, { "epoch": 2.318773290240108, "grad_norm": 0.23498305678367615, "learning_rate": 0.00022655090153907967, "loss": 5.0559, "step": 384550 }, { "epoch": 2.319074782021442, "grad_norm": 1.6340826749801636, "learning_rate": 0.00022653340127164567, "loss": 5.0354, "step": 384600 }, { "epoch": 2.319376273802776, "grad_norm": 1.4736876487731934, "learning_rate": 0.00022651589959572362, "loss": 5.288, "step": 384650 }, { "epoch": 2.3196777655841103, "grad_norm": 2.351961374282837, "learning_rate": 0.0002264983965116356, "loss": 4.6971, "step": 384700 }, { "epoch": 2.319979257365444, "grad_norm": 2.02006196975708, "learning_rate": 0.00022648089201970378, "loss": 5.1296, "step": 384750 }, { "epoch": 2.3202807491467783, "grad_norm": 1.7770259380340576, "learning_rate": 0.00022646338612025035, "loss": 4.8483, "step": 384800 }, { "epoch": 2.320582240928112, "grad_norm": 1.795157790184021, "learning_rate": 0.00022644587881359734, "loss": 4.9033, "step": 384850 }, { "epoch": 2.3208837327094463, "grad_norm": 2.1385889053344727, "learning_rate": 0.00022642837010006706, "loss": 4.5875, "step": 384900 }, { "epoch": 2.3211852244907805, "grad_norm": 1.298032283782959, "learning_rate": 0.00022641085997998173, "loss": 5.2022, "step": 384950 }, { "epoch": 2.3214867162721147, "grad_norm": 2.147371292114258, "learning_rate": 0.0002263933484536635, "loss": 4.5077, "step": 385000 }, { "epoch": 2.3217882080534484, "grad_norm": 2.2442080974578857, "learning_rate": 0.00022637583552143474, "loss": 5.4494, "step": 385050 }, { "epoch": 2.3220896998347826, "grad_norm": 1.5033931732177734, "learning_rate": 0.00022635832118361772, "loss": 4.9308, "step": 385100 }, { "epoch": 2.3223911916161164, "grad_norm": 1.7755283117294312, "learning_rate": 0.00022634080544053478, "loss": 4.908, "step": 385150 }, { "epoch": 2.3226926833974506, "grad_norm": 0.784486711025238, "learning_rate": 0.00022632328829250826, "loss": 4.9635, "step": 385200 }, { "epoch": 2.322994175178785, "grad_norm": 1.3022226095199585, "learning_rate": 0.00022630576973986052, "loss": 4.8412, "step": 385250 }, { "epoch": 2.3232956669601186, "grad_norm": 2.122169256210327, "learning_rate": 0.00022628824978291394, "loss": 5.2171, "step": 385300 }, { "epoch": 2.323597158741453, "grad_norm": 1.0283857583999634, "learning_rate": 0.00022627072842199103, "loss": 4.8867, "step": 385350 }, { "epoch": 2.3238986505227865, "grad_norm": 1.9200390577316284, "learning_rate": 0.0002262532056574143, "loss": 4.9583, "step": 385400 }, { "epoch": 2.3242001423041208, "grad_norm": 1.7217296361923218, "learning_rate": 0.000226235681489506, "loss": 4.5976, "step": 385450 }, { "epoch": 2.324501634085455, "grad_norm": 1.3701345920562744, "learning_rate": 0.00022621815591858884, "loss": 4.7827, "step": 385500 }, { "epoch": 2.3248031258667887, "grad_norm": 0.5032483339309692, "learning_rate": 0.00022620062894498528, "loss": 5.159, "step": 385550 }, { "epoch": 2.325104617648123, "grad_norm": 2.416299819946289, "learning_rate": 0.00022618310056901788, "loss": 5.2083, "step": 385600 }, { "epoch": 2.325406109429457, "grad_norm": 0.8910948634147644, "learning_rate": 0.00022616557079100928, "loss": 5.1725, "step": 385650 }, { "epoch": 2.325707601210791, "grad_norm": 2.3951034545898438, "learning_rate": 0.00022614803961128196, "loss": 4.9669, "step": 385700 }, { "epoch": 2.326009092992125, "grad_norm": 3.9855456352233887, "learning_rate": 0.00022613050703015867, "loss": 5.339, "step": 385750 }, { "epoch": 2.3263105847734593, "grad_norm": 1.5456827878952026, "learning_rate": 0.0002261129730479621, "loss": 5.2938, "step": 385800 }, { "epoch": 2.326612076554793, "grad_norm": 1.4030948877334595, "learning_rate": 0.00022609543766501482, "loss": 4.7984, "step": 385850 }, { "epoch": 2.3269135683361273, "grad_norm": 1.450782060623169, "learning_rate": 0.00022607790088163962, "loss": 4.984, "step": 385900 }, { "epoch": 2.327215060117461, "grad_norm": 1.814547061920166, "learning_rate": 0.00022606036269815923, "loss": 5.0478, "step": 385950 }, { "epoch": 2.3275165518987952, "grad_norm": 1.034628987312317, "learning_rate": 0.0002260428231148964, "loss": 4.5585, "step": 386000 }, { "epoch": 2.3278180436801295, "grad_norm": 1.3574833869934082, "learning_rate": 0.0002260252821321739, "loss": 5.1944, "step": 386050 }, { "epoch": 2.328119535461463, "grad_norm": 2.390455961227417, "learning_rate": 0.00022600773975031465, "loss": 5.2011, "step": 386100 }, { "epoch": 2.3284210272427974, "grad_norm": 1.273376703262329, "learning_rate": 0.0002259901959696414, "loss": 5.0969, "step": 386150 }, { "epoch": 2.328722519024131, "grad_norm": 1.6957461833953857, "learning_rate": 0.000225972650790477, "loss": 5.0227, "step": 386200 }, { "epoch": 2.3290240108054654, "grad_norm": 2.6349148750305176, "learning_rate": 0.00022595510421314442, "loss": 5.2827, "step": 386250 }, { "epoch": 2.3293255025867996, "grad_norm": 1.4700790643692017, "learning_rate": 0.00022593755623796647, "loss": 4.7311, "step": 386300 }, { "epoch": 2.3296269943681334, "grad_norm": 1.342423439025879, "learning_rate": 0.00022592000686526623, "loss": 4.5844, "step": 386350 }, { "epoch": 2.3299284861494676, "grad_norm": 1.0551280975341797, "learning_rate": 0.00022590245609536663, "loss": 4.736, "step": 386400 }, { "epoch": 2.3302299779308018, "grad_norm": 1.4556893110275269, "learning_rate": 0.00022588490392859054, "loss": 4.7624, "step": 386450 }, { "epoch": 2.3305314697121355, "grad_norm": 2.1546132564544678, "learning_rate": 0.0002258673503652612, "loss": 5.1131, "step": 386500 }, { "epoch": 2.3308329614934697, "grad_norm": 1.5338613986968994, "learning_rate": 0.00022584979540570148, "loss": 5.0379, "step": 386550 }, { "epoch": 2.331134453274804, "grad_norm": 1.9235084056854248, "learning_rate": 0.00022583223905023455, "loss": 5.2954, "step": 386600 }, { "epoch": 2.3314359450561377, "grad_norm": 1.4768192768096924, "learning_rate": 0.00022581468129918345, "loss": 5.1482, "step": 386650 }, { "epoch": 2.331737436837472, "grad_norm": 1.5089786052703857, "learning_rate": 0.00022579712215287136, "loss": 4.7467, "step": 386700 }, { "epoch": 2.3320389286188057, "grad_norm": 0.4101142883300781, "learning_rate": 0.00022577956161162137, "loss": 4.7224, "step": 386750 }, { "epoch": 2.33234042040014, "grad_norm": 1.5088366270065308, "learning_rate": 0.00022576199967575675, "loss": 5.075, "step": 386800 }, { "epoch": 2.332641912181474, "grad_norm": 1.3111358880996704, "learning_rate": 0.0002257444363456006, "loss": 5.0141, "step": 386850 }, { "epoch": 2.332943403962808, "grad_norm": 1.8741462230682373, "learning_rate": 0.0002257268716214762, "loss": 4.9175, "step": 386900 }, { "epoch": 2.333244895744142, "grad_norm": 1.7511807680130005, "learning_rate": 0.0002257093055037068, "loss": 4.8552, "step": 386950 }, { "epoch": 2.333546387525476, "grad_norm": 2.626128911972046, "learning_rate": 0.0002256917379926157, "loss": 4.875, "step": 387000 }, { "epoch": 2.333546387525476, "eval_loss": 5.476101875305176, "eval_runtime": 38.956, "eval_samples_per_second": 13.143, "eval_steps_per_second": 6.572, "eval_tts_loss": 7.362416212374382, "step": 387000 }, { "epoch": 2.33384787930681, "grad_norm": 1.7056037187576294, "learning_rate": 0.00022567416908852612, "loss": 4.7314, "step": 387050 }, { "epoch": 2.3341493710881442, "grad_norm": 1.4231382608413696, "learning_rate": 0.00022565659879176148, "loss": 4.6261, "step": 387100 }, { "epoch": 2.334450862869478, "grad_norm": 1.7188522815704346, "learning_rate": 0.0002256390271026451, "loss": 4.9862, "step": 387150 }, { "epoch": 2.334752354650812, "grad_norm": 1.4170082807540894, "learning_rate": 0.00022562145402150037, "loss": 5.3766, "step": 387200 }, { "epoch": 2.3350538464321464, "grad_norm": 2.50057315826416, "learning_rate": 0.00022560387954865074, "loss": 5.451, "step": 387250 }, { "epoch": 2.33535533821348, "grad_norm": 1.774854302406311, "learning_rate": 0.00022558630368441952, "loss": 4.9452, "step": 387300 }, { "epoch": 2.3356568299948144, "grad_norm": 3.9640095233917236, "learning_rate": 0.00022556872642913033, "loss": 5.0515, "step": 387350 }, { "epoch": 2.3359583217761486, "grad_norm": 1.3424046039581299, "learning_rate": 0.00022555114778310655, "loss": 5.1705, "step": 387400 }, { "epoch": 2.3362598135574824, "grad_norm": 1.9718639850616455, "learning_rate": 0.00022553356774667168, "loss": 5.3654, "step": 387450 }, { "epoch": 2.3365613053388166, "grad_norm": 1.4893274307250977, "learning_rate": 0.0002255159863201493, "loss": 5.1943, "step": 387500 }, { "epoch": 2.3368627971201503, "grad_norm": 1.8397818803787231, "learning_rate": 0.00022549840350386297, "loss": 4.7012, "step": 387550 }, { "epoch": 2.3371642889014845, "grad_norm": 1.3049665689468384, "learning_rate": 0.00022548081929813627, "loss": 5.1783, "step": 387600 }, { "epoch": 2.3374657806828187, "grad_norm": 1.526561975479126, "learning_rate": 0.0002254632337032928, "loss": 4.899, "step": 387650 }, { "epoch": 2.3377672724641525, "grad_norm": 1.4354832172393799, "learning_rate": 0.0002254456467196562, "loss": 5.0352, "step": 387700 }, { "epoch": 2.3380687642454867, "grad_norm": 2.197990655899048, "learning_rate": 0.0002254280583475501, "loss": 5.0183, "step": 387750 }, { "epoch": 2.3383702560268205, "grad_norm": 1.6275888681411743, "learning_rate": 0.00022541046858729833, "loss": 5.0054, "step": 387800 }, { "epoch": 2.3386717478081547, "grad_norm": 1.4703987836837769, "learning_rate": 0.00022539287743922443, "loss": 5.1437, "step": 387850 }, { "epoch": 2.338973239589489, "grad_norm": 0.9098920226097107, "learning_rate": 0.0002253752849036522, "loss": 5.0309, "step": 387900 }, { "epoch": 2.3392747313708226, "grad_norm": 1.4106637239456177, "learning_rate": 0.00022535769098090547, "loss": 4.9598, "step": 387950 }, { "epoch": 2.339576223152157, "grad_norm": 1.3440495729446411, "learning_rate": 0.00022534009567130796, "loss": 4.8977, "step": 388000 }, { "epoch": 2.339877714933491, "grad_norm": 1.9195541143417358, "learning_rate": 0.00022532249897518346, "loss": 4.6456, "step": 388050 }, { "epoch": 2.340179206714825, "grad_norm": 1.7293884754180908, "learning_rate": 0.00022530490089285593, "loss": 4.702, "step": 388100 }, { "epoch": 2.340480698496159, "grad_norm": 1.3160632848739624, "learning_rate": 0.00022528730142464912, "loss": 5.2892, "step": 388150 }, { "epoch": 2.3407821902774932, "grad_norm": 0.7623602747917175, "learning_rate": 0.00022526970057088697, "loss": 5.0981, "step": 388200 }, { "epoch": 2.341083682058827, "grad_norm": 2.6150166988372803, "learning_rate": 0.00022525209833189339, "loss": 4.5803, "step": 388250 }, { "epoch": 2.341385173840161, "grad_norm": 2.2879691123962402, "learning_rate": 0.00022523449470799236, "loss": 4.946, "step": 388300 }, { "epoch": 2.341686665621495, "grad_norm": 2.405651569366455, "learning_rate": 0.00022521688969950779, "loss": 5.0577, "step": 388350 }, { "epoch": 2.341988157402829, "grad_norm": 1.8270117044448853, "learning_rate": 0.00022519928330676375, "loss": 5.2659, "step": 388400 }, { "epoch": 2.3422896491841634, "grad_norm": 0.8319032788276672, "learning_rate": 0.00022518167553008418, "loss": 4.9988, "step": 388450 }, { "epoch": 2.342591140965497, "grad_norm": 0.779445469379425, "learning_rate": 0.00022516406636979315, "loss": 4.7971, "step": 388500 }, { "epoch": 2.3428926327468313, "grad_norm": 1.0221179723739624, "learning_rate": 0.0002251464558262148, "loss": 4.9106, "step": 388550 }, { "epoch": 2.3431941245281656, "grad_norm": 1.8927507400512695, "learning_rate": 0.00022512884389967313, "loss": 4.8937, "step": 388600 }, { "epoch": 2.3434956163094993, "grad_norm": 1.3765991926193237, "learning_rate": 0.00022511123059049226, "loss": 5.2037, "step": 388650 }, { "epoch": 2.3437971080908335, "grad_norm": 2.157902956008911, "learning_rate": 0.00022509361589899646, "loss": 5.0624, "step": 388700 }, { "epoch": 2.3440985998721677, "grad_norm": 1.3158308267593384, "learning_rate": 0.0002250759998255098, "loss": 5.4336, "step": 388750 }, { "epoch": 2.3444000916535015, "grad_norm": 1.6267948150634766, "learning_rate": 0.00022505838237035647, "loss": 5.1417, "step": 388800 }, { "epoch": 2.3447015834348357, "grad_norm": 1.401137113571167, "learning_rate": 0.00022504076353386076, "loss": 5.2933, "step": 388850 }, { "epoch": 2.3450030752161695, "grad_norm": 0.9076070785522461, "learning_rate": 0.00022502314331634686, "loss": 5.0832, "step": 388900 }, { "epoch": 2.3453045669975037, "grad_norm": 1.6587446928024292, "learning_rate": 0.00022500552171813903, "loss": 4.9896, "step": 388950 }, { "epoch": 2.345606058778838, "grad_norm": 1.4904402494430542, "learning_rate": 0.00022498789873956173, "loss": 5.0583, "step": 389000 }, { "epoch": 2.3459075505601716, "grad_norm": 1.9293969869613647, "learning_rate": 0.0002249702743809391, "loss": 4.9771, "step": 389050 }, { "epoch": 2.346209042341506, "grad_norm": 1.731654405593872, "learning_rate": 0.00022495264864259552, "loss": 5.145, "step": 389100 }, { "epoch": 2.3465105341228396, "grad_norm": 1.6018630266189575, "learning_rate": 0.00022493502152485544, "loss": 4.9417, "step": 389150 }, { "epoch": 2.346812025904174, "grad_norm": 2.5062668323516846, "learning_rate": 0.00022491739302804326, "loss": 4.5187, "step": 389200 }, { "epoch": 2.347113517685508, "grad_norm": 1.8865779638290405, "learning_rate": 0.00022489976315248335, "loss": 5.1556, "step": 389250 }, { "epoch": 2.347415009466842, "grad_norm": 1.968945026397705, "learning_rate": 0.00022488213189850018, "loss": 5.1346, "step": 389300 }, { "epoch": 2.347716501248176, "grad_norm": 4.268087863922119, "learning_rate": 0.00022486449926641824, "loss": 5.0653, "step": 389350 }, { "epoch": 2.34801799302951, "grad_norm": 1.4190630912780762, "learning_rate": 0.00022484686525656205, "loss": 4.6137, "step": 389400 }, { "epoch": 2.348319484810844, "grad_norm": 0.7209135890007019, "learning_rate": 0.0002248292298692561, "loss": 5.2114, "step": 389450 }, { "epoch": 2.348620976592178, "grad_norm": 1.7924877405166626, "learning_rate": 0.00022481159310482494, "loss": 4.3659, "step": 389500 }, { "epoch": 2.3489224683735124, "grad_norm": 1.4544042348861694, "learning_rate": 0.00022479395496359316, "loss": 5.0883, "step": 389550 }, { "epoch": 2.349223960154846, "grad_norm": 1.632355809211731, "learning_rate": 0.0002247763154458854, "loss": 5.2353, "step": 389600 }, { "epoch": 2.3495254519361803, "grad_norm": 2.6372196674346924, "learning_rate": 0.0002247586745520263, "loss": 5.0571, "step": 389650 }, { "epoch": 2.349826943717514, "grad_norm": 2.5326905250549316, "learning_rate": 0.0002247410322823404, "loss": 4.8889, "step": 389700 }, { "epoch": 2.3501284354988483, "grad_norm": 0.5645126700401306, "learning_rate": 0.00022472338863715257, "loss": 4.5691, "step": 389750 }, { "epoch": 2.3504299272801825, "grad_norm": 1.8626564741134644, "learning_rate": 0.00022470574361678733, "loss": 4.6341, "step": 389800 }, { "epoch": 2.3507314190615163, "grad_norm": 1.4638503789901733, "learning_rate": 0.0002246880972215695, "loss": 4.8561, "step": 389850 }, { "epoch": 2.3510329108428505, "grad_norm": 1.797676920890808, "learning_rate": 0.00022467044945182385, "loss": 4.9726, "step": 389900 }, { "epoch": 2.3513344026241842, "grad_norm": 1.549825668334961, "learning_rate": 0.0002246528003078751, "loss": 5.0594, "step": 389950 }, { "epoch": 2.3516358944055185, "grad_norm": 1.3784704208374023, "learning_rate": 0.00022463514979004814, "loss": 5.2659, "step": 390000 }, { "epoch": 2.3516358944055185, "eval_loss": 5.4583282470703125, "eval_runtime": 39.0579, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.41317049267895, "step": 390000 }, { "epoch": 2.3519373861868527, "grad_norm": 1.7838120460510254, "learning_rate": 0.00022461749789866772, "loss": 4.8472, "step": 390050 }, { "epoch": 2.3522388779681864, "grad_norm": 1.9096208810806274, "learning_rate": 0.00022459984463405873, "loss": 4.9008, "step": 390100 }, { "epoch": 2.3525403697495206, "grad_norm": 1.378006100654602, "learning_rate": 0.0002245821899965461, "loss": 5.0183, "step": 390150 }, { "epoch": 2.352841861530855, "grad_norm": 1.4802803993225098, "learning_rate": 0.0002245645339864547, "loss": 4.9238, "step": 390200 }, { "epoch": 2.3531433533121886, "grad_norm": 1.0844480991363525, "learning_rate": 0.00022454687660410942, "loss": 4.8955, "step": 390250 }, { "epoch": 2.353444845093523, "grad_norm": 1.5699883699417114, "learning_rate": 0.0002245292178498353, "loss": 4.7469, "step": 390300 }, { "epoch": 2.353746336874857, "grad_norm": 1.4195142984390259, "learning_rate": 0.00022451155772395726, "loss": 4.9275, "step": 390350 }, { "epoch": 2.3540478286561908, "grad_norm": 2.4461262226104736, "learning_rate": 0.00022449389622680035, "loss": 4.9895, "step": 390400 }, { "epoch": 2.354349320437525, "grad_norm": 1.791741967201233, "learning_rate": 0.0002244762333586896, "loss": 5.0709, "step": 390450 }, { "epoch": 2.3546508122188587, "grad_norm": 1.190754771232605, "learning_rate": 0.00022445856911995, "loss": 4.8076, "step": 390500 }, { "epoch": 2.354952304000193, "grad_norm": 1.5935155153274536, "learning_rate": 0.00022444090351090673, "loss": 5.1076, "step": 390550 }, { "epoch": 2.355253795781527, "grad_norm": 2.498333215713501, "learning_rate": 0.0002244232365318849, "loss": 4.8649, "step": 390600 }, { "epoch": 2.355555287562861, "grad_norm": 1.4966236352920532, "learning_rate": 0.00022440556818320952, "loss": 5.0984, "step": 390650 }, { "epoch": 2.355856779344195, "grad_norm": 2.217879056930542, "learning_rate": 0.0002243878984652059, "loss": 4.969, "step": 390700 }, { "epoch": 2.356158271125529, "grad_norm": 1.7635915279388428, "learning_rate": 0.0002243702273781992, "loss": 5.3371, "step": 390750 }, { "epoch": 2.356459762906863, "grad_norm": 1.4859753847122192, "learning_rate": 0.00022435255492251452, "loss": 4.5057, "step": 390800 }, { "epoch": 2.3567612546881973, "grad_norm": 1.7167826890945435, "learning_rate": 0.00022433488109847722, "loss": 4.7572, "step": 390850 }, { "epoch": 2.357062746469531, "grad_norm": 0.6447906494140625, "learning_rate": 0.0002243172059064125, "loss": 4.8064, "step": 390900 }, { "epoch": 2.3573642382508653, "grad_norm": 2.3448426723480225, "learning_rate": 0.00022429952934664566, "loss": 4.7215, "step": 390950 }, { "epoch": 2.3576657300321995, "grad_norm": 1.5273280143737793, "learning_rate": 0.00022428185141950204, "loss": 5.0402, "step": 391000 }, { "epoch": 2.3579672218135332, "grad_norm": 2.8519949913024902, "learning_rate": 0.00022426417212530693, "loss": 5.1781, "step": 391050 }, { "epoch": 2.3582687135948674, "grad_norm": 2.0714762210845947, "learning_rate": 0.0002242464914643857, "loss": 5.2333, "step": 391100 }, { "epoch": 2.3585702053762017, "grad_norm": 1.7447134256362915, "learning_rate": 0.00022422880943706377, "loss": 4.9225, "step": 391150 }, { "epoch": 2.3588716971575354, "grad_norm": 4.439245223999023, "learning_rate": 0.00022421112604366654, "loss": 4.7136, "step": 391200 }, { "epoch": 2.3591731889388696, "grad_norm": 0.9912755489349365, "learning_rate": 0.00022419344128451942, "loss": 5.1776, "step": 391250 }, { "epoch": 2.3594746807202034, "grad_norm": 2.1352365016937256, "learning_rate": 0.00022417575515994788, "loss": 4.9966, "step": 391300 }, { "epoch": 2.3597761725015376, "grad_norm": 1.0273157358169556, "learning_rate": 0.00022415806767027744, "loss": 4.6898, "step": 391350 }, { "epoch": 2.360077664282872, "grad_norm": 1.5872386693954468, "learning_rate": 0.00022414037881583358, "loss": 4.828, "step": 391400 }, { "epoch": 2.3603791560642056, "grad_norm": 1.328079104423523, "learning_rate": 0.00022412268859694188, "loss": 4.9234, "step": 391450 }, { "epoch": 2.3606806478455398, "grad_norm": 2.096014976501465, "learning_rate": 0.00022410499701392786, "loss": 4.4976, "step": 391500 }, { "epoch": 2.3609821396268735, "grad_norm": 2.508197784423828, "learning_rate": 0.00022408730406711715, "loss": 5.296, "step": 391550 }, { "epoch": 2.3612836314082077, "grad_norm": 1.243333339691162, "learning_rate": 0.0002240696097568353, "loss": 5.1138, "step": 391600 }, { "epoch": 2.361585123189542, "grad_norm": 1.4728933572769165, "learning_rate": 0.000224051914083408, "loss": 4.6951, "step": 391650 }, { "epoch": 2.3618866149708757, "grad_norm": 2.0292558670043945, "learning_rate": 0.00022403421704716086, "loss": 5.0537, "step": 391700 }, { "epoch": 2.36218810675221, "grad_norm": 0.5119836330413818, "learning_rate": 0.00022401651864841965, "loss": 5.1902, "step": 391750 }, { "epoch": 2.362489598533544, "grad_norm": 0.45017296075820923, "learning_rate": 0.00022399881888751007, "loss": 5.3035, "step": 391800 }, { "epoch": 2.362791090314878, "grad_norm": 1.4609556198120117, "learning_rate": 0.00022398111776475777, "loss": 5.0184, "step": 391850 }, { "epoch": 2.363092582096212, "grad_norm": 1.4387742280960083, "learning_rate": 0.00022396341528048858, "loss": 5.0634, "step": 391900 }, { "epoch": 2.3633940738775463, "grad_norm": 1.791248083114624, "learning_rate": 0.00022394571143502832, "loss": 5.1301, "step": 391950 }, { "epoch": 2.36369556565888, "grad_norm": 2.6318604946136475, "learning_rate": 0.00022392800622870275, "loss": 5.074, "step": 392000 }, { "epoch": 2.3639970574402143, "grad_norm": 1.4442191123962402, "learning_rate": 0.0002239102996618377, "loss": 5.0416, "step": 392050 }, { "epoch": 2.364298549221548, "grad_norm": 2.0618269443511963, "learning_rate": 0.0002238925917347591, "loss": 5.0673, "step": 392100 }, { "epoch": 2.3646000410028822, "grad_norm": 1.2476677894592285, "learning_rate": 0.0002238748824477928, "loss": 5.3544, "step": 392150 }, { "epoch": 2.3649015327842164, "grad_norm": 0.915245771408081, "learning_rate": 0.0002238571718012647, "loss": 5.3589, "step": 392200 }, { "epoch": 2.36520302456555, "grad_norm": 1.468183994293213, "learning_rate": 0.0002238394597955007, "loss": 4.9837, "step": 392250 }, { "epoch": 2.3655045163468844, "grad_norm": 2.3880228996276855, "learning_rate": 0.0002238217464308269, "loss": 4.5723, "step": 392300 }, { "epoch": 2.3658060081282186, "grad_norm": 1.868654727935791, "learning_rate": 0.0002238040317075692, "loss": 5.3209, "step": 392350 }, { "epoch": 2.3661074999095524, "grad_norm": 1.4170730113983154, "learning_rate": 0.00022378631562605357, "loss": 5.4017, "step": 392400 }, { "epoch": 2.3664089916908866, "grad_norm": 3.8078067302703857, "learning_rate": 0.00022376859818660615, "loss": 5.1234, "step": 392450 }, { "epoch": 2.366710483472221, "grad_norm": 1.3787331581115723, "learning_rate": 0.0002237508793895529, "loss": 5.1965, "step": 392500 }, { "epoch": 2.3670119752535546, "grad_norm": 1.952039361000061, "learning_rate": 0.00022373315923522003, "loss": 5.1143, "step": 392550 }, { "epoch": 2.3673134670348888, "grad_norm": 1.2137529850006104, "learning_rate": 0.00022371543772393352, "loss": 4.7175, "step": 392600 }, { "epoch": 2.3676149588162225, "grad_norm": 0.6323267221450806, "learning_rate": 0.00022369771485601962, "loss": 5.0592, "step": 392650 }, { "epoch": 2.3679164505975567, "grad_norm": 1.3805665969848633, "learning_rate": 0.0002236799906318044, "loss": 5.2873, "step": 392700 }, { "epoch": 2.368217942378891, "grad_norm": 1.9939959049224854, "learning_rate": 0.00022366226505161413, "loss": 4.9901, "step": 392750 }, { "epoch": 2.3685194341602247, "grad_norm": 1.4352816343307495, "learning_rate": 0.00022364453811577503, "loss": 4.8861, "step": 392800 }, { "epoch": 2.368820925941559, "grad_norm": 0.5182697772979736, "learning_rate": 0.00022362680982461328, "loss": 5.0185, "step": 392850 }, { "epoch": 2.3691224177228927, "grad_norm": 1.69892418384552, "learning_rate": 0.00022360908017845512, "loss": 4.8768, "step": 392900 }, { "epoch": 2.369423909504227, "grad_norm": 1.5548583269119263, "learning_rate": 0.00022359134917762695, "loss": 5.0252, "step": 392950 }, { "epoch": 2.369725401285561, "grad_norm": 2.4201743602752686, "learning_rate": 0.000223573616822455, "loss": 4.6571, "step": 393000 }, { "epoch": 2.369725401285561, "eval_loss": 5.458495140075684, "eval_runtime": 39.022, "eval_samples_per_second": 13.121, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.4175249248706505, "step": 393000 }, { "epoch": 2.370026893066895, "grad_norm": 1.5227891206741333, "learning_rate": 0.00022355588311326563, "loss": 4.8124, "step": 393050 }, { "epoch": 2.370328384848229, "grad_norm": 0.8985530138015747, "learning_rate": 0.00022353814805038517, "loss": 4.6639, "step": 393100 }, { "epoch": 2.3706298766295633, "grad_norm": 2.7543694972991943, "learning_rate": 0.00022352041163414005, "loss": 5.299, "step": 393150 }, { "epoch": 2.370931368410897, "grad_norm": 2.0776731967926025, "learning_rate": 0.0002235026738648567, "loss": 5.0147, "step": 393200 }, { "epoch": 2.3712328601922312, "grad_norm": 1.910136103630066, "learning_rate": 0.0002234849347428615, "loss": 5.1016, "step": 393250 }, { "epoch": 2.3715343519735654, "grad_norm": 2.089845657348633, "learning_rate": 0.00022346719426848097, "loss": 4.9173, "step": 393300 }, { "epoch": 2.371835843754899, "grad_norm": 1.0542925596237183, "learning_rate": 0.00022344945244204154, "loss": 5.06, "step": 393350 }, { "epoch": 2.3721373355362334, "grad_norm": 2.144347906112671, "learning_rate": 0.00022343170926386979, "loss": 5.0037, "step": 393400 }, { "epoch": 2.372438827317567, "grad_norm": 1.3191733360290527, "learning_rate": 0.0002234139647342922, "loss": 4.7183, "step": 393450 }, { "epoch": 2.3727403190989014, "grad_norm": 1.339491367340088, "learning_rate": 0.00022339621885363538, "loss": 4.9386, "step": 393500 }, { "epoch": 2.3730418108802356, "grad_norm": 1.3930777311325073, "learning_rate": 0.00022337847162222588, "loss": 5.2725, "step": 393550 }, { "epoch": 2.3733433026615693, "grad_norm": 2.065983533859253, "learning_rate": 0.0002233607230403903, "loss": 5.1162, "step": 393600 }, { "epoch": 2.3736447944429035, "grad_norm": 1.3582336902618408, "learning_rate": 0.00022334297310845535, "loss": 4.7898, "step": 393650 }, { "epoch": 2.3739462862242373, "grad_norm": 2.60538911819458, "learning_rate": 0.00022332522182674764, "loss": 5.0985, "step": 393700 }, { "epoch": 2.3742477780055715, "grad_norm": 1.0194659233093262, "learning_rate": 0.00022330746919559384, "loss": 5.0506, "step": 393750 }, { "epoch": 2.3745492697869057, "grad_norm": 1.916258454322815, "learning_rate": 0.00022328971521532072, "loss": 5.1681, "step": 393800 }, { "epoch": 2.3748507615682395, "grad_norm": 2.0593719482421875, "learning_rate": 0.00022327195988625496, "loss": 5.1829, "step": 393850 }, { "epoch": 2.3751522533495737, "grad_norm": 1.8236057758331299, "learning_rate": 0.00022325420320872333, "loss": 4.9638, "step": 393900 }, { "epoch": 2.375453745130908, "grad_norm": 1.5325239896774292, "learning_rate": 0.00022323644518305264, "loss": 5.0174, "step": 393950 }, { "epoch": 2.3757552369122417, "grad_norm": 1.5230437517166138, "learning_rate": 0.00022321868580956966, "loss": 4.9124, "step": 394000 }, { "epoch": 2.376056728693576, "grad_norm": 1.8094985485076904, "learning_rate": 0.0002232009250886013, "loss": 5.2227, "step": 394050 }, { "epoch": 2.37635822047491, "grad_norm": 1.7121886014938354, "learning_rate": 0.00022318316302047438, "loss": 5.0832, "step": 394100 }, { "epoch": 2.376659712256244, "grad_norm": 2.4937193393707275, "learning_rate": 0.00022316539960551577, "loss": 5.0695, "step": 394150 }, { "epoch": 2.376961204037578, "grad_norm": 1.8117692470550537, "learning_rate": 0.0002231476348440524, "loss": 5.0564, "step": 394200 }, { "epoch": 2.377262695818912, "grad_norm": 1.2811223268508911, "learning_rate": 0.00022312986873641117, "loss": 4.6063, "step": 394250 }, { "epoch": 2.377564187600246, "grad_norm": 1.2505391836166382, "learning_rate": 0.00022311210128291906, "loss": 5.4039, "step": 394300 }, { "epoch": 2.37786567938158, "grad_norm": 1.7125564813613892, "learning_rate": 0.00022309433248390314, "loss": 4.6532, "step": 394350 }, { "epoch": 2.378167171162914, "grad_norm": 0.9771890044212341, "learning_rate": 0.00022307656233969028, "loss": 5.0619, "step": 394400 }, { "epoch": 2.378468662944248, "grad_norm": 1.8670930862426758, "learning_rate": 0.00022305879085060754, "loss": 4.6256, "step": 394450 }, { "epoch": 2.378770154725582, "grad_norm": 1.2435050010681152, "learning_rate": 0.00022304101801698208, "loss": 4.8035, "step": 394500 }, { "epoch": 2.379071646506916, "grad_norm": 1.6015019416809082, "learning_rate": 0.0002230232438391409, "loss": 4.8365, "step": 394550 }, { "epoch": 2.3793731382882504, "grad_norm": 1.3345295190811157, "learning_rate": 0.0002230054683174111, "loss": 5.0215, "step": 394600 }, { "epoch": 2.379674630069584, "grad_norm": 2.213066577911377, "learning_rate": 0.00022298769145211988, "loss": 5.1435, "step": 394650 }, { "epoch": 2.3799761218509183, "grad_norm": 1.470656394958496, "learning_rate": 0.00022296991324359435, "loss": 5.2644, "step": 394700 }, { "epoch": 2.3802776136322525, "grad_norm": 1.193877935409546, "learning_rate": 0.0002229521336921617, "loss": 4.7927, "step": 394750 }, { "epoch": 2.3805791054135863, "grad_norm": 2.097886323928833, "learning_rate": 0.00022293435279814914, "loss": 5.3591, "step": 394800 }, { "epoch": 2.3808805971949205, "grad_norm": 2.5093424320220947, "learning_rate": 0.00022291657056188388, "loss": 5.0754, "step": 394850 }, { "epoch": 2.3811820889762547, "grad_norm": 0.7046369910240173, "learning_rate": 0.00022289878698369318, "loss": 5.1214, "step": 394900 }, { "epoch": 2.3814835807575885, "grad_norm": 1.5362151861190796, "learning_rate": 0.00022288100206390435, "loss": 4.688, "step": 394950 }, { "epoch": 2.3817850725389227, "grad_norm": 1.7615050077438354, "learning_rate": 0.00022286321580284476, "loss": 4.9835, "step": 395000 }, { "epoch": 2.3820865643202564, "grad_norm": 1.4556944370269775, "learning_rate": 0.00022284542820084156, "loss": 4.5323, "step": 395050 }, { "epoch": 2.3823880561015907, "grad_norm": 1.9233101606369019, "learning_rate": 0.00022282763925822224, "loss": 5.3546, "step": 395100 }, { "epoch": 2.382689547882925, "grad_norm": 1.9762436151504517, "learning_rate": 0.0002228098489753142, "loss": 5.0204, "step": 395150 }, { "epoch": 2.3829910396642586, "grad_norm": 2.5815956592559814, "learning_rate": 0.00022279205735244476, "loss": 4.7604, "step": 395200 }, { "epoch": 2.383292531445593, "grad_norm": 1.4116532802581787, "learning_rate": 0.00022277426438994138, "loss": 5.2779, "step": 395250 }, { "epoch": 2.3835940232269266, "grad_norm": 1.4719996452331543, "learning_rate": 0.00022275647008813151, "loss": 5.2798, "step": 395300 }, { "epoch": 2.383895515008261, "grad_norm": 1.6265616416931152, "learning_rate": 0.00022273867444734267, "loss": 5.3852, "step": 395350 }, { "epoch": 2.384197006789595, "grad_norm": 1.8417890071868896, "learning_rate": 0.0002227208774679023, "loss": 5.0093, "step": 395400 }, { "epoch": 2.3844984985709288, "grad_norm": 1.5086075067520142, "learning_rate": 0.00022270307915013795, "loss": 5.2671, "step": 395450 }, { "epoch": 2.384799990352263, "grad_norm": 2.24946665763855, "learning_rate": 0.00022268527949437719, "loss": 5.3585, "step": 395500 }, { "epoch": 2.385101482133597, "grad_norm": 2.232821464538574, "learning_rate": 0.00022266747850094762, "loss": 4.6771, "step": 395550 }, { "epoch": 2.385402973914931, "grad_norm": 1.394866943359375, "learning_rate": 0.0002226496761701768, "loss": 4.9532, "step": 395600 }, { "epoch": 2.385704465696265, "grad_norm": 2.6309196949005127, "learning_rate": 0.00022263187250239232, "loss": 5.0799, "step": 395650 }, { "epoch": 2.3860059574775994, "grad_norm": 0.9458669424057007, "learning_rate": 0.00022261406749792193, "loss": 4.9226, "step": 395700 }, { "epoch": 2.386307449258933, "grad_norm": 1.6398614645004272, "learning_rate": 0.00022259626115709325, "loss": 4.8731, "step": 395750 }, { "epoch": 2.3866089410402673, "grad_norm": 1.4620420932769775, "learning_rate": 0.00022257845348023394, "loss": 5.0085, "step": 395800 }, { "epoch": 2.386910432821601, "grad_norm": 1.4427663087844849, "learning_rate": 0.0002225606444676718, "loss": 5.0037, "step": 395850 }, { "epoch": 2.3872119246029353, "grad_norm": 1.538336992263794, "learning_rate": 0.00022254283411973454, "loss": 5.0335, "step": 395900 }, { "epoch": 2.3875134163842695, "grad_norm": 1.9443747997283936, "learning_rate": 0.00022252502243674997, "loss": 4.9971, "step": 395950 }, { "epoch": 2.3878149081656033, "grad_norm": 1.1237069368362427, "learning_rate": 0.00022250720941904584, "loss": 5.0653, "step": 396000 }, { "epoch": 2.3878149081656033, "eval_loss": 5.448690414428711, "eval_runtime": 39.0624, "eval_samples_per_second": 13.107, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.41064857416295, "step": 396000 }, { "epoch": 2.3881163999469375, "grad_norm": 1.2133551836013794, "learning_rate": 0.00022248939506694997, "loss": 5.1672, "step": 396050 }, { "epoch": 2.3884178917282717, "grad_norm": 2.1376953125, "learning_rate": 0.00022247157938079024, "loss": 4.9147, "step": 396100 }, { "epoch": 2.3887193835096054, "grad_norm": 1.6323051452636719, "learning_rate": 0.00022245376236089455, "loss": 4.9623, "step": 396150 }, { "epoch": 2.3890208752909396, "grad_norm": 1.5657134056091309, "learning_rate": 0.00022243594400759073, "loss": 5.084, "step": 396200 }, { "epoch": 2.389322367072274, "grad_norm": 1.886521577835083, "learning_rate": 0.00022241812432120676, "loss": 5.1645, "step": 396250 }, { "epoch": 2.3896238588536076, "grad_norm": 1.2246155738830566, "learning_rate": 0.00022240030330207054, "loss": 4.7117, "step": 396300 }, { "epoch": 2.389925350634942, "grad_norm": 1.696428894996643, "learning_rate": 0.00022238248095051002, "loss": 4.7199, "step": 396350 }, { "epoch": 2.3902268424162756, "grad_norm": 2.5000226497650146, "learning_rate": 0.00022236465726685332, "loss": 4.9345, "step": 396400 }, { "epoch": 2.39052833419761, "grad_norm": 1.3716243505477905, "learning_rate": 0.0002223468322514283, "loss": 4.7517, "step": 396450 }, { "epoch": 2.390829825978944, "grad_norm": 1.8148773908615112, "learning_rate": 0.00022232900590456309, "loss": 5.2628, "step": 396500 }, { "epoch": 2.3911313177602778, "grad_norm": 2.114485263824463, "learning_rate": 0.00022231117822658577, "loss": 5.1575, "step": 396550 }, { "epoch": 2.391432809541612, "grad_norm": 1.9386225938796997, "learning_rate": 0.00022229334921782442, "loss": 5.0131, "step": 396600 }, { "epoch": 2.3917343013229457, "grad_norm": 1.7194496393203735, "learning_rate": 0.0002222755188786071, "loss": 4.8502, "step": 396650 }, { "epoch": 2.39203579310428, "grad_norm": 1.6070438623428345, "learning_rate": 0.000222257687209262, "loss": 5.0401, "step": 396700 }, { "epoch": 2.392337284885614, "grad_norm": 0.989215075969696, "learning_rate": 0.00022223985421011732, "loss": 5.1512, "step": 396750 }, { "epoch": 2.392638776666948, "grad_norm": 1.5448211431503296, "learning_rate": 0.00022222201988150118, "loss": 4.9019, "step": 396800 }, { "epoch": 2.392940268448282, "grad_norm": 1.9892479181289673, "learning_rate": 0.00022220418422374185, "loss": 5.2469, "step": 396850 }, { "epoch": 2.3932417602296163, "grad_norm": 1.7350844144821167, "learning_rate": 0.00022218634723716754, "loss": 4.974, "step": 396900 }, { "epoch": 2.39354325201095, "grad_norm": 1.5045170783996582, "learning_rate": 0.00022216850892210654, "loss": 5.1061, "step": 396950 }, { "epoch": 2.3938447437922843, "grad_norm": 1.470298171043396, "learning_rate": 0.00022215066927888715, "loss": 5.2062, "step": 397000 }, { "epoch": 2.3941462355736185, "grad_norm": 1.88271164894104, "learning_rate": 0.00022213282830783762, "loss": 4.8916, "step": 397050 }, { "epoch": 2.3944477273549523, "grad_norm": 1.5898152589797974, "learning_rate": 0.00022211498600928627, "loss": 5.2691, "step": 397100 }, { "epoch": 2.3947492191362865, "grad_norm": 1.145816445350647, "learning_rate": 0.00022209714238356163, "loss": 4.9104, "step": 397150 }, { "epoch": 2.3950507109176202, "grad_norm": 0.46114635467529297, "learning_rate": 0.0002220792974309919, "loss": 4.511, "step": 397200 }, { "epoch": 2.3953522026989544, "grad_norm": 2.4945590496063232, "learning_rate": 0.00022206145115190553, "loss": 5.0086, "step": 397250 }, { "epoch": 2.3956536944802886, "grad_norm": 1.7594949007034302, "learning_rate": 0.00022204360354663108, "loss": 5.1948, "step": 397300 }, { "epoch": 2.3959551862616224, "grad_norm": 1.4653619527816772, "learning_rate": 0.00022202575461549687, "loss": 4.9767, "step": 397350 }, { "epoch": 2.3962566780429566, "grad_norm": 2.1965532302856445, "learning_rate": 0.00022200790435883145, "loss": 5.4103, "step": 397400 }, { "epoch": 2.3965581698242904, "grad_norm": 2.9194180965423584, "learning_rate": 0.0002219900527769633, "loss": 4.6857, "step": 397450 }, { "epoch": 2.3968596616056246, "grad_norm": 1.5020846128463745, "learning_rate": 0.00022197219987022093, "loss": 4.9066, "step": 397500 }, { "epoch": 2.397161153386959, "grad_norm": 1.851584792137146, "learning_rate": 0.000221954345638933, "loss": 5.2988, "step": 397550 }, { "epoch": 2.3974626451682925, "grad_norm": 1.0970051288604736, "learning_rate": 0.000221936490083428, "loss": 5.0517, "step": 397600 }, { "epoch": 2.3977641369496268, "grad_norm": 1.3560484647750854, "learning_rate": 0.00022191863320403452, "loss": 4.9136, "step": 397650 }, { "epoch": 2.398065628730961, "grad_norm": 0.5405916571617126, "learning_rate": 0.00022190077500108124, "loss": 4.8865, "step": 397700 }, { "epoch": 2.3983671205122947, "grad_norm": 1.278210997581482, "learning_rate": 0.00022188291547489683, "loss": 4.9144, "step": 397750 }, { "epoch": 2.398668612293629, "grad_norm": 1.5604469776153564, "learning_rate": 0.00022186505462580994, "loss": 4.8892, "step": 397800 }, { "epoch": 2.398970104074963, "grad_norm": 2.383021116256714, "learning_rate": 0.00022184719245414925, "loss": 5.4719, "step": 397850 }, { "epoch": 2.399271595856297, "grad_norm": 1.200305700302124, "learning_rate": 0.00022182932896024356, "loss": 4.6415, "step": 397900 }, { "epoch": 2.399573087637631, "grad_norm": 1.6284743547439575, "learning_rate": 0.00022181146414442153, "loss": 5.2839, "step": 397950 }, { "epoch": 2.399874579418965, "grad_norm": 2.1461048126220703, "learning_rate": 0.00022179359800701204, "loss": 5.1866, "step": 398000 }, { "epoch": 2.400176071200299, "grad_norm": 2.096226692199707, "learning_rate": 0.00022177573054834378, "loss": 4.9253, "step": 398050 }, { "epoch": 2.4004775629816333, "grad_norm": 1.2383326292037964, "learning_rate": 0.0002217578617687456, "loss": 5.1483, "step": 398100 }, { "epoch": 2.400779054762967, "grad_norm": 1.6497483253479004, "learning_rate": 0.00022173999166854647, "loss": 4.8095, "step": 398150 }, { "epoch": 2.4010805465443013, "grad_norm": 1.896498441696167, "learning_rate": 0.00022172212024807513, "loss": 5.2045, "step": 398200 }, { "epoch": 2.401382038325635, "grad_norm": 2.076822280883789, "learning_rate": 0.00022170424750766048, "loss": 4.8765, "step": 398250 }, { "epoch": 2.4016835301069692, "grad_norm": 3.092991590499878, "learning_rate": 0.00022168637344763153, "loss": 4.8947, "step": 398300 }, { "epoch": 2.4019850218883034, "grad_norm": 1.7845375537872314, "learning_rate": 0.00022166849806831715, "loss": 4.8144, "step": 398350 }, { "epoch": 2.402286513669637, "grad_norm": 1.6422598361968994, "learning_rate": 0.00022165062137004637, "loss": 4.9341, "step": 398400 }, { "epoch": 2.4025880054509714, "grad_norm": 1.3774687051773071, "learning_rate": 0.00022163274335314814, "loss": 4.8542, "step": 398450 }, { "epoch": 2.4028894972323056, "grad_norm": 1.7528467178344727, "learning_rate": 0.0002216148640179515, "loss": 4.9439, "step": 398500 }, { "epoch": 2.4031909890136394, "grad_norm": 1.4066665172576904, "learning_rate": 0.0002215969833647855, "loss": 4.9841, "step": 398550 }, { "epoch": 2.4034924807949736, "grad_norm": 1.4723806381225586, "learning_rate": 0.00022157910139397916, "loss": 5.2007, "step": 398600 }, { "epoch": 2.403793972576308, "grad_norm": 1.9694020748138428, "learning_rate": 0.00022156121810586163, "loss": 4.8021, "step": 398650 }, { "epoch": 2.4040954643576415, "grad_norm": 1.8380489349365234, "learning_rate": 0.000221543333500762, "loss": 4.7797, "step": 398700 }, { "epoch": 2.4043969561389757, "grad_norm": 2.6034693717956543, "learning_rate": 0.00022152544757900942, "loss": 4.9504, "step": 398750 }, { "epoch": 2.4046984479203095, "grad_norm": 1.6523873805999756, "learning_rate": 0.00022150756034093307, "loss": 4.7467, "step": 398800 }, { "epoch": 2.4049999397016437, "grad_norm": 1.3421990871429443, "learning_rate": 0.00022148967178686204, "loss": 5.4263, "step": 398850 }, { "epoch": 2.405301431482978, "grad_norm": 2.5726747512817383, "learning_rate": 0.0002214717819171257, "loss": 4.6184, "step": 398900 }, { "epoch": 2.4056029232643117, "grad_norm": 1.8656848669052124, "learning_rate": 0.00022145389073205317, "loss": 4.8267, "step": 398950 }, { "epoch": 2.405904415045646, "grad_norm": 1.3954519033432007, "learning_rate": 0.00022143599823197375, "loss": 5.2972, "step": 399000 }, { "epoch": 2.405904415045646, "eval_loss": 5.449582099914551, "eval_runtime": 39.1903, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 7.349736426080825, "step": 399000 }, { "epoch": 2.4062059068269797, "grad_norm": 2.312695264816284, "learning_rate": 0.00022141810441721676, "loss": 5.1901, "step": 399050 }, { "epoch": 2.406507398608314, "grad_norm": 1.7749125957489014, "learning_rate": 0.0002214002092881114, "loss": 4.9821, "step": 399100 }, { "epoch": 2.406808890389648, "grad_norm": 0.5201131105422974, "learning_rate": 0.0002213823128449872, "loss": 4.9124, "step": 399150 }, { "epoch": 2.407110382170982, "grad_norm": 1.591413140296936, "learning_rate": 0.00022136441508817335, "loss": 5.0586, "step": 399200 }, { "epoch": 2.407411873952316, "grad_norm": 1.970687747001648, "learning_rate": 0.00022134651601799925, "loss": 4.9471, "step": 399250 }, { "epoch": 2.4077133657336502, "grad_norm": 1.4187288284301758, "learning_rate": 0.00022132861563479438, "loss": 5.0539, "step": 399300 }, { "epoch": 2.408014857514984, "grad_norm": 1.601684808731079, "learning_rate": 0.00022131071393888813, "loss": 4.7246, "step": 399350 }, { "epoch": 2.408316349296318, "grad_norm": 1.2964645624160767, "learning_rate": 0.0002212928109306099, "loss": 4.7262, "step": 399400 }, { "epoch": 2.4086178410776524, "grad_norm": 1.8788516521453857, "learning_rate": 0.00022127490661028927, "loss": 4.9805, "step": 399450 }, { "epoch": 2.408919332858986, "grad_norm": 1.2260818481445312, "learning_rate": 0.0002212570009782557, "loss": 4.8678, "step": 399500 }, { "epoch": 2.4092208246403204, "grad_norm": 2.4352571964263916, "learning_rate": 0.0002212390940348387, "loss": 5.0812, "step": 399550 }, { "epoch": 2.409522316421654, "grad_norm": 1.4737792015075684, "learning_rate": 0.00022122118578036786, "loss": 5.1067, "step": 399600 }, { "epoch": 2.4098238082029884, "grad_norm": 2.665726661682129, "learning_rate": 0.00022120327621517273, "loss": 4.6039, "step": 399650 }, { "epoch": 2.4101252999843226, "grad_norm": 1.3896311521530151, "learning_rate": 0.00022118536533958288, "loss": 5.1199, "step": 399700 }, { "epoch": 2.4104267917656563, "grad_norm": 1.914383053779602, "learning_rate": 0.00022116745315392804, "loss": 5.1172, "step": 399750 }, { "epoch": 2.4107282835469905, "grad_norm": 2.385684013366699, "learning_rate": 0.0002211495396585377, "loss": 5.0371, "step": 399800 }, { "epoch": 2.4110297753283247, "grad_norm": 1.4178334474563599, "learning_rate": 0.00022113162485374167, "loss": 4.7295, "step": 399850 }, { "epoch": 2.4113312671096585, "grad_norm": 2.2846240997314453, "learning_rate": 0.00022111370873986955, "loss": 5.0237, "step": 399900 }, { "epoch": 2.4116327588909927, "grad_norm": 2.4566762447357178, "learning_rate": 0.0002210957913172512, "loss": 5.4963, "step": 399950 }, { "epoch": 2.411934250672327, "grad_norm": 1.7784819602966309, "learning_rate": 0.00022107787258621617, "loss": 5.0341, "step": 400000 }, { "epoch": 2.4122357424536607, "grad_norm": 1.3611475229263306, "learning_rate": 0.00022105995254709438, "loss": 4.9573, "step": 400050 }, { "epoch": 2.412537234234995, "grad_norm": 1.9807535409927368, "learning_rate": 0.00022104203120021555, "loss": 5.0665, "step": 400100 }, { "epoch": 2.4128387260163286, "grad_norm": 1.4062421321868896, "learning_rate": 0.00022102410854590954, "loss": 5.3081, "step": 400150 }, { "epoch": 2.413140217797663, "grad_norm": 1.4229108095169067, "learning_rate": 0.00022100618458450615, "loss": 4.9912, "step": 400200 }, { "epoch": 2.413441709578997, "grad_norm": 1.3832820653915405, "learning_rate": 0.00022098825931633528, "loss": 5.315, "step": 400250 }, { "epoch": 2.413743201360331, "grad_norm": 0.9731420874595642, "learning_rate": 0.00022097033274172677, "loss": 4.9076, "step": 400300 }, { "epoch": 2.414044693141665, "grad_norm": 0.5909666419029236, "learning_rate": 0.0002209524048610106, "loss": 4.6005, "step": 400350 }, { "epoch": 2.414346184922999, "grad_norm": 1.6207263469696045, "learning_rate": 0.00022093447567451667, "loss": 4.7441, "step": 400400 }, { "epoch": 2.414647676704333, "grad_norm": 1.6950125694274902, "learning_rate": 0.00022091654518257488, "loss": 4.9448, "step": 400450 }, { "epoch": 2.414949168485667, "grad_norm": 1.7157882452011108, "learning_rate": 0.00022089861338551537, "loss": 5.0418, "step": 400500 }, { "epoch": 2.415250660267001, "grad_norm": 1.4788960218429565, "learning_rate": 0.00022088068028366803, "loss": 4.9405, "step": 400550 }, { "epoch": 2.415552152048335, "grad_norm": 1.527696132659912, "learning_rate": 0.0002208627458773629, "loss": 5.0791, "step": 400600 }, { "epoch": 2.4158536438296694, "grad_norm": 1.3645936250686646, "learning_rate": 0.00022084481016693009, "loss": 4.5567, "step": 400650 }, { "epoch": 2.416155135611003, "grad_norm": 2.1905200481414795, "learning_rate": 0.0002208268731526996, "loss": 5.4077, "step": 400700 }, { "epoch": 2.4164566273923374, "grad_norm": 1.2187188863754272, "learning_rate": 0.00022080893483500164, "loss": 5.0295, "step": 400750 }, { "epoch": 2.4167581191736716, "grad_norm": 2.145832061767578, "learning_rate": 0.00022079099521416626, "loss": 5.2997, "step": 400800 }, { "epoch": 2.4170596109550053, "grad_norm": 0.7192820310592651, "learning_rate": 0.0002207730542905236, "loss": 4.7138, "step": 400850 }, { "epoch": 2.4173611027363395, "grad_norm": 1.4587204456329346, "learning_rate": 0.00022075511206440395, "loss": 5.2126, "step": 400900 }, { "epoch": 2.4176625945176733, "grad_norm": 0.5613594651222229, "learning_rate": 0.00022073716853613742, "loss": 5.0897, "step": 400950 }, { "epoch": 2.4179640862990075, "grad_norm": 2.023611068725586, "learning_rate": 0.0002207192237060542, "loss": 4.5351, "step": 401000 }, { "epoch": 2.4182655780803417, "grad_norm": 1.4378854036331177, "learning_rate": 0.00022070127757448462, "loss": 4.9806, "step": 401050 }, { "epoch": 2.4185670698616755, "grad_norm": 2.271674156188965, "learning_rate": 0.00022068333014175893, "loss": 5.313, "step": 401100 }, { "epoch": 2.4188685616430097, "grad_norm": 3.6257245540618896, "learning_rate": 0.00022066538140820741, "loss": 5.1755, "step": 401150 }, { "epoch": 2.4191700534243434, "grad_norm": 1.6852113008499146, "learning_rate": 0.0002206474313741604, "loss": 5.2149, "step": 401200 }, { "epoch": 2.4194715452056776, "grad_norm": 2.209195137023926, "learning_rate": 0.00022062948003994824, "loss": 4.9466, "step": 401250 }, { "epoch": 2.419773036987012, "grad_norm": 0.5255229473114014, "learning_rate": 0.0002206115274059012, "loss": 5.1137, "step": 401300 }, { "epoch": 2.4200745287683456, "grad_norm": 1.593942403793335, "learning_rate": 0.00022059357347234992, "loss": 4.6568, "step": 401350 }, { "epoch": 2.42037602054968, "grad_norm": 1.7688429355621338, "learning_rate": 0.0002205756182396246, "loss": 5.0016, "step": 401400 }, { "epoch": 2.420677512331014, "grad_norm": 0.939060628414154, "learning_rate": 0.0002205576617080557, "loss": 4.7238, "step": 401450 }, { "epoch": 2.420979004112348, "grad_norm": 1.8382493257522583, "learning_rate": 0.00022053970387797377, "loss": 5.1844, "step": 401500 }, { "epoch": 2.421280495893682, "grad_norm": 2.648010492324829, "learning_rate": 0.00022052174474970925, "loss": 4.932, "step": 401550 }, { "epoch": 2.421581987675016, "grad_norm": 0.8251957893371582, "learning_rate": 0.00022050378432359263, "loss": 4.9107, "step": 401600 }, { "epoch": 2.42188347945635, "grad_norm": 2.5583608150482178, "learning_rate": 0.0002204858225999545, "loss": 4.9528, "step": 401650 }, { "epoch": 2.422184971237684, "grad_norm": 1.1300402879714966, "learning_rate": 0.0002204678595791254, "loss": 4.4742, "step": 401700 }, { "epoch": 2.422486463019018, "grad_norm": 1.423042893409729, "learning_rate": 0.0002204498952614359, "loss": 5.3273, "step": 401750 }, { "epoch": 2.422787954800352, "grad_norm": 1.742882251739502, "learning_rate": 0.0002204319296472166, "loss": 5.281, "step": 401800 }, { "epoch": 2.4230894465816863, "grad_norm": 2.372239351272583, "learning_rate": 0.0002204139627367982, "loss": 5.1521, "step": 401850 }, { "epoch": 2.42339093836302, "grad_norm": 2.1881985664367676, "learning_rate": 0.0002203959945305112, "loss": 5.2515, "step": 401900 }, { "epoch": 2.4236924301443543, "grad_norm": 2.7066028118133545, "learning_rate": 0.00022037802502868648, "loss": 4.9036, "step": 401950 }, { "epoch": 2.423993921925688, "grad_norm": 1.961729645729065, "learning_rate": 0.0002203600542316546, "loss": 5.0855, "step": 402000 }, { "epoch": 2.423993921925688, "eval_loss": 5.44149112701416, "eval_runtime": 39.1544, "eval_samples_per_second": 13.076, "eval_steps_per_second": 6.538, "eval_tts_loss": 7.417742037624404, "step": 402000 }, { "epoch": 2.4242954137070223, "grad_norm": 1.9279327392578125, "learning_rate": 0.0002203420821397463, "loss": 4.9573, "step": 402050 }, { "epoch": 2.4245969054883565, "grad_norm": 1.138176441192627, "learning_rate": 0.00022032410875329244, "loss": 5.065, "step": 402100 }, { "epoch": 2.4248983972696903, "grad_norm": 0.3899115025997162, "learning_rate": 0.00022030613407262366, "loss": 4.9312, "step": 402150 }, { "epoch": 2.4251998890510245, "grad_norm": 0.9150970578193665, "learning_rate": 0.0002202881580980708, "loss": 4.776, "step": 402200 }, { "epoch": 2.4255013808323587, "grad_norm": 0.8825558423995972, "learning_rate": 0.00022027018082996478, "loss": 4.8702, "step": 402250 }, { "epoch": 2.4258028726136924, "grad_norm": 3.960470199584961, "learning_rate": 0.0002202522022686363, "loss": 4.8699, "step": 402300 }, { "epoch": 2.4261043643950266, "grad_norm": 2.544496774673462, "learning_rate": 0.0002202342224144163, "loss": 5.1376, "step": 402350 }, { "epoch": 2.426405856176361, "grad_norm": 1.9227712154388428, "learning_rate": 0.0002202162412676357, "loss": 4.7055, "step": 402400 }, { "epoch": 2.4267073479576946, "grad_norm": 2.6109113693237305, "learning_rate": 0.0002201982588286253, "loss": 5.5176, "step": 402450 }, { "epoch": 2.427008839739029, "grad_norm": 1.5037823915481567, "learning_rate": 0.0002201802750977162, "loss": 4.9376, "step": 402500 }, { "epoch": 2.4273103315203626, "grad_norm": 1.4125511646270752, "learning_rate": 0.00022016229007523926, "loss": 5.0657, "step": 402550 }, { "epoch": 2.427611823301697, "grad_norm": 2.139039993286133, "learning_rate": 0.00022014430376152548, "loss": 4.9858, "step": 402600 }, { "epoch": 2.427913315083031, "grad_norm": 0.9405465126037598, "learning_rate": 0.00022012631615690593, "loss": 4.9199, "step": 402650 }, { "epoch": 2.4282148068643647, "grad_norm": 2.0391175746917725, "learning_rate": 0.00022010832726171157, "loss": 4.8057, "step": 402700 }, { "epoch": 2.428516298645699, "grad_norm": 1.7398383617401123, "learning_rate": 0.00022009033707627354, "loss": 4.7098, "step": 402750 }, { "epoch": 2.4288177904270327, "grad_norm": 3.1351430416107178, "learning_rate": 0.00022007234560092286, "loss": 4.8269, "step": 402800 }, { "epoch": 2.429119282208367, "grad_norm": 1.3181979656219482, "learning_rate": 0.00022005435283599064, "loss": 5.0674, "step": 402850 }, { "epoch": 2.429420773989701, "grad_norm": 1.8485560417175293, "learning_rate": 0.00022003635878180803, "loss": 4.4904, "step": 402900 }, { "epoch": 2.429722265771035, "grad_norm": 2.030529260635376, "learning_rate": 0.00022001836343870623, "loss": 5.021, "step": 402950 }, { "epoch": 2.430023757552369, "grad_norm": 2.299678087234497, "learning_rate": 0.00022000036680701636, "loss": 5.0936, "step": 403000 }, { "epoch": 2.4303252493337033, "grad_norm": 2.7160394191741943, "learning_rate": 0.00021998236888706957, "loss": 4.9837, "step": 403050 }, { "epoch": 2.430626741115037, "grad_norm": 1.5512908697128296, "learning_rate": 0.0002199643696791972, "loss": 5.2436, "step": 403100 }, { "epoch": 2.4309282328963713, "grad_norm": 1.5691943168640137, "learning_rate": 0.00021994636918373049, "loss": 4.9326, "step": 403150 }, { "epoch": 2.4312297246777055, "grad_norm": 1.5093270540237427, "learning_rate": 0.0002199283674010006, "loss": 5.2109, "step": 403200 }, { "epoch": 2.4315312164590392, "grad_norm": 2.0650830268859863, "learning_rate": 0.00021991036433133894, "loss": 4.7897, "step": 403250 }, { "epoch": 2.4318327082403735, "grad_norm": 2.0631651878356934, "learning_rate": 0.00021989235997507678, "loss": 5.0074, "step": 403300 }, { "epoch": 2.432134200021707, "grad_norm": 1.578500509262085, "learning_rate": 0.00021987435433254549, "loss": 4.6373, "step": 403350 }, { "epoch": 2.4324356918030414, "grad_norm": 0.7688920497894287, "learning_rate": 0.00021985634740407642, "loss": 5.2084, "step": 403400 }, { "epoch": 2.4327371835843756, "grad_norm": 1.4483532905578613, "learning_rate": 0.00021983833919000096, "loss": 5.0007, "step": 403450 }, { "epoch": 2.4330386753657094, "grad_norm": 2.0289483070373535, "learning_rate": 0.0002198203296906505, "loss": 5.0141, "step": 403500 }, { "epoch": 2.4333401671470436, "grad_norm": 2.0656192302703857, "learning_rate": 0.0002198023189063566, "loss": 5.3088, "step": 403550 }, { "epoch": 2.433641658928378, "grad_norm": 0.7134799957275391, "learning_rate": 0.00021978430683745057, "loss": 4.6257, "step": 403600 }, { "epoch": 2.4339431507097116, "grad_norm": 1.540412187576294, "learning_rate": 0.00021976629348426392, "loss": 5.1269, "step": 403650 }, { "epoch": 2.4342446424910458, "grad_norm": 1.1767220497131348, "learning_rate": 0.00021974827884712827, "loss": 5.2372, "step": 403700 }, { "epoch": 2.43454613427238, "grad_norm": 1.3708769083023071, "learning_rate": 0.0002197302629263751, "loss": 5.0936, "step": 403750 }, { "epoch": 2.4348476260537137, "grad_norm": 1.5464930534362793, "learning_rate": 0.00021971224572233584, "loss": 5.1064, "step": 403800 }, { "epoch": 2.435149117835048, "grad_norm": 1.92662513256073, "learning_rate": 0.0002196942272353423, "loss": 5.1653, "step": 403850 }, { "epoch": 2.4354506096163817, "grad_norm": 1.6912274360656738, "learning_rate": 0.00021967620746572592, "loss": 4.9324, "step": 403900 }, { "epoch": 2.435752101397716, "grad_norm": 1.3946729898452759, "learning_rate": 0.00021965818641381837, "loss": 4.713, "step": 403950 }, { "epoch": 2.43605359317905, "grad_norm": 1.6522958278656006, "learning_rate": 0.00021964016407995132, "loss": 4.5789, "step": 404000 }, { "epoch": 2.436355084960384, "grad_norm": 0.48584434390068054, "learning_rate": 0.00021962214046445638, "loss": 4.5002, "step": 404050 }, { "epoch": 2.436656576741718, "grad_norm": 1.939421534538269, "learning_rate": 0.00021960411556766536, "loss": 4.8675, "step": 404100 }, { "epoch": 2.436958068523052, "grad_norm": 1.6311386823654175, "learning_rate": 0.00021958608938990997, "loss": 5.3524, "step": 404150 }, { "epoch": 2.437259560304386, "grad_norm": 1.3967509269714355, "learning_rate": 0.00021956806193152182, "loss": 4.4591, "step": 404200 }, { "epoch": 2.4375610520857203, "grad_norm": 2.0919339656829834, "learning_rate": 0.00021955003319283278, "loss": 4.9163, "step": 404250 }, { "epoch": 2.437862543867054, "grad_norm": 1.9408190250396729, "learning_rate": 0.00021953200317417467, "loss": 4.9299, "step": 404300 }, { "epoch": 2.4381640356483882, "grad_norm": 2.1485612392425537, "learning_rate": 0.00021951397187587924, "loss": 4.9347, "step": 404350 }, { "epoch": 2.4384655274297224, "grad_norm": 1.2948178052902222, "learning_rate": 0.00021949593929827837, "loss": 4.9116, "step": 404400 }, { "epoch": 2.438767019211056, "grad_norm": 0.6350423693656921, "learning_rate": 0.0002194779054417039, "loss": 4.6437, "step": 404450 }, { "epoch": 2.4390685109923904, "grad_norm": 1.7534141540527344, "learning_rate": 0.00021945987030648774, "loss": 4.7138, "step": 404500 }, { "epoch": 2.4393700027737246, "grad_norm": 2.0769364833831787, "learning_rate": 0.0002194418338929618, "loss": 5.0875, "step": 404550 }, { "epoch": 2.4396714945550584, "grad_norm": 1.7097752094268799, "learning_rate": 0.00021942379620145804, "loss": 4.9743, "step": 404600 }, { "epoch": 2.4399729863363926, "grad_norm": 2.6318717002868652, "learning_rate": 0.00021940575723230831, "loss": 5.2016, "step": 404650 }, { "epoch": 2.4402744781177264, "grad_norm": 2.0439600944519043, "learning_rate": 0.0002193877169858447, "loss": 5.2371, "step": 404700 }, { "epoch": 2.4405759698990606, "grad_norm": 1.8277145624160767, "learning_rate": 0.0002193696754623992, "loss": 4.6342, "step": 404750 }, { "epoch": 2.4408774616803948, "grad_norm": 2.112311601638794, "learning_rate": 0.0002193516326623038, "loss": 5.2071, "step": 404800 }, { "epoch": 2.4411789534617285, "grad_norm": 1.7644425630569458, "learning_rate": 0.00021933358858589058, "loss": 4.8396, "step": 404850 }, { "epoch": 2.4414804452430627, "grad_norm": 1.581359624862671, "learning_rate": 0.0002193155432334916, "loss": 5.0374, "step": 404900 }, { "epoch": 2.4417819370243965, "grad_norm": 2.3717479705810547, "learning_rate": 0.00021929749660543895, "loss": 5.241, "step": 404950 }, { "epoch": 2.4420834288057307, "grad_norm": 1.7734328508377075, "learning_rate": 0.00021927944870206477, "loss": 5.0546, "step": 405000 }, { "epoch": 2.4420834288057307, "eval_loss": 5.446295738220215, "eval_runtime": 38.9515, "eval_samples_per_second": 13.145, "eval_steps_per_second": 6.572, "eval_tts_loss": 7.3464091240322285, "step": 405000 }, { "epoch": 2.442384920587065, "grad_norm": 1.5572805404663086, "learning_rate": 0.00021926139952370123, "loss": 4.7371, "step": 405050 }, { "epoch": 2.4426864123683987, "grad_norm": 1.0207878351211548, "learning_rate": 0.00021924334907068042, "loss": 4.761, "step": 405100 }, { "epoch": 2.442987904149733, "grad_norm": 0.8421022891998291, "learning_rate": 0.00021922529734333466, "loss": 4.3644, "step": 405150 }, { "epoch": 2.443289395931067, "grad_norm": 1.4044723510742188, "learning_rate": 0.00021920724434199606, "loss": 5.2466, "step": 405200 }, { "epoch": 2.443590887712401, "grad_norm": 1.4181965589523315, "learning_rate": 0.0002191891900669969, "loss": 5.3297, "step": 405250 }, { "epoch": 2.443892379493735, "grad_norm": 1.1825388669967651, "learning_rate": 0.0002191711345186694, "loss": 4.9844, "step": 405300 }, { "epoch": 2.4441938712750693, "grad_norm": 1.1268835067749023, "learning_rate": 0.00021915307769734593, "loss": 4.9346, "step": 405350 }, { "epoch": 2.444495363056403, "grad_norm": 2.075920581817627, "learning_rate": 0.0002191350196033587, "loss": 4.7783, "step": 405400 }, { "epoch": 2.4447968548377372, "grad_norm": 2.3067610263824463, "learning_rate": 0.00021911696023704016, "loss": 4.8176, "step": 405450 }, { "epoch": 2.445098346619071, "grad_norm": 1.7929580211639404, "learning_rate": 0.00021909889959872254, "loss": 4.8775, "step": 405500 }, { "epoch": 2.445399838400405, "grad_norm": 1.2867883443832397, "learning_rate": 0.0002190808376887383, "loss": 4.8461, "step": 405550 }, { "epoch": 2.4457013301817394, "grad_norm": 2.056096315383911, "learning_rate": 0.00021906277450741985, "loss": 4.6902, "step": 405600 }, { "epoch": 2.446002821963073, "grad_norm": 1.655070185661316, "learning_rate": 0.00021904471005509957, "loss": 4.8393, "step": 405650 }, { "epoch": 2.4463043137444074, "grad_norm": 2.1471095085144043, "learning_rate": 0.00021902664433210992, "loss": 5.0915, "step": 405700 }, { "epoch": 2.446605805525741, "grad_norm": 1.9359861612319946, "learning_rate": 0.00021900857733878342, "loss": 5.1685, "step": 405750 }, { "epoch": 2.4469072973070753, "grad_norm": 2.723355531692505, "learning_rate": 0.0002189905090754525, "loss": 5.1851, "step": 405800 }, { "epoch": 2.4472087890884096, "grad_norm": 2.526996612548828, "learning_rate": 0.00021897243954244972, "loss": 5.1982, "step": 405850 }, { "epoch": 2.4475102808697433, "grad_norm": 1.4595788717269897, "learning_rate": 0.00021895436874010767, "loss": 4.801, "step": 405900 }, { "epoch": 2.4478117726510775, "grad_norm": 2.150219678878784, "learning_rate": 0.00021893629666875882, "loss": 5.1278, "step": 405950 }, { "epoch": 2.4481132644324117, "grad_norm": 2.130345344543457, "learning_rate": 0.00021891822332873581, "loss": 4.9492, "step": 406000 }, { "epoch": 2.4484147562137455, "grad_norm": 1.6311607360839844, "learning_rate": 0.00021890014872037126, "loss": 5.0275, "step": 406050 }, { "epoch": 2.4487162479950797, "grad_norm": 1.376933217048645, "learning_rate": 0.0002188820728439978, "loss": 5.2314, "step": 406100 }, { "epoch": 2.449017739776414, "grad_norm": 2.4815266132354736, "learning_rate": 0.0002188639956999481, "loss": 5.0764, "step": 406150 }, { "epoch": 2.4493192315577477, "grad_norm": 2.4148552417755127, "learning_rate": 0.00021884591728855483, "loss": 5.2148, "step": 406200 }, { "epoch": 2.449620723339082, "grad_norm": 1.9125621318817139, "learning_rate": 0.00021882783761015067, "loss": 5.2711, "step": 406250 }, { "epoch": 2.4499222151204156, "grad_norm": 0.4502065181732178, "learning_rate": 0.0002188097566650684, "loss": 4.5883, "step": 406300 }, { "epoch": 2.45022370690175, "grad_norm": 1.6806986331939697, "learning_rate": 0.0002187916744536408, "loss": 5.0318, "step": 406350 }, { "epoch": 2.450525198683084, "grad_norm": 1.8095611333847046, "learning_rate": 0.00021877359097620053, "loss": 4.7941, "step": 406400 }, { "epoch": 2.450826690464418, "grad_norm": 1.323620080947876, "learning_rate": 0.0002187555062330805, "loss": 5.2255, "step": 406450 }, { "epoch": 2.451128182245752, "grad_norm": 2.0589301586151123, "learning_rate": 0.00021873742022461348, "loss": 5.1734, "step": 406500 }, { "epoch": 2.451429674027086, "grad_norm": 0.9307945370674133, "learning_rate": 0.00021871933295113233, "loss": 4.8865, "step": 406550 }, { "epoch": 2.45173116580842, "grad_norm": 1.4549609422683716, "learning_rate": 0.00021870124441296994, "loss": 5.1156, "step": 406600 }, { "epoch": 2.452032657589754, "grad_norm": 1.3704859018325806, "learning_rate": 0.0002186831546104592, "loss": 5.0486, "step": 406650 }, { "epoch": 2.452334149371088, "grad_norm": 2.129194736480713, "learning_rate": 0.000218665063543933, "loss": 4.8777, "step": 406700 }, { "epoch": 2.452635641152422, "grad_norm": 1.639875054359436, "learning_rate": 0.00021864697121372432, "loss": 5.0689, "step": 406750 }, { "epoch": 2.4529371329337564, "grad_norm": 1.2832746505737305, "learning_rate": 0.00021862887762016608, "loss": 5.1573, "step": 406800 }, { "epoch": 2.45323862471509, "grad_norm": 1.5755749940872192, "learning_rate": 0.00021861078276359127, "loss": 5.3401, "step": 406850 }, { "epoch": 2.4535401164964243, "grad_norm": 1.1074858903884888, "learning_rate": 0.00021859268664433293, "loss": 5.0081, "step": 406900 }, { "epoch": 2.4538416082777585, "grad_norm": 1.8692177534103394, "learning_rate": 0.00021857458926272414, "loss": 5.09, "step": 406950 }, { "epoch": 2.4541431000590923, "grad_norm": 1.7267659902572632, "learning_rate": 0.00021855649061909777, "loss": 5.1005, "step": 407000 }, { "epoch": 2.4544445918404265, "grad_norm": 1.546557068824768, "learning_rate": 0.00021853839071378712, "loss": 5.2226, "step": 407050 }, { "epoch": 2.4547460836217603, "grad_norm": 1.678471565246582, "learning_rate": 0.00021852028954712517, "loss": 5.1401, "step": 407100 }, { "epoch": 2.4550475754030945, "grad_norm": 0.7004038095474243, "learning_rate": 0.00021850218711944508, "loss": 4.724, "step": 407150 }, { "epoch": 2.4553490671844287, "grad_norm": 1.6252903938293457, "learning_rate": 0.00021848408343107996, "loss": 4.847, "step": 407200 }, { "epoch": 2.4556505589657625, "grad_norm": 1.833946704864502, "learning_rate": 0.00021846597848236308, "loss": 5.2134, "step": 407250 }, { "epoch": 2.4559520507470967, "grad_norm": 1.335918664932251, "learning_rate": 0.00021844787227362752, "loss": 5.0161, "step": 407300 }, { "epoch": 2.456253542528431, "grad_norm": 1.831375002861023, "learning_rate": 0.0002184297648052066, "loss": 4.8862, "step": 407350 }, { "epoch": 2.4565550343097646, "grad_norm": 1.9115467071533203, "learning_rate": 0.00021841165607743344, "loss": 5.1872, "step": 407400 }, { "epoch": 2.456856526091099, "grad_norm": 2.1808390617370605, "learning_rate": 0.00021839354609064142, "loss": 5.041, "step": 407450 }, { "epoch": 2.457158017872433, "grad_norm": 1.5380754470825195, "learning_rate": 0.0002183754348451638, "loss": 5.168, "step": 407500 }, { "epoch": 2.457459509653767, "grad_norm": 1.5516141653060913, "learning_rate": 0.00021835732234133385, "loss": 5.1239, "step": 407550 }, { "epoch": 2.457761001435101, "grad_norm": 1.970680832862854, "learning_rate": 0.00021833920857948497, "loss": 4.9115, "step": 407600 }, { "epoch": 2.4580624932164348, "grad_norm": 1.9184753894805908, "learning_rate": 0.00021832109355995045, "loss": 5.0909, "step": 407650 }, { "epoch": 2.458363984997769, "grad_norm": 0.9735426306724548, "learning_rate": 0.00021830297728306374, "loss": 5.1792, "step": 407700 }, { "epoch": 2.458665476779103, "grad_norm": 2.4570939540863037, "learning_rate": 0.0002182848597491582, "loss": 4.8875, "step": 407750 }, { "epoch": 2.458966968560437, "grad_norm": 2.0401716232299805, "learning_rate": 0.00021826674095856724, "loss": 4.6587, "step": 407800 }, { "epoch": 2.459268460341771, "grad_norm": 1.24132239818573, "learning_rate": 0.0002182486209116243, "loss": 5.028, "step": 407850 }, { "epoch": 2.459569952123105, "grad_norm": 3.5255239009857178, "learning_rate": 0.00021823049960866293, "loss": 5.3329, "step": 407900 }, { "epoch": 2.459871443904439, "grad_norm": 2.1363673210144043, "learning_rate": 0.0002182123770500166, "loss": 4.9783, "step": 407950 }, { "epoch": 2.4601729356857733, "grad_norm": 1.5424338579177856, "learning_rate": 0.00021819425323601878, "loss": 5.1003, "step": 408000 }, { "epoch": 2.4601729356857733, "eval_loss": 5.432572364807129, "eval_runtime": 39.1087, "eval_samples_per_second": 13.092, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.41975550564155, "step": 408000 }, { "epoch": 2.460474427467107, "grad_norm": 1.4487032890319824, "learning_rate": 0.00021817612816700305, "loss": 4.9966, "step": 408050 }, { "epoch": 2.4607759192484413, "grad_norm": 1.708829402923584, "learning_rate": 0.000218158001843303, "loss": 4.6275, "step": 408100 }, { "epoch": 2.4610774110297755, "grad_norm": 1.7623180150985718, "learning_rate": 0.00021813987426525219, "loss": 4.9557, "step": 408150 }, { "epoch": 2.4613789028111093, "grad_norm": 1.6345771551132202, "learning_rate": 0.00021812174543318422, "loss": 5.1311, "step": 408200 }, { "epoch": 2.4616803945924435, "grad_norm": 2.5927658081054688, "learning_rate": 0.00021810361534743277, "loss": 4.9381, "step": 408250 }, { "epoch": 2.4619818863737777, "grad_norm": 1.8005027770996094, "learning_rate": 0.00021808548400833143, "loss": 4.7468, "step": 408300 }, { "epoch": 2.4622833781551114, "grad_norm": 1.656917691230774, "learning_rate": 0.00021806735141621396, "loss": 4.6853, "step": 408350 }, { "epoch": 2.4625848699364457, "grad_norm": 1.8566738367080688, "learning_rate": 0.00021804921757141398, "loss": 4.9207, "step": 408400 }, { "epoch": 2.4628863617177794, "grad_norm": 2.0454280376434326, "learning_rate": 0.00021803108247426525, "loss": 4.7448, "step": 408450 }, { "epoch": 2.4631878534991136, "grad_norm": 1.122841238975525, "learning_rate": 0.00021801294612510156, "loss": 4.9225, "step": 408500 }, { "epoch": 2.463489345280448, "grad_norm": 1.5715738534927368, "learning_rate": 0.00021799480852425671, "loss": 4.753, "step": 408550 }, { "epoch": 2.4637908370617816, "grad_norm": 1.625836730003357, "learning_rate": 0.00021797666967206437, "loss": 4.8414, "step": 408600 }, { "epoch": 2.464092328843116, "grad_norm": 1.0393636226654053, "learning_rate": 0.00021795852956885844, "loss": 4.5012, "step": 408650 }, { "epoch": 2.4643938206244496, "grad_norm": 0.9951134324073792, "learning_rate": 0.0002179403882149728, "loss": 4.8747, "step": 408700 }, { "epoch": 2.4646953124057838, "grad_norm": 2.2055606842041016, "learning_rate": 0.00021792224561074125, "loss": 5.1764, "step": 408750 }, { "epoch": 2.464996804187118, "grad_norm": 1.7514716386795044, "learning_rate": 0.0002179041017564977, "loss": 5.1897, "step": 408800 }, { "epoch": 2.4652982959684517, "grad_norm": 1.34211266040802, "learning_rate": 0.00021788595665257602, "loss": 4.7132, "step": 408850 }, { "epoch": 2.465599787749786, "grad_norm": 2.638629674911499, "learning_rate": 0.00021786781029931024, "loss": 4.7227, "step": 408900 }, { "epoch": 2.46590127953112, "grad_norm": 1.6962075233459473, "learning_rate": 0.00021784966269703423, "loss": 5.2387, "step": 408950 }, { "epoch": 2.466202771312454, "grad_norm": 1.8509228229522705, "learning_rate": 0.00021783151384608204, "loss": 5.5372, "step": 409000 }, { "epoch": 2.466504263093788, "grad_norm": 2.1137351989746094, "learning_rate": 0.0002178133637467876, "loss": 4.9994, "step": 409050 }, { "epoch": 2.4668057548751223, "grad_norm": 1.2663171291351318, "learning_rate": 0.00021779521239948497, "loss": 5.237, "step": 409100 }, { "epoch": 2.467107246656456, "grad_norm": 1.4014694690704346, "learning_rate": 0.00021777705980450824, "loss": 5.0547, "step": 409150 }, { "epoch": 2.4674087384377903, "grad_norm": 2.74814772605896, "learning_rate": 0.00021775890596219144, "loss": 5.3351, "step": 409200 }, { "epoch": 2.467710230219124, "grad_norm": 1.4225780963897705, "learning_rate": 0.00021774075087286867, "loss": 5.2059, "step": 409250 }, { "epoch": 2.4680117220004583, "grad_norm": 1.5149468183517456, "learning_rate": 0.00021772259453687405, "loss": 4.6739, "step": 409300 }, { "epoch": 2.4683132137817925, "grad_norm": 1.3310648202896118, "learning_rate": 0.00021770443695454174, "loss": 5.125, "step": 409350 }, { "epoch": 2.4686147055631262, "grad_norm": 1.604642391204834, "learning_rate": 0.00021768627812620585, "loss": 4.6598, "step": 409400 }, { "epoch": 2.4689161973444604, "grad_norm": 1.8411434888839722, "learning_rate": 0.0002176681180522006, "loss": 5.2918, "step": 409450 }, { "epoch": 2.469217689125794, "grad_norm": 1.3541598320007324, "learning_rate": 0.00021764995673286022, "loss": 4.7439, "step": 409500 }, { "epoch": 2.4695191809071284, "grad_norm": 1.606815218925476, "learning_rate": 0.00021763179416851898, "loss": 4.9662, "step": 409550 }, { "epoch": 2.4698206726884626, "grad_norm": 1.8169761896133423, "learning_rate": 0.000217613630359511, "loss": 4.9625, "step": 409600 }, { "epoch": 2.4701221644697964, "grad_norm": 2.459615468978882, "learning_rate": 0.0002175954653061707, "loss": 4.9597, "step": 409650 }, { "epoch": 2.4704236562511306, "grad_norm": 2.172954797744751, "learning_rate": 0.0002175772990088323, "loss": 4.8514, "step": 409700 }, { "epoch": 2.470725148032465, "grad_norm": 1.5398132801055908, "learning_rate": 0.00021755913146783015, "loss": 4.7822, "step": 409750 }, { "epoch": 2.4710266398137986, "grad_norm": 1.5777175426483154, "learning_rate": 0.00021754096268349863, "loss": 4.7068, "step": 409800 }, { "epoch": 2.4713281315951328, "grad_norm": 2.3124189376831055, "learning_rate": 0.00021752279265617206, "loss": 4.9283, "step": 409850 }, { "epoch": 2.471629623376467, "grad_norm": 1.5743417739868164, "learning_rate": 0.00021750462138618484, "loss": 5.1422, "step": 409900 }, { "epoch": 2.4719311151578007, "grad_norm": 1.518612027168274, "learning_rate": 0.0002174864488738714, "loss": 5.2084, "step": 409950 }, { "epoch": 2.472232606939135, "grad_norm": 1.5023547410964966, "learning_rate": 0.0002174682751195662, "loss": 4.6285, "step": 410000 }, { "epoch": 2.4725340987204687, "grad_norm": 1.4674961566925049, "learning_rate": 0.0002174501001236036, "loss": 4.9169, "step": 410050 }, { "epoch": 2.472835590501803, "grad_norm": 1.0915076732635498, "learning_rate": 0.00021743192388631825, "loss": 5.2668, "step": 410100 }, { "epoch": 2.473137082283137, "grad_norm": 2.178046226501465, "learning_rate": 0.00021741374640804457, "loss": 5.0364, "step": 410150 }, { "epoch": 2.473438574064471, "grad_norm": 1.3961602449417114, "learning_rate": 0.00021739556768911704, "loss": 5.1596, "step": 410200 }, { "epoch": 2.473740065845805, "grad_norm": 3.5476412773132324, "learning_rate": 0.00021737738772987028, "loss": 4.8063, "step": 410250 }, { "epoch": 2.474041557627139, "grad_norm": 2.7989461421966553, "learning_rate": 0.00021735920653063887, "loss": 5.2531, "step": 410300 }, { "epoch": 2.474343049408473, "grad_norm": 1.4855706691741943, "learning_rate": 0.0002173410240917574, "loss": 5.3359, "step": 410350 }, { "epoch": 2.4746445411898073, "grad_norm": 2.4605722427368164, "learning_rate": 0.00021732284041356046, "loss": 5.0093, "step": 410400 }, { "epoch": 2.474946032971141, "grad_norm": 2.92691969871521, "learning_rate": 0.00021730465549638274, "loss": 4.836, "step": 410450 }, { "epoch": 2.4752475247524752, "grad_norm": 2.6153335571289062, "learning_rate": 0.00021728646934055884, "loss": 5.1704, "step": 410500 }, { "epoch": 2.4755490165338094, "grad_norm": 1.8033236265182495, "learning_rate": 0.00021726828194642352, "loss": 4.9047, "step": 410550 }, { "epoch": 2.475850508315143, "grad_norm": 1.4785431623458862, "learning_rate": 0.0002172500933143115, "loss": 5.1071, "step": 410600 }, { "epoch": 2.4761520000964774, "grad_norm": 2.0552120208740234, "learning_rate": 0.0002172319034445574, "loss": 5.4435, "step": 410650 }, { "epoch": 2.4764534918778116, "grad_norm": 1.8731619119644165, "learning_rate": 0.00021721371233749613, "loss": 5.006, "step": 410700 }, { "epoch": 2.4767549836591454, "grad_norm": 1.5919727087020874, "learning_rate": 0.00021719551999346243, "loss": 5.3638, "step": 410750 }, { "epoch": 2.4770564754404796, "grad_norm": 2.053466320037842, "learning_rate": 0.00021717732641279099, "loss": 4.7507, "step": 410800 }, { "epoch": 2.4773579672218133, "grad_norm": 1.7634600400924683, "learning_rate": 0.0002171591315958168, "loss": 4.8216, "step": 410850 }, { "epoch": 2.4776594590031475, "grad_norm": 1.3821120262145996, "learning_rate": 0.00021714093554287459, "loss": 4.6837, "step": 410900 }, { "epoch": 2.4779609507844818, "grad_norm": 2.1593165397644043, "learning_rate": 0.00021712273825429929, "loss": 4.6335, "step": 410950 }, { "epoch": 2.4782624425658155, "grad_norm": 1.6157103776931763, "learning_rate": 0.00021710453973042577, "loss": 4.7173, "step": 411000 }, { "epoch": 2.4782624425658155, "eval_loss": 5.448861122131348, "eval_runtime": 38.9988, "eval_samples_per_second": 13.129, "eval_steps_per_second": 6.564, "eval_tts_loss": 7.356982376926529, "step": 411000 }, { "epoch": 2.4785639343471497, "grad_norm": 1.826833963394165, "learning_rate": 0.00021708633997158898, "loss": 4.9202, "step": 411050 }, { "epoch": 2.478865426128484, "grad_norm": 2.1254794597625732, "learning_rate": 0.00021706813897812385, "loss": 4.7768, "step": 411100 }, { "epoch": 2.4791669179098177, "grad_norm": 1.3658850193023682, "learning_rate": 0.00021704993675036532, "loss": 5.1719, "step": 411150 }, { "epoch": 2.479468409691152, "grad_norm": 1.4637216329574585, "learning_rate": 0.00021703173328864832, "loss": 4.6341, "step": 411200 }, { "epoch": 2.479769901472486, "grad_norm": 2.037604331970215, "learning_rate": 0.00021701352859330804, "loss": 5.3601, "step": 411250 }, { "epoch": 2.48007139325382, "grad_norm": 1.695335030555725, "learning_rate": 0.00021699532266467934, "loss": 5.1056, "step": 411300 }, { "epoch": 2.480372885035154, "grad_norm": 0.9284546375274658, "learning_rate": 0.00021697711550309735, "loss": 4.6888, "step": 411350 }, { "epoch": 2.480674376816488, "grad_norm": 2.841130495071411, "learning_rate": 0.00021695890710889712, "loss": 5.2005, "step": 411400 }, { "epoch": 2.480975868597822, "grad_norm": 2.7221271991729736, "learning_rate": 0.0002169406974824138, "loss": 4.472, "step": 411450 }, { "epoch": 2.4812773603791562, "grad_norm": 2.0793161392211914, "learning_rate": 0.00021692248662398244, "loss": 5.0288, "step": 411500 }, { "epoch": 2.48157885216049, "grad_norm": 1.1658982038497925, "learning_rate": 0.00021690427453393822, "loss": 4.5525, "step": 411550 }, { "epoch": 2.481880343941824, "grad_norm": 0.5914915800094604, "learning_rate": 0.00021688606121261628, "loss": 5.0908, "step": 411600 }, { "epoch": 2.482181835723158, "grad_norm": 1.0794368982315063, "learning_rate": 0.00021686784666035186, "loss": 4.9871, "step": 411650 }, { "epoch": 2.482483327504492, "grad_norm": 2.0707998275756836, "learning_rate": 0.00021684963087748016, "loss": 4.9532, "step": 411700 }, { "epoch": 2.4827848192858264, "grad_norm": 1.8418834209442139, "learning_rate": 0.0002168314138643364, "loss": 5.3057, "step": 411750 }, { "epoch": 2.48308631106716, "grad_norm": 1.0817320346832275, "learning_rate": 0.0002168131956212558, "loss": 5.0729, "step": 411800 }, { "epoch": 2.4833878028484944, "grad_norm": 2.5046041011810303, "learning_rate": 0.00021679497614857373, "loss": 5.2837, "step": 411850 }, { "epoch": 2.4836892946298286, "grad_norm": 1.3678525686264038, "learning_rate": 0.0002167767554466254, "loss": 4.9137, "step": 411900 }, { "epoch": 2.4839907864111623, "grad_norm": 2.6099255084991455, "learning_rate": 0.00021675853351574623, "loss": 5.0469, "step": 411950 }, { "epoch": 2.4842922781924965, "grad_norm": 1.9263005256652832, "learning_rate": 0.00021674031035627151, "loss": 5.1783, "step": 412000 }, { "epoch": 2.4845937699738307, "grad_norm": 0.9723213315010071, "learning_rate": 0.0002167220859685366, "loss": 5.0847, "step": 412050 }, { "epoch": 2.4848952617551645, "grad_norm": 1.3613039255142212, "learning_rate": 0.00021670386035287692, "loss": 4.7601, "step": 412100 }, { "epoch": 2.4851967535364987, "grad_norm": 1.8465627431869507, "learning_rate": 0.00021668563350962788, "loss": 4.985, "step": 412150 }, { "epoch": 2.4854982453178325, "grad_norm": 1.0589160919189453, "learning_rate": 0.00021666740543912496, "loss": 4.5932, "step": 412200 }, { "epoch": 2.4857997370991667, "grad_norm": 2.407349109649658, "learning_rate": 0.0002166491761417035, "loss": 5.236, "step": 412250 }, { "epoch": 2.486101228880501, "grad_norm": 1.307974100112915, "learning_rate": 0.00021663094561769913, "loss": 4.8738, "step": 412300 }, { "epoch": 2.4864027206618347, "grad_norm": 1.98446524143219, "learning_rate": 0.0002166127138674473, "loss": 5.0933, "step": 412350 }, { "epoch": 2.486704212443169, "grad_norm": 1.7228606939315796, "learning_rate": 0.0002165944808912835, "loss": 5.0343, "step": 412400 }, { "epoch": 2.4870057042245026, "grad_norm": 0.9126201272010803, "learning_rate": 0.00021657624668954334, "loss": 5.006, "step": 412450 }, { "epoch": 2.487307196005837, "grad_norm": 0.935261070728302, "learning_rate": 0.00021655801126256236, "loss": 4.9393, "step": 412500 }, { "epoch": 2.487608687787171, "grad_norm": 1.9437938928604126, "learning_rate": 0.00021653977461067613, "loss": 5.0406, "step": 412550 }, { "epoch": 2.487910179568505, "grad_norm": 1.345807433128357, "learning_rate": 0.00021652153673422036, "loss": 4.9955, "step": 412600 }, { "epoch": 2.488211671349839, "grad_norm": 1.4763636589050293, "learning_rate": 0.00021650329763353058, "loss": 5.1582, "step": 412650 }, { "epoch": 2.488513163131173, "grad_norm": 1.8384608030319214, "learning_rate": 0.00021648505730894262, "loss": 4.7727, "step": 412700 }, { "epoch": 2.488814654912507, "grad_norm": 2.029528856277466, "learning_rate": 0.000216466815760792, "loss": 4.9909, "step": 412750 }, { "epoch": 2.489116146693841, "grad_norm": 0.5779051780700684, "learning_rate": 0.00021644857298941443, "loss": 5.1613, "step": 412800 }, { "epoch": 2.4894176384751754, "grad_norm": 1.5314757823944092, "learning_rate": 0.00021643032899514573, "loss": 5.0202, "step": 412850 }, { "epoch": 2.489719130256509, "grad_norm": 2.350517988204956, "learning_rate": 0.00021641208377832166, "loss": 5.0802, "step": 412900 }, { "epoch": 2.4900206220378434, "grad_norm": 1.4355885982513428, "learning_rate": 0.00021639383733927794, "loss": 5.0385, "step": 412950 }, { "epoch": 2.490322113819177, "grad_norm": 2.865818500518799, "learning_rate": 0.0002163755896783504, "loss": 4.8859, "step": 413000 }, { "epoch": 2.4906236056005113, "grad_norm": 1.594516634941101, "learning_rate": 0.00021635734079587487, "loss": 4.3757, "step": 413050 }, { "epoch": 2.4909250973818455, "grad_norm": 1.982903003692627, "learning_rate": 0.00021633909069218715, "loss": 4.8842, "step": 413100 }, { "epoch": 2.4912265891631793, "grad_norm": 1.424961805343628, "learning_rate": 0.00021632083936762317, "loss": 5.003, "step": 413150 }, { "epoch": 2.4915280809445135, "grad_norm": 1.9693365097045898, "learning_rate": 0.00021630258682251878, "loss": 4.991, "step": 413200 }, { "epoch": 2.4918295727258473, "grad_norm": 1.5793927907943726, "learning_rate": 0.00021628433305720984, "loss": 4.8438, "step": 413250 }, { "epoch": 2.4921310645071815, "grad_norm": 1.8333888053894043, "learning_rate": 0.00021626607807203243, "loss": 4.8803, "step": 413300 }, { "epoch": 2.4924325562885157, "grad_norm": 1.946319341659546, "learning_rate": 0.00021624782186732239, "loss": 5.2321, "step": 413350 }, { "epoch": 2.4927340480698494, "grad_norm": 2.218961238861084, "learning_rate": 0.0002162295644434157, "loss": 5.0174, "step": 413400 }, { "epoch": 2.4930355398511836, "grad_norm": 2.433393716812134, "learning_rate": 0.0002162113058006484, "loss": 5.0432, "step": 413450 }, { "epoch": 2.493337031632518, "grad_norm": 2.306981325149536, "learning_rate": 0.00021619304593935653, "loss": 4.7247, "step": 413500 }, { "epoch": 2.4936385234138516, "grad_norm": 1.5646494626998901, "learning_rate": 0.0002161747848598761, "loss": 5.2165, "step": 413550 }, { "epoch": 2.493940015195186, "grad_norm": 1.389992356300354, "learning_rate": 0.00021615652256254316, "loss": 4.7333, "step": 413600 }, { "epoch": 2.49424150697652, "grad_norm": 1.5363869667053223, "learning_rate": 0.0002161382590476939, "loss": 5.3298, "step": 413650 }, { "epoch": 2.494542998757854, "grad_norm": 1.4164040088653564, "learning_rate": 0.00021611999431566433, "loss": 4.9991, "step": 413700 }, { "epoch": 2.494844490539188, "grad_norm": 1.7569663524627686, "learning_rate": 0.0002161017283667906, "loss": 4.8422, "step": 413750 }, { "epoch": 2.4951459823205218, "grad_norm": 1.5592014789581299, "learning_rate": 0.0002160834612014089, "loss": 5.3752, "step": 413800 }, { "epoch": 2.495447474101856, "grad_norm": 2.894568920135498, "learning_rate": 0.00021606519281985542, "loss": 4.9245, "step": 413850 }, { "epoch": 2.49574896588319, "grad_norm": 2.0541207790374756, "learning_rate": 0.00021604692322246634, "loss": 5.0429, "step": 413900 }, { "epoch": 2.496050457664524, "grad_norm": 0.2496439665555954, "learning_rate": 0.0002160286524095779, "loss": 4.8121, "step": 413950 }, { "epoch": 2.496351949445858, "grad_norm": 1.643494963645935, "learning_rate": 0.0002160103803815263, "loss": 5.1945, "step": 414000 }, { "epoch": 2.496351949445858, "eval_loss": 5.444316387176514, "eval_runtime": 39.0129, "eval_samples_per_second": 13.124, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.424004347504371, "step": 414000 }, { "epoch": 2.496653441227192, "grad_norm": 1.6074131727218628, "learning_rate": 0.00021599210713864794, "loss": 4.6083, "step": 414050 }, { "epoch": 2.496954933008526, "grad_norm": 0.4262903034687042, "learning_rate": 0.00021597383268127896, "loss": 4.8936, "step": 414100 }, { "epoch": 2.4972564247898603, "grad_norm": 1.402692437171936, "learning_rate": 0.00021595555700975572, "loss": 5.1868, "step": 414150 }, { "epoch": 2.497557916571194, "grad_norm": 1.5453397035598755, "learning_rate": 0.00021593728012441463, "loss": 5.1206, "step": 414200 }, { "epoch": 2.4978594083525283, "grad_norm": 2.516983985900879, "learning_rate": 0.00021591900202559192, "loss": 5.0575, "step": 414250 }, { "epoch": 2.4981609001338625, "grad_norm": 0.5871146321296692, "learning_rate": 0.00021590072271362415, "loss": 5.4401, "step": 414300 }, { "epoch": 2.4984623919151963, "grad_norm": 1.9867496490478516, "learning_rate": 0.0002158824421888476, "loss": 5.0713, "step": 414350 }, { "epoch": 2.4987638836965305, "grad_norm": 2.4403111934661865, "learning_rate": 0.00021586416045159867, "loss": 5.0933, "step": 414400 }, { "epoch": 2.4990653754778647, "grad_norm": 2.2304885387420654, "learning_rate": 0.0002158458775022139, "loss": 5.4043, "step": 414450 }, { "epoch": 2.4993668672591984, "grad_norm": 2.5284411907196045, "learning_rate": 0.00021582759334102972, "loss": 4.7496, "step": 414500 }, { "epoch": 2.4996683590405326, "grad_norm": 1.9858622550964355, "learning_rate": 0.0002158093079683826, "loss": 4.7752, "step": 414550 }, { "epoch": 2.4999698508218664, "grad_norm": 1.5633331537246704, "learning_rate": 0.0002157910213846091, "loss": 4.9134, "step": 414600 }, { "epoch": 2.5002713426032006, "grad_norm": 2.0271668434143066, "learning_rate": 0.00021577273359004573, "loss": 4.5758, "step": 414650 }, { "epoch": 2.500572834384535, "grad_norm": 2.508913278579712, "learning_rate": 0.00021575444458502905, "loss": 5.1119, "step": 414700 }, { "epoch": 2.5008743261658686, "grad_norm": 2.835437297821045, "learning_rate": 0.00021573615436989567, "loss": 4.8288, "step": 414750 }, { "epoch": 2.501175817947203, "grad_norm": 1.7382960319519043, "learning_rate": 0.0002157178629449822, "loss": 5.2169, "step": 414800 }, { "epoch": 2.5014773097285365, "grad_norm": 1.5926048755645752, "learning_rate": 0.0002156995703106252, "loss": 5.1156, "step": 414850 }, { "epoch": 2.5017788015098708, "grad_norm": 1.26852548122406, "learning_rate": 0.0002156812764671614, "loss": 4.7222, "step": 414900 }, { "epoch": 2.502080293291205, "grad_norm": 2.025730848312378, "learning_rate": 0.00021566298141492744, "loss": 4.6706, "step": 414950 }, { "epoch": 2.502381785072539, "grad_norm": 1.181594729423523, "learning_rate": 0.00021564468515425996, "loss": 4.7385, "step": 415000 }, { "epoch": 2.502683276853873, "grad_norm": 1.5839887857437134, "learning_rate": 0.0002156263876854958, "loss": 4.6572, "step": 415050 }, { "epoch": 2.502984768635207, "grad_norm": 2.422574520111084, "learning_rate": 0.0002156080890089716, "loss": 5.0997, "step": 415100 }, { "epoch": 2.503286260416541, "grad_norm": 2.2253172397613525, "learning_rate": 0.00021558978912502417, "loss": 5.1022, "step": 415150 }, { "epoch": 2.503587752197875, "grad_norm": 2.17917799949646, "learning_rate": 0.00021557148803399024, "loss": 5.4451, "step": 415200 }, { "epoch": 2.5038892439792093, "grad_norm": 2.026826858520508, "learning_rate": 0.0002155531857362067, "loss": 5.0516, "step": 415250 }, { "epoch": 2.504190735760543, "grad_norm": 1.650363802909851, "learning_rate": 0.00021553488223201027, "loss": 5.3286, "step": 415300 }, { "epoch": 2.5044922275418773, "grad_norm": 1.5575965642929077, "learning_rate": 0.00021551657752173789, "loss": 5.065, "step": 415350 }, { "epoch": 2.504793719323211, "grad_norm": 1.370139241218567, "learning_rate": 0.00021549827160572638, "loss": 5.1057, "step": 415400 }, { "epoch": 2.5050952111045452, "grad_norm": 1.8771356344223022, "learning_rate": 0.00021547996448431265, "loss": 4.7674, "step": 415450 }, { "epoch": 2.5053967028858795, "grad_norm": 1.4456268548965454, "learning_rate": 0.00021546165615783365, "loss": 4.7333, "step": 415500 }, { "epoch": 2.505698194667213, "grad_norm": 2.057443141937256, "learning_rate": 0.00021544334662662625, "loss": 5.2442, "step": 415550 }, { "epoch": 2.5059996864485474, "grad_norm": 1.9148682355880737, "learning_rate": 0.00021542503589102742, "loss": 5.1474, "step": 415600 }, { "epoch": 2.506301178229881, "grad_norm": 1.2595590353012085, "learning_rate": 0.00021540672395137424, "loss": 4.9612, "step": 415650 }, { "epoch": 2.5066026700112154, "grad_norm": 0.617211639881134, "learning_rate": 0.00021538841080800361, "loss": 5.1627, "step": 415700 }, { "epoch": 2.5069041617925496, "grad_norm": 2.2012734413146973, "learning_rate": 0.0002153700964612526, "loss": 5.0725, "step": 415750 }, { "epoch": 2.507205653573884, "grad_norm": 1.9759736061096191, "learning_rate": 0.00021535178091145824, "loss": 4.9887, "step": 415800 }, { "epoch": 2.5075071453552176, "grad_norm": 1.946066975593567, "learning_rate": 0.00021533346415895761, "loss": 4.973, "step": 415850 }, { "epoch": 2.5078086371365518, "grad_norm": 2.1494157314300537, "learning_rate": 0.00021531514620408784, "loss": 5.1526, "step": 415900 }, { "epoch": 2.5081101289178855, "grad_norm": 1.427943468093872, "learning_rate": 0.00021529682704718602, "loss": 4.6736, "step": 415950 }, { "epoch": 2.5084116206992197, "grad_norm": 0.95756596326828, "learning_rate": 0.00021527850668858923, "loss": 5.0792, "step": 416000 }, { "epoch": 2.508713112480554, "grad_norm": 0.7963663935661316, "learning_rate": 0.00021526018512863472, "loss": 5.0512, "step": 416050 }, { "epoch": 2.5090146042618877, "grad_norm": 1.2957830429077148, "learning_rate": 0.00021524186236765964, "loss": 4.774, "step": 416100 }, { "epoch": 2.509316096043222, "grad_norm": 1.5563851594924927, "learning_rate": 0.00021522353840600116, "loss": 5.1946, "step": 416150 }, { "epoch": 2.5096175878245557, "grad_norm": 1.5229873657226562, "learning_rate": 0.00021520521324399657, "loss": 5.1358, "step": 416200 }, { "epoch": 2.50991907960589, "grad_norm": 1.775650978088379, "learning_rate": 0.00021518688688198305, "loss": 5.3644, "step": 416250 }, { "epoch": 2.510220571387224, "grad_norm": 2.1246464252471924, "learning_rate": 0.00021516855932029793, "loss": 5.1593, "step": 416300 }, { "epoch": 2.510522063168558, "grad_norm": 2.6500871181488037, "learning_rate": 0.00021515023055927848, "loss": 5.0085, "step": 416350 }, { "epoch": 2.510823554949892, "grad_norm": 0.752747654914856, "learning_rate": 0.00021513190059926197, "loss": 4.8962, "step": 416400 }, { "epoch": 2.511125046731226, "grad_norm": 1.7084442377090454, "learning_rate": 0.0002151135694405858, "loss": 5.1694, "step": 416450 }, { "epoch": 2.51142653851256, "grad_norm": 1.4062426090240479, "learning_rate": 0.0002150952370835873, "loss": 5.535, "step": 416500 }, { "epoch": 2.5117280302938942, "grad_norm": 2.274604320526123, "learning_rate": 0.00021507690352860388, "loss": 5.2236, "step": 416550 }, { "epoch": 2.5120295220752284, "grad_norm": 1.716734528541565, "learning_rate": 0.0002150585687759729, "loss": 4.6949, "step": 416600 }, { "epoch": 2.512331013856562, "grad_norm": 1.5222011804580688, "learning_rate": 0.0002150402328260318, "loss": 5.0804, "step": 416650 }, { "epoch": 2.5126325056378964, "grad_norm": 3.029667615890503, "learning_rate": 0.00021502189567911806, "loss": 5.1369, "step": 416700 }, { "epoch": 2.51293399741923, "grad_norm": 2.3565499782562256, "learning_rate": 0.00021500355733556907, "loss": 5.0653, "step": 416750 }, { "epoch": 2.5132354892005644, "grad_norm": 1.8960013389587402, "learning_rate": 0.00021498521779572243, "loss": 5.1606, "step": 416800 }, { "epoch": 2.5135369809818986, "grad_norm": 2.0262153148651123, "learning_rate": 0.00021496687705991555, "loss": 5.0875, "step": 416850 }, { "epoch": 2.5138384727632324, "grad_norm": 1.5433884859085083, "learning_rate": 0.00021494853512848605, "loss": 5.312, "step": 416900 }, { "epoch": 2.5141399645445666, "grad_norm": 2.1629598140716553, "learning_rate": 0.00021493019200177145, "loss": 4.5919, "step": 416950 }, { "epoch": 2.5144414563259003, "grad_norm": 1.7479983568191528, "learning_rate": 0.00021491184768010927, "loss": 5.1036, "step": 417000 }, { "epoch": 2.5144414563259003, "eval_loss": 5.425546646118164, "eval_runtime": 39.1763, "eval_samples_per_second": 13.069, "eval_steps_per_second": 6.535, "eval_tts_loss": 7.3919147100519345, "step": 417000 }, { "epoch": 2.5147429481072345, "grad_norm": 1.710964560508728, "learning_rate": 0.0002148935021638372, "loss": 4.8173, "step": 417050 }, { "epoch": 2.5150444398885687, "grad_norm": 1.3747729063034058, "learning_rate": 0.00021487515545329288, "loss": 5.1994, "step": 417100 }, { "epoch": 2.515345931669903, "grad_norm": 1.914894938468933, "learning_rate": 0.00021485680754881385, "loss": 4.7812, "step": 417150 }, { "epoch": 2.5156474234512367, "grad_norm": 1.9385780096054077, "learning_rate": 0.0002148384584507378, "loss": 5.2464, "step": 417200 }, { "epoch": 2.515948915232571, "grad_norm": 1.5297154188156128, "learning_rate": 0.0002148201081594025, "loss": 5.1339, "step": 417250 }, { "epoch": 2.5162504070139047, "grad_norm": 1.8912707567214966, "learning_rate": 0.00021480175667514562, "loss": 4.9401, "step": 417300 }, { "epoch": 2.516551898795239, "grad_norm": 2.158046245574951, "learning_rate": 0.00021478340399830488, "loss": 5.1626, "step": 417350 }, { "epoch": 2.516853390576573, "grad_norm": 1.0224155187606812, "learning_rate": 0.00021476505012921803, "loss": 4.8011, "step": 417400 }, { "epoch": 2.517154882357907, "grad_norm": 1.773127555847168, "learning_rate": 0.00021474669506822287, "loss": 5.6121, "step": 417450 }, { "epoch": 2.517456374139241, "grad_norm": 1.4945800304412842, "learning_rate": 0.00021472833881565717, "loss": 5.089, "step": 417500 }, { "epoch": 2.517757865920575, "grad_norm": 1.985727071762085, "learning_rate": 0.0002147099813718588, "loss": 5.2692, "step": 417550 }, { "epoch": 2.518059357701909, "grad_norm": 1.8105002641677856, "learning_rate": 0.0002146916227371655, "loss": 4.9295, "step": 417600 }, { "epoch": 2.5183608494832432, "grad_norm": 2.252906560897827, "learning_rate": 0.00021467326291191524, "loss": 5.411, "step": 417650 }, { "epoch": 2.518662341264577, "grad_norm": 1.848681926727295, "learning_rate": 0.0002146549018964459, "loss": 5.1334, "step": 417700 }, { "epoch": 2.518963833045911, "grad_norm": 2.1905574798583984, "learning_rate": 0.00021463653969109532, "loss": 5.1154, "step": 417750 }, { "epoch": 2.519265324827245, "grad_norm": 1.485280156135559, "learning_rate": 0.00021461817629620148, "loss": 5.1392, "step": 417800 }, { "epoch": 2.519566816608579, "grad_norm": 2.3242690563201904, "learning_rate": 0.00021459981171210231, "loss": 4.741, "step": 417850 }, { "epoch": 2.5198683083899134, "grad_norm": 2.179414749145508, "learning_rate": 0.0002145814459391358, "loss": 4.7488, "step": 417900 }, { "epoch": 2.5201698001712476, "grad_norm": 1.5641307830810547, "learning_rate": 0.00021456307897764, "loss": 5.1009, "step": 417950 }, { "epoch": 2.5204712919525813, "grad_norm": 1.2862275838851929, "learning_rate": 0.00021454471082795278, "loss": 5.395, "step": 418000 }, { "epoch": 2.5207727837339156, "grad_norm": 2.098515033721924, "learning_rate": 0.00021452634149041227, "loss": 4.8351, "step": 418050 }, { "epoch": 2.5210742755152493, "grad_norm": 2.791541337966919, "learning_rate": 0.0002145079709653566, "loss": 5.1085, "step": 418100 }, { "epoch": 2.5213757672965835, "grad_norm": 1.3494024276733398, "learning_rate": 0.0002144895992531237, "loss": 5.4691, "step": 418150 }, { "epoch": 2.5216772590779177, "grad_norm": 1.2095019817352295, "learning_rate": 0.0002144712263540518, "loss": 4.7043, "step": 418200 }, { "epoch": 2.5219787508592515, "grad_norm": 1.7016528844833374, "learning_rate": 0.00021445285226847897, "loss": 4.8759, "step": 418250 }, { "epoch": 2.5222802426405857, "grad_norm": 2.117001533508301, "learning_rate": 0.00021443447699674346, "loss": 4.9873, "step": 418300 }, { "epoch": 2.5225817344219195, "grad_norm": 1.403304934501648, "learning_rate": 0.00021441610053918322, "loss": 4.7656, "step": 418350 }, { "epoch": 2.5228832262032537, "grad_norm": 2.1458849906921387, "learning_rate": 0.00021439772289613666, "loss": 5.1789, "step": 418400 }, { "epoch": 2.523184717984588, "grad_norm": 2.2983670234680176, "learning_rate": 0.0002143793440679419, "loss": 5.3488, "step": 418450 }, { "epoch": 2.5234862097659216, "grad_norm": 1.3644680976867676, "learning_rate": 0.00021436096405493717, "loss": 5.2467, "step": 418500 }, { "epoch": 2.523787701547256, "grad_norm": 1.5750573873519897, "learning_rate": 0.00021434258285746074, "loss": 5.2213, "step": 418550 }, { "epoch": 2.5240891933285896, "grad_norm": 1.4452975988388062, "learning_rate": 0.0002143242004758509, "loss": 4.9448, "step": 418600 }, { "epoch": 2.524390685109924, "grad_norm": 2.372547149658203, "learning_rate": 0.00021430581691044595, "loss": 5.1796, "step": 418650 }, { "epoch": 2.524692176891258, "grad_norm": 2.578754186630249, "learning_rate": 0.00021428743216158423, "loss": 4.9888, "step": 418700 }, { "epoch": 2.5249936686725922, "grad_norm": 2.7117671966552734, "learning_rate": 0.00021426904622960403, "loss": 4.5608, "step": 418750 }, { "epoch": 2.525295160453926, "grad_norm": 1.937497854232788, "learning_rate": 0.00021425065911484374, "loss": 5.2332, "step": 418800 }, { "epoch": 2.52559665223526, "grad_norm": 2.537116527557373, "learning_rate": 0.0002142322708176418, "loss": 5.6783, "step": 418850 }, { "epoch": 2.525898144016594, "grad_norm": 1.4955376386642456, "learning_rate": 0.00021421388133833656, "loss": 4.8378, "step": 418900 }, { "epoch": 2.526199635797928, "grad_norm": 0.37027284502983093, "learning_rate": 0.00021419549067726646, "loss": 5.2553, "step": 418950 }, { "epoch": 2.5265011275792624, "grad_norm": 1.6733505725860596, "learning_rate": 0.00021417709883476998, "loss": 5.3776, "step": 419000 }, { "epoch": 2.526802619360596, "grad_norm": 10.15834903717041, "learning_rate": 0.00021415870581118558, "loss": 5.1723, "step": 419050 }, { "epoch": 2.5271041111419303, "grad_norm": 1.1303606033325195, "learning_rate": 0.00021414031160685176, "loss": 4.9355, "step": 419100 }, { "epoch": 2.527405602923264, "grad_norm": 1.6573264598846436, "learning_rate": 0.00021412191622210706, "loss": 4.9888, "step": 419150 }, { "epoch": 2.5277070947045983, "grad_norm": 2.5752780437469482, "learning_rate": 0.00021410351965728996, "loss": 4.5841, "step": 419200 }, { "epoch": 2.5280085864859325, "grad_norm": 2.042680025100708, "learning_rate": 0.00021408512191273905, "loss": 5.196, "step": 419250 }, { "epoch": 2.5283100782672663, "grad_norm": 1.4299389123916626, "learning_rate": 0.000214066722988793, "loss": 4.9841, "step": 419300 }, { "epoch": 2.5286115700486005, "grad_norm": 1.882906198501587, "learning_rate": 0.00021404832288579028, "loss": 4.7273, "step": 419350 }, { "epoch": 2.5289130618299343, "grad_norm": 1.4092326164245605, "learning_rate": 0.00021402992160406958, "loss": 4.8318, "step": 419400 }, { "epoch": 2.5292145536112685, "grad_norm": 2.393930673599243, "learning_rate": 0.00021401151914396957, "loss": 4.9331, "step": 419450 }, { "epoch": 2.5295160453926027, "grad_norm": 2.445303440093994, "learning_rate": 0.0002139931155058289, "loss": 4.8068, "step": 419500 }, { "epoch": 2.529817537173937, "grad_norm": 1.500832438468933, "learning_rate": 0.0002139747106899863, "loss": 4.9247, "step": 419550 }, { "epoch": 2.5301190289552706, "grad_norm": 3.0521044731140137, "learning_rate": 0.00021395630469678038, "loss": 4.9211, "step": 419600 }, { "epoch": 2.530420520736605, "grad_norm": 1.6314510107040405, "learning_rate": 0.00021393789752654992, "loss": 4.6345, "step": 419650 }, { "epoch": 2.5307220125179386, "grad_norm": 1.4487990140914917, "learning_rate": 0.0002139194891796338, "loss": 4.5726, "step": 419700 }, { "epoch": 2.531023504299273, "grad_norm": 1.7112089395523071, "learning_rate": 0.00021390107965637065, "loss": 5.1793, "step": 419750 }, { "epoch": 2.531324996080607, "grad_norm": 2.770873785018921, "learning_rate": 0.00021388266895709928, "loss": 5.2715, "step": 419800 }, { "epoch": 2.5316264878619408, "grad_norm": 3.503709077835083, "learning_rate": 0.0002138642570821586, "loss": 4.9302, "step": 419850 }, { "epoch": 2.531927979643275, "grad_norm": 1.8862783908843994, "learning_rate": 0.00021384584403188743, "loss": 5.3476, "step": 419900 }, { "epoch": 2.5322294714246087, "grad_norm": 1.628469705581665, "learning_rate": 0.00021382742980662455, "loss": 5.243, "step": 419950 }, { "epoch": 2.532530963205943, "grad_norm": 2.629941701889038, "learning_rate": 0.00021380901440670892, "loss": 5.1793, "step": 420000 }, { "epoch": 2.532530963205943, "eval_loss": 5.443547248840332, "eval_runtime": 39.0803, "eval_samples_per_second": 13.101, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.293684179025643, "step": 420000 }, { "epoch": 2.532832454987277, "grad_norm": 1.7934505939483643, "learning_rate": 0.00021379059783247948, "loss": 5.2195, "step": 420050 }, { "epoch": 2.533133946768611, "grad_norm": 2.600430965423584, "learning_rate": 0.0002137721800842751, "loss": 4.7488, "step": 420100 }, { "epoch": 2.533435438549945, "grad_norm": 1.5007591247558594, "learning_rate": 0.00021375376116243474, "loss": 4.8594, "step": 420150 }, { "epoch": 2.5337369303312793, "grad_norm": 0.7689090371131897, "learning_rate": 0.00021373534106729742, "loss": 5.1531, "step": 420200 }, { "epoch": 2.534038422112613, "grad_norm": 2.1137754917144775, "learning_rate": 0.00021371691979920208, "loss": 5.1741, "step": 420250 }, { "epoch": 2.5343399138939473, "grad_norm": 1.6881871223449707, "learning_rate": 0.00021369849735848777, "loss": 5.6455, "step": 420300 }, { "epoch": 2.5346414056752815, "grad_norm": 2.011117696762085, "learning_rate": 0.00021368007374549354, "loss": 5.068, "step": 420350 }, { "epoch": 2.5349428974566153, "grad_norm": 0.6200054883956909, "learning_rate": 0.00021366164896055836, "loss": 4.7009, "step": 420400 }, { "epoch": 2.5352443892379495, "grad_norm": 2.2893311977386475, "learning_rate": 0.00021364322300402143, "loss": 4.9939, "step": 420450 }, { "epoch": 2.5355458810192832, "grad_norm": 1.9783929586410522, "learning_rate": 0.00021362479587622182, "loss": 5.135, "step": 420500 }, { "epoch": 2.5358473728006174, "grad_norm": 1.6566721200942993, "learning_rate": 0.00021360636757749858, "loss": 4.3269, "step": 420550 }, { "epoch": 2.5361488645819517, "grad_norm": 0.46983063220977783, "learning_rate": 0.00021358793810819098, "loss": 5.2145, "step": 420600 }, { "epoch": 2.5364503563632854, "grad_norm": 1.8325685262680054, "learning_rate": 0.0002135695074686381, "loss": 5.0318, "step": 420650 }, { "epoch": 2.5367518481446196, "grad_norm": 1.1181976795196533, "learning_rate": 0.00021355107565917915, "loss": 4.8219, "step": 420700 }, { "epoch": 2.5370533399259534, "grad_norm": 2.748225450515747, "learning_rate": 0.00021353264268015335, "loss": 5.3121, "step": 420750 }, { "epoch": 2.5373548317072876, "grad_norm": 0.9717361330986023, "learning_rate": 0.00021351420853189993, "loss": 5.16, "step": 420800 }, { "epoch": 2.537656323488622, "grad_norm": 2.0219967365264893, "learning_rate": 0.0002134957732147581, "loss": 4.9758, "step": 420850 }, { "epoch": 2.537957815269956, "grad_norm": 1.434787392616272, "learning_rate": 0.00021347733672906727, "loss": 4.7595, "step": 420900 }, { "epoch": 2.5382593070512898, "grad_norm": 2.3399174213409424, "learning_rate": 0.00021345889907516656, "loss": 5.1416, "step": 420950 }, { "epoch": 2.538560798832624, "grad_norm": 1.3550716638565063, "learning_rate": 0.0002134404602533954, "loss": 4.9593, "step": 421000 }, { "epoch": 2.5388622906139577, "grad_norm": 1.6576874256134033, "learning_rate": 0.0002134220202640931, "loss": 5.0365, "step": 421050 }, { "epoch": 2.539163782395292, "grad_norm": 2.8573832511901855, "learning_rate": 0.000213403579107599, "loss": 4.8333, "step": 421100 }, { "epoch": 2.539465274176626, "grad_norm": 2.552948474884033, "learning_rate": 0.00021338513678425256, "loss": 5.0434, "step": 421150 }, { "epoch": 2.53976676595796, "grad_norm": 1.7834923267364502, "learning_rate": 0.0002133666932943931, "loss": 5.0423, "step": 421200 }, { "epoch": 2.540068257739294, "grad_norm": 0.984713613986969, "learning_rate": 0.0002133482486383601, "loss": 5.0738, "step": 421250 }, { "epoch": 2.540369749520628, "grad_norm": 1.5305012464523315, "learning_rate": 0.00021332980281649298, "loss": 5.1355, "step": 421300 }, { "epoch": 2.540671241301962, "grad_norm": 2.3901586532592773, "learning_rate": 0.0002133113558291312, "loss": 5.2545, "step": 421350 }, { "epoch": 2.5409727330832963, "grad_norm": 1.7593632936477661, "learning_rate": 0.00021329290767661425, "loss": 5.2477, "step": 421400 }, { "epoch": 2.54127422486463, "grad_norm": 2.3454997539520264, "learning_rate": 0.0002132744583592817, "loss": 4.7555, "step": 421450 }, { "epoch": 2.5415757166459643, "grad_norm": 2.181087017059326, "learning_rate": 0.00021325600787747305, "loss": 5.2675, "step": 421500 }, { "epoch": 2.541877208427298, "grad_norm": 1.3670369386672974, "learning_rate": 0.00021323755623152782, "loss": 5.3999, "step": 421550 }, { "epoch": 2.5421787002086322, "grad_norm": 1.3552520275115967, "learning_rate": 0.0002132191034217856, "loss": 5.1958, "step": 421600 }, { "epoch": 2.5424801919899664, "grad_norm": 1.6503562927246094, "learning_rate": 0.00021320064944858603, "loss": 4.8058, "step": 421650 }, { "epoch": 2.5427816837713006, "grad_norm": 3.0021111965179443, "learning_rate": 0.00021318219431226866, "loss": 4.7624, "step": 421700 }, { "epoch": 2.5430831755526344, "grad_norm": 1.9366484880447388, "learning_rate": 0.00021316373801317322, "loss": 5.1359, "step": 421750 }, { "epoch": 2.5433846673339686, "grad_norm": 1.996944785118103, "learning_rate": 0.0002131452805516393, "loss": 4.9586, "step": 421800 }, { "epoch": 2.5436861591153024, "grad_norm": 1.0180977582931519, "learning_rate": 0.0002131268219280066, "loss": 5.0795, "step": 421850 }, { "epoch": 2.5439876508966366, "grad_norm": 1.8637950420379639, "learning_rate": 0.00021310836214261485, "loss": 5.0933, "step": 421900 }, { "epoch": 2.544289142677971, "grad_norm": 1.087514042854309, "learning_rate": 0.00021308990119580374, "loss": 4.7121, "step": 421950 }, { "epoch": 2.5445906344593046, "grad_norm": 1.795600414276123, "learning_rate": 0.000213071439087913, "loss": 4.629, "step": 422000 }, { "epoch": 2.5448921262406388, "grad_norm": 1.5380791425704956, "learning_rate": 0.00021305297581928243, "loss": 5.3971, "step": 422050 }, { "epoch": 2.5451936180219725, "grad_norm": 2.4753990173339844, "learning_rate": 0.00021303451139025192, "loss": 4.878, "step": 422100 }, { "epoch": 2.5454951098033067, "grad_norm": 2.713536262512207, "learning_rate": 0.00021301604580116108, "loss": 4.6627, "step": 422150 }, { "epoch": 2.545796601584641, "grad_norm": 1.5320104360580444, "learning_rate": 0.00021299757905234986, "loss": 4.833, "step": 422200 }, { "epoch": 2.5460980933659747, "grad_norm": 2.168733596801758, "learning_rate": 0.0002129791111441581, "loss": 4.7678, "step": 422250 }, { "epoch": 2.546399585147309, "grad_norm": 0.6986207365989685, "learning_rate": 0.00021296064207692575, "loss": 4.7228, "step": 422300 }, { "epoch": 2.5467010769286427, "grad_norm": 1.4305529594421387, "learning_rate": 0.00021294217185099257, "loss": 4.9791, "step": 422350 }, { "epoch": 2.547002568709977, "grad_norm": 1.5141892433166504, "learning_rate": 0.00021292370046669851, "loss": 4.9593, "step": 422400 }, { "epoch": 2.547304060491311, "grad_norm": 0.8756690621376038, "learning_rate": 0.00021290522792438357, "loss": 4.6218, "step": 422450 }, { "epoch": 2.5476055522726453, "grad_norm": 1.5788122415542603, "learning_rate": 0.0002128867542243877, "loss": 4.9018, "step": 422500 }, { "epoch": 2.547907044053979, "grad_norm": 1.5288728475570679, "learning_rate": 0.00021286827936705088, "loss": 4.7905, "step": 422550 }, { "epoch": 2.5482085358353133, "grad_norm": 1.6966372728347778, "learning_rate": 0.00021284980335271303, "loss": 4.8508, "step": 422600 }, { "epoch": 2.548510027616647, "grad_norm": 1.7925292253494263, "learning_rate": 0.00021283132618171426, "loss": 5.2535, "step": 422650 }, { "epoch": 2.5488115193979812, "grad_norm": 2.2459051609039307, "learning_rate": 0.00021281284785439462, "loss": 5.3127, "step": 422700 }, { "epoch": 2.5491130111793154, "grad_norm": 0.8796519041061401, "learning_rate": 0.0002127943683710942, "loss": 5.1337, "step": 422750 }, { "epoch": 2.549414502960649, "grad_norm": 2.0007574558258057, "learning_rate": 0.00021277588773215296, "loss": 4.9057, "step": 422800 }, { "epoch": 2.5497159947419834, "grad_norm": 0.9163351655006409, "learning_rate": 0.00021275740593791113, "loss": 5.5325, "step": 422850 }, { "epoch": 2.550017486523317, "grad_norm": 2.316559076309204, "learning_rate": 0.0002127389229887088, "loss": 4.6582, "step": 422900 }, { "epoch": 2.5503189783046514, "grad_norm": 3.622502326965332, "learning_rate": 0.0002127204388848861, "loss": 5.2855, "step": 422950 }, { "epoch": 2.5506204700859856, "grad_norm": 1.7728513479232788, "learning_rate": 0.0002127019536267832, "loss": 5.6815, "step": 423000 }, { "epoch": 2.5506204700859856, "eval_loss": 5.432579040527344, "eval_runtime": 39.0521, "eval_samples_per_second": 13.111, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.382750212073503, "step": 423000 }, { "epoch": 2.5509219618673193, "grad_norm": 2.0487749576568604, "learning_rate": 0.00021268346721474042, "loss": 4.8091, "step": 423050 }, { "epoch": 2.5512234536486535, "grad_norm": 2.8243520259857178, "learning_rate": 0.00021266497964909787, "loss": 5.0356, "step": 423100 }, { "epoch": 2.5515249454299873, "grad_norm": 2.226423740386963, "learning_rate": 0.0002126464909301957, "loss": 4.8612, "step": 423150 }, { "epoch": 2.5518264372113215, "grad_norm": 1.64911687374115, "learning_rate": 0.00021262800105837436, "loss": 5.0113, "step": 423200 }, { "epoch": 2.5521279289926557, "grad_norm": 1.6644483804702759, "learning_rate": 0.000212609510033974, "loss": 5.1992, "step": 423250 }, { "epoch": 2.55242942077399, "grad_norm": 1.5039016008377075, "learning_rate": 0.00021259101785733498, "loss": 5.0297, "step": 423300 }, { "epoch": 2.5527309125553237, "grad_norm": 1.9635252952575684, "learning_rate": 0.00021257252452879755, "loss": 4.9277, "step": 423350 }, { "epoch": 2.553032404336658, "grad_norm": 1.1381502151489258, "learning_rate": 0.00021255403004870213, "loss": 4.6182, "step": 423400 }, { "epoch": 2.5533338961179917, "grad_norm": 1.7303415536880493, "learning_rate": 0.00021253553441738906, "loss": 5.0997, "step": 423450 }, { "epoch": 2.553635387899326, "grad_norm": 2.2562568187713623, "learning_rate": 0.00021251703763519873, "loss": 5.0705, "step": 423500 }, { "epoch": 2.55393687968066, "grad_norm": 1.6506770849227905, "learning_rate": 0.00021249853970247153, "loss": 5.1062, "step": 423550 }, { "epoch": 2.554238371461994, "grad_norm": 1.871446967124939, "learning_rate": 0.0002124800406195479, "loss": 5.0181, "step": 423600 }, { "epoch": 2.554539863243328, "grad_norm": 2.0990209579467773, "learning_rate": 0.00021246154038676824, "loss": 5.3021, "step": 423650 }, { "epoch": 2.554841355024662, "grad_norm": 2.7564797401428223, "learning_rate": 0.00021244303900447317, "loss": 5.2734, "step": 423700 }, { "epoch": 2.555142846805996, "grad_norm": 1.4455965757369995, "learning_rate": 0.00021242453647300294, "loss": 4.8158, "step": 423750 }, { "epoch": 2.55544433858733, "grad_norm": 1.84609055519104, "learning_rate": 0.0002124060327926983, "loss": 4.9951, "step": 423800 }, { "epoch": 2.555745830368664, "grad_norm": 1.4668971300125122, "learning_rate": 0.00021238752796389967, "loss": 5.3977, "step": 423850 }, { "epoch": 2.556047322149998, "grad_norm": 1.3618099689483643, "learning_rate": 0.00021236902198694762, "loss": 5.409, "step": 423900 }, { "epoch": 2.5563488139313324, "grad_norm": 0.8206979036331177, "learning_rate": 0.00021235051486218274, "loss": 4.5917, "step": 423950 }, { "epoch": 2.556650305712666, "grad_norm": 2.665447473526001, "learning_rate": 0.00021233200658994558, "loss": 5.0006, "step": 424000 }, { "epoch": 2.5569517974940004, "grad_norm": 1.9175881147384644, "learning_rate": 0.0002123134971705768, "loss": 4.7876, "step": 424050 }, { "epoch": 2.5572532892753346, "grad_norm": 2.0093696117401123, "learning_rate": 0.00021229498660441703, "loss": 5.0767, "step": 424100 }, { "epoch": 2.5575547810566683, "grad_norm": 2.108736515045166, "learning_rate": 0.0002122764748918069, "loss": 5.2705, "step": 424150 }, { "epoch": 2.5578562728380025, "grad_norm": 1.387010931968689, "learning_rate": 0.00021225796203308716, "loss": 4.7109, "step": 424200 }, { "epoch": 2.5581577646193363, "grad_norm": 2.4181385040283203, "learning_rate": 0.00021223944802859846, "loss": 5.1261, "step": 424250 }, { "epoch": 2.5584592564006705, "grad_norm": 1.5594189167022705, "learning_rate": 0.00021222093287868155, "loss": 4.564, "step": 424300 }, { "epoch": 2.5587607481820047, "grad_norm": 1.840827465057373, "learning_rate": 0.00021220241658367715, "loss": 5.2254, "step": 424350 }, { "epoch": 2.5590622399633385, "grad_norm": 1.3325536251068115, "learning_rate": 0.00021218389914392607, "loss": 4.6631, "step": 424400 }, { "epoch": 2.5593637317446727, "grad_norm": 2.2882425785064697, "learning_rate": 0.00021216538055976905, "loss": 5.1748, "step": 424450 }, { "epoch": 2.5596652235260065, "grad_norm": 1.7731163501739502, "learning_rate": 0.00021214686083154694, "loss": 5.2913, "step": 424500 }, { "epoch": 2.5599667153073407, "grad_norm": 1.4015518426895142, "learning_rate": 0.00021212833995960052, "loss": 4.4144, "step": 424550 }, { "epoch": 2.560268207088675, "grad_norm": 1.9220877885818481, "learning_rate": 0.00021210981794427065, "loss": 4.7021, "step": 424600 }, { "epoch": 2.560569698870009, "grad_norm": 1.8579332828521729, "learning_rate": 0.00021209129478589825, "loss": 4.863, "step": 424650 }, { "epoch": 2.560871190651343, "grad_norm": 1.1203234195709229, "learning_rate": 0.00021207277048482416, "loss": 4.564, "step": 424700 }, { "epoch": 2.561172682432677, "grad_norm": 1.6948200464248657, "learning_rate": 0.0002120542450413893, "loss": 4.8277, "step": 424750 }, { "epoch": 2.561474174214011, "grad_norm": 1.3919850587844849, "learning_rate": 0.00021203571845593462, "loss": 5.3763, "step": 424800 }, { "epoch": 2.561775665995345, "grad_norm": 3.0718371868133545, "learning_rate": 0.00021201719072880108, "loss": 4.7685, "step": 424850 }, { "epoch": 2.562077157776679, "grad_norm": 2.8036177158355713, "learning_rate": 0.00021199866186032967, "loss": 4.8487, "step": 424900 }, { "epoch": 2.562378649558013, "grad_norm": 1.6836321353912354, "learning_rate": 0.00021198013185086135, "loss": 5.3925, "step": 424950 }, { "epoch": 2.562680141339347, "grad_norm": 2.1097192764282227, "learning_rate": 0.00021196160070073716, "loss": 5.1342, "step": 425000 }, { "epoch": 2.562981633120681, "grad_norm": 2.348322868347168, "learning_rate": 0.0002119430684102981, "loss": 5.1314, "step": 425050 }, { "epoch": 2.563283124902015, "grad_norm": 1.6441493034362793, "learning_rate": 0.00021192453497988531, "loss": 4.9318, "step": 425100 }, { "epoch": 2.5635846166833494, "grad_norm": 2.792287826538086, "learning_rate": 0.0002119060004098398, "loss": 4.9943, "step": 425150 }, { "epoch": 2.563886108464683, "grad_norm": 1.7568774223327637, "learning_rate": 0.00021188746470050265, "loss": 4.9458, "step": 425200 }, { "epoch": 2.5641876002460173, "grad_norm": 2.323045253753662, "learning_rate": 0.00021186892785221508, "loss": 5.3793, "step": 425250 }, { "epoch": 2.564489092027351, "grad_norm": 1.4742705821990967, "learning_rate": 0.00021185038986531822, "loss": 5.0479, "step": 425300 }, { "epoch": 2.5647905838086853, "grad_norm": 1.6015931367874146, "learning_rate": 0.00021183185074015313, "loss": 4.9799, "step": 425350 }, { "epoch": 2.5650920755900195, "grad_norm": 1.4848779439926147, "learning_rate": 0.00021181331047706112, "loss": 5.0368, "step": 425400 }, { "epoch": 2.5653935673713537, "grad_norm": 1.2071497440338135, "learning_rate": 0.00021179476907638332, "loss": 5.078, "step": 425450 }, { "epoch": 2.5656950591526875, "grad_norm": 0.5560643672943115, "learning_rate": 0.00021177622653846098, "loss": 4.7079, "step": 425500 }, { "epoch": 2.5659965509340217, "grad_norm": 1.9852700233459473, "learning_rate": 0.00021175768286363535, "loss": 5.173, "step": 425550 }, { "epoch": 2.5662980427153554, "grad_norm": 2.128666639328003, "learning_rate": 0.0002117391380522477, "loss": 5.1515, "step": 425600 }, { "epoch": 2.5665995344966896, "grad_norm": 1.6896480321884155, "learning_rate": 0.00021172059210463934, "loss": 4.6339, "step": 425650 }, { "epoch": 2.566901026278024, "grad_norm": 1.6576464176177979, "learning_rate": 0.00021170204502115152, "loss": 4.8534, "step": 425700 }, { "epoch": 2.5672025180593576, "grad_norm": 2.7850563526153564, "learning_rate": 0.00021168349680212564, "loss": 5.0716, "step": 425750 }, { "epoch": 2.567504009840692, "grad_norm": 1.4856584072113037, "learning_rate": 0.000211664947447903, "loss": 4.6193, "step": 425800 }, { "epoch": 2.5678055016220256, "grad_norm": 1.7477819919586182, "learning_rate": 0.00021164639695882504, "loss": 4.8872, "step": 425850 }, { "epoch": 2.56810699340336, "grad_norm": 2.166832685470581, "learning_rate": 0.00021162784533523314, "loss": 5.15, "step": 425900 }, { "epoch": 2.568408485184694, "grad_norm": 2.1870615482330322, "learning_rate": 0.00021160929257746863, "loss": 4.9731, "step": 425950 }, { "epoch": 2.5687099769660278, "grad_norm": 1.4744091033935547, "learning_rate": 0.00021159073868587307, "loss": 5.2901, "step": 426000 }, { "epoch": 2.5687099769660278, "eval_loss": 5.438728332519531, "eval_runtime": 39.143, "eval_samples_per_second": 13.08, "eval_steps_per_second": 6.54, "eval_tts_loss": 7.3920589456561165, "step": 426000 }, { "epoch": 2.569011468747362, "grad_norm": 1.7967417240142822, "learning_rate": 0.00021157218366078782, "loss": 4.9014, "step": 426050 }, { "epoch": 2.5693129605286957, "grad_norm": 0.6654082536697388, "learning_rate": 0.0002115536275025544, "loss": 4.6468, "step": 426100 }, { "epoch": 2.56961445231003, "grad_norm": 0.862086296081543, "learning_rate": 0.00021153507021151434, "loss": 4.9564, "step": 426150 }, { "epoch": 2.569915944091364, "grad_norm": 1.250229001045227, "learning_rate": 0.00021151651178800908, "loss": 5.1456, "step": 426200 }, { "epoch": 2.5702174358726984, "grad_norm": 2.096863269805908, "learning_rate": 0.00021149795223238027, "loss": 4.9264, "step": 426250 }, { "epoch": 2.570518927654032, "grad_norm": 0.8303429484367371, "learning_rate": 0.00021147939154496937, "loss": 5.128, "step": 426300 }, { "epoch": 2.5708204194353663, "grad_norm": 2.2037928104400635, "learning_rate": 0.00021146082972611798, "loss": 5.16, "step": 426350 }, { "epoch": 2.5711219112167, "grad_norm": 2.162827491760254, "learning_rate": 0.00021144226677616778, "loss": 4.5328, "step": 426400 }, { "epoch": 2.5714234029980343, "grad_norm": 1.2609028816223145, "learning_rate": 0.00021142370269546033, "loss": 4.8031, "step": 426450 }, { "epoch": 2.5717248947793685, "grad_norm": 1.4578821659088135, "learning_rate": 0.0002114051374843373, "loss": 5.3376, "step": 426500 }, { "epoch": 2.5720263865607023, "grad_norm": 1.8382235765457153, "learning_rate": 0.0002113865711431403, "loss": 4.5677, "step": 426550 }, { "epoch": 2.5723278783420365, "grad_norm": 1.9415616989135742, "learning_rate": 0.0002113680036722111, "loss": 4.9473, "step": 426600 }, { "epoch": 2.5726293701233702, "grad_norm": 2.0255112648010254, "learning_rate": 0.00021134943507189135, "loss": 4.6954, "step": 426650 }, { "epoch": 2.5729308619047044, "grad_norm": 2.113623857498169, "learning_rate": 0.0002113308653425228, "loss": 4.8079, "step": 426700 }, { "epoch": 2.5732323536860386, "grad_norm": 1.6419122219085693, "learning_rate": 0.0002113122944844472, "loss": 4.9818, "step": 426750 }, { "epoch": 2.5735338454673724, "grad_norm": 1.7208352088928223, "learning_rate": 0.0002112937224980063, "loss": 5.2185, "step": 426800 }, { "epoch": 2.5738353372487066, "grad_norm": 2.1216986179351807, "learning_rate": 0.00021127514938354194, "loss": 5.2421, "step": 426850 }, { "epoch": 2.5741368290300404, "grad_norm": 2.6113133430480957, "learning_rate": 0.0002112565751413959, "loss": 4.878, "step": 426900 }, { "epoch": 2.5744383208113746, "grad_norm": 1.9708194732666016, "learning_rate": 0.00021123799977191, "loss": 4.8524, "step": 426950 }, { "epoch": 2.574739812592709, "grad_norm": 1.6084227561950684, "learning_rate": 0.00021121942327542611, "loss": 4.9002, "step": 427000 }, { "epoch": 2.575041304374043, "grad_norm": 2.5538344383239746, "learning_rate": 0.0002112008456522861, "loss": 5.1493, "step": 427050 }, { "epoch": 2.5753427961553768, "grad_norm": 1.3628551959991455, "learning_rate": 0.00021118226690283187, "loss": 4.8049, "step": 427100 }, { "epoch": 2.575644287936711, "grad_norm": 2.262824058532715, "learning_rate": 0.0002111636870274053, "loss": 5.0142, "step": 427150 }, { "epoch": 2.5759457797180447, "grad_norm": 1.1221317052841187, "learning_rate": 0.00021114510602634838, "loss": 4.6256, "step": 427200 }, { "epoch": 2.576247271499379, "grad_norm": 1.732358455657959, "learning_rate": 0.00021112652390000304, "loss": 5.0345, "step": 427250 }, { "epoch": 2.576548763280713, "grad_norm": 1.7225995063781738, "learning_rate": 0.00021110794064871125, "loss": 4.7873, "step": 427300 }, { "epoch": 2.576850255062047, "grad_norm": 1.7759149074554443, "learning_rate": 0.000211089356272815, "loss": 4.8097, "step": 427350 }, { "epoch": 2.577151746843381, "grad_norm": 1.3132575750350952, "learning_rate": 0.00021107077077265636, "loss": 4.9742, "step": 427400 }, { "epoch": 2.577453238624715, "grad_norm": 1.8613638877868652, "learning_rate": 0.0002110521841485773, "loss": 5.0765, "step": 427450 }, { "epoch": 2.577754730406049, "grad_norm": 1.2297677993774414, "learning_rate": 0.00021103359640091998, "loss": 4.7033, "step": 427500 }, { "epoch": 2.5780562221873833, "grad_norm": 2.165574550628662, "learning_rate": 0.00021101500753002635, "loss": 5.0665, "step": 427550 }, { "epoch": 2.578357713968717, "grad_norm": 2.182286262512207, "learning_rate": 0.00021099641753623863, "loss": 5.0581, "step": 427600 }, { "epoch": 2.5786592057500513, "grad_norm": 1.1033179759979248, "learning_rate": 0.0002109778264198989, "loss": 5.0116, "step": 427650 }, { "epoch": 2.5789606975313855, "grad_norm": 2.096179485321045, "learning_rate": 0.00021095923418134924, "loss": 5.0017, "step": 427700 }, { "epoch": 2.5792621893127192, "grad_norm": 1.9498591423034668, "learning_rate": 0.00021094064082093192, "loss": 5.1438, "step": 427750 }, { "epoch": 2.5795636810940534, "grad_norm": 1.4383617639541626, "learning_rate": 0.00021092204633898899, "loss": 5.5458, "step": 427800 }, { "epoch": 2.5798651728753876, "grad_norm": 1.7094765901565552, "learning_rate": 0.00021090345073586286, "loss": 5.0323, "step": 427850 }, { "epoch": 2.5801666646567214, "grad_norm": 4.90307092666626, "learning_rate": 0.00021088485401189557, "loss": 5.2308, "step": 427900 }, { "epoch": 2.5804681564380556, "grad_norm": 2.52518367767334, "learning_rate": 0.00021086625616742937, "loss": 5.1246, "step": 427950 }, { "epoch": 2.5807696482193894, "grad_norm": 1.6383116245269775, "learning_rate": 0.00021084765720280667, "loss": 4.7351, "step": 428000 }, { "epoch": 2.5810711400007236, "grad_norm": 1.860108494758606, "learning_rate": 0.00021082905711836963, "loss": 4.7738, "step": 428050 }, { "epoch": 2.581372631782058, "grad_norm": 2.453223466873169, "learning_rate": 0.00021081045591446064, "loss": 5.2645, "step": 428100 }, { "epoch": 2.5816741235633915, "grad_norm": 0.8130080699920654, "learning_rate": 0.00021079185359142196, "loss": 4.934, "step": 428150 }, { "epoch": 2.5819756153447257, "grad_norm": 1.20054030418396, "learning_rate": 0.000210773250149596, "loss": 5.2325, "step": 428200 }, { "epoch": 2.5822771071260595, "grad_norm": 2.6718547344207764, "learning_rate": 0.0002107546455893251, "loss": 4.9261, "step": 428250 }, { "epoch": 2.5825785989073937, "grad_norm": 1.9032419919967651, "learning_rate": 0.0002107360399109516, "loss": 5.0412, "step": 428300 }, { "epoch": 2.582880090688728, "grad_norm": 1.6206403970718384, "learning_rate": 0.000210717433114818, "loss": 4.8634, "step": 428350 }, { "epoch": 2.583181582470062, "grad_norm": 1.9246288537979126, "learning_rate": 0.00021069882520126666, "loss": 4.9801, "step": 428400 }, { "epoch": 2.583483074251396, "grad_norm": 1.8151098489761353, "learning_rate": 0.00021068021617064013, "loss": 5.2244, "step": 428450 }, { "epoch": 2.58378456603273, "grad_norm": 1.3262354135513306, "learning_rate": 0.00021066160602328075, "loss": 4.8666, "step": 428500 }, { "epoch": 2.584086057814064, "grad_norm": 1.574084758758545, "learning_rate": 0.00021064299475953105, "loss": 4.668, "step": 428550 }, { "epoch": 2.584387549595398, "grad_norm": 1.767253041267395, "learning_rate": 0.00021062438237973367, "loss": 4.9168, "step": 428600 }, { "epoch": 2.5846890413767323, "grad_norm": 1.0300639867782593, "learning_rate": 0.000210605768884231, "loss": 4.5842, "step": 428650 }, { "epoch": 2.584990533158066, "grad_norm": 2.0200486183166504, "learning_rate": 0.00021058715427336563, "loss": 4.8353, "step": 428700 }, { "epoch": 2.5852920249394002, "grad_norm": 1.5956798791885376, "learning_rate": 0.00021056853854748016, "loss": 5.2705, "step": 428750 }, { "epoch": 2.585593516720734, "grad_norm": 1.6400682926177979, "learning_rate": 0.0002105499217069172, "loss": 4.9552, "step": 428800 }, { "epoch": 2.585895008502068, "grad_norm": 0.5568464398384094, "learning_rate": 0.00021053130375201933, "loss": 4.6745, "step": 428850 }, { "epoch": 2.5861965002834024, "grad_norm": 1.5761618614196777, "learning_rate": 0.0002105126846831292, "loss": 4.8978, "step": 428900 }, { "epoch": 2.586497992064736, "grad_norm": 2.3632595539093018, "learning_rate": 0.00021049406450058942, "loss": 4.828, "step": 428950 }, { "epoch": 2.5867994838460704, "grad_norm": 2.4223217964172363, "learning_rate": 0.00021047544320474272, "loss": 4.9907, "step": 429000 }, { "epoch": 2.5867994838460704, "eval_loss": 5.443655014038086, "eval_runtime": 39.2336, "eval_samples_per_second": 13.05, "eval_steps_per_second": 6.525, "eval_tts_loss": 7.432236665005835, "step": 429000 }, { "epoch": 2.587100975627404, "grad_norm": 1.4144049882888794, "learning_rate": 0.00021045682079593182, "loss": 4.9024, "step": 429050 }, { "epoch": 2.5874024674087384, "grad_norm": 1.5995838642120361, "learning_rate": 0.00021043819727449944, "loss": 5.1288, "step": 429100 }, { "epoch": 2.5877039591900726, "grad_norm": 1.2925347089767456, "learning_rate": 0.00021041957264078821, "loss": 5.1078, "step": 429150 }, { "epoch": 2.5880054509714068, "grad_norm": 2.0342533588409424, "learning_rate": 0.000210400946895141, "loss": 5.2612, "step": 429200 }, { "epoch": 2.5883069427527405, "grad_norm": 1.868704915046692, "learning_rate": 0.0002103823200379006, "loss": 4.7244, "step": 429250 }, { "epoch": 2.5886084345340747, "grad_norm": 1.0116186141967773, "learning_rate": 0.00021036369206940973, "loss": 4.727, "step": 429300 }, { "epoch": 2.5889099263154085, "grad_norm": 0.7202889919281006, "learning_rate": 0.00021034506299001126, "loss": 4.8811, "step": 429350 }, { "epoch": 2.5892114180967427, "grad_norm": 1.9350417852401733, "learning_rate": 0.00021032643280004802, "loss": 4.7856, "step": 429400 }, { "epoch": 2.589512909878077, "grad_norm": 2.0922632217407227, "learning_rate": 0.0002103078014998629, "loss": 5.0342, "step": 429450 }, { "epoch": 2.5898144016594107, "grad_norm": 0.5161727666854858, "learning_rate": 0.00021028916908979873, "loss": 4.9802, "step": 429500 }, { "epoch": 2.590115893440745, "grad_norm": 1.341262698173523, "learning_rate": 0.00021027053557019838, "loss": 5.1619, "step": 429550 }, { "epoch": 2.5904173852220787, "grad_norm": 1.7084345817565918, "learning_rate": 0.00021025190094140493, "loss": 5.0427, "step": 429600 }, { "epoch": 2.590718877003413, "grad_norm": 2.4819302558898926, "learning_rate": 0.0002102332652037612, "loss": 5.0245, "step": 429650 }, { "epoch": 2.591020368784747, "grad_norm": 1.6827774047851562, "learning_rate": 0.00021021462835761018, "loss": 4.9599, "step": 429700 }, { "epoch": 2.591321860566081, "grad_norm": 1.7200435400009155, "learning_rate": 0.00021019599040329485, "loss": 4.7833, "step": 429750 }, { "epoch": 2.591623352347415, "grad_norm": 2.710354804992676, "learning_rate": 0.00021017735134115824, "loss": 4.8636, "step": 429800 }, { "epoch": 2.591924844128749, "grad_norm": 1.6863160133361816, "learning_rate": 0.00021015871117154332, "loss": 5.5422, "step": 429850 }, { "epoch": 2.592226335910083, "grad_norm": 1.3629528284072876, "learning_rate": 0.00021014006989479323, "loss": 5.132, "step": 429900 }, { "epoch": 2.592527827691417, "grad_norm": 1.4817181825637817, "learning_rate": 0.00021012142751125092, "loss": 4.8952, "step": 429950 }, { "epoch": 2.5928293194727514, "grad_norm": 2.002751588821411, "learning_rate": 0.00021010278402125955, "loss": 4.5721, "step": 430000 }, { "epoch": 2.593130811254085, "grad_norm": 1.0604078769683838, "learning_rate": 0.00021008413942516223, "loss": 4.8287, "step": 430050 }, { "epoch": 2.5934323030354194, "grad_norm": 0.9638405442237854, "learning_rate": 0.00021006549372330207, "loss": 4.7496, "step": 430100 }, { "epoch": 2.593733794816753, "grad_norm": 1.6076321601867676, "learning_rate": 0.00021004684691602218, "loss": 4.9707, "step": 430150 }, { "epoch": 2.5940352865980874, "grad_norm": 1.4437663555145264, "learning_rate": 0.00021002819900366578, "loss": 4.7702, "step": 430200 }, { "epoch": 2.5943367783794216, "grad_norm": 0.6488787531852722, "learning_rate": 0.00021000954998657604, "loss": 4.7083, "step": 430250 }, { "epoch": 2.5946382701607553, "grad_norm": 1.6266508102416992, "learning_rate": 0.00020999089986509622, "loss": 4.8427, "step": 430300 }, { "epoch": 2.5949397619420895, "grad_norm": 1.8996864557266235, "learning_rate": 0.00020997224863956944, "loss": 5.0539, "step": 430350 }, { "epoch": 2.5952412537234233, "grad_norm": 1.435868501663208, "learning_rate": 0.00020995359631033903, "loss": 5.2636, "step": 430400 }, { "epoch": 2.5955427455047575, "grad_norm": 1.442514419555664, "learning_rate": 0.0002099349428777482, "loss": 4.9542, "step": 430450 }, { "epoch": 2.5958442372860917, "grad_norm": 1.5477800369262695, "learning_rate": 0.00020991628834214035, "loss": 4.737, "step": 430500 }, { "epoch": 2.5961457290674255, "grad_norm": 2.1602680683135986, "learning_rate": 0.00020989763270385867, "loss": 4.6124, "step": 430550 }, { "epoch": 2.5964472208487597, "grad_norm": 2.2628870010375977, "learning_rate": 0.0002098789759632465, "loss": 4.7329, "step": 430600 }, { "epoch": 2.5967487126300934, "grad_norm": 0.6379478573799133, "learning_rate": 0.00020986031812064734, "loss": 4.5595, "step": 430650 }, { "epoch": 2.5970502044114276, "grad_norm": 1.2052439451217651, "learning_rate": 0.00020984165917640438, "loss": 5.0795, "step": 430700 }, { "epoch": 2.597351696192762, "grad_norm": 2.1082026958465576, "learning_rate": 0.00020982299913086108, "loss": 4.8136, "step": 430750 }, { "epoch": 2.597653187974096, "grad_norm": 1.7384589910507202, "learning_rate": 0.00020980433798436086, "loss": 4.9339, "step": 430800 }, { "epoch": 2.59795467975543, "grad_norm": 0.9823183417320251, "learning_rate": 0.00020978567573724715, "loss": 4.3942, "step": 430850 }, { "epoch": 2.598256171536764, "grad_norm": 1.9272751808166504, "learning_rate": 0.0002097670123898634, "loss": 4.7958, "step": 430900 }, { "epoch": 2.598557663318098, "grad_norm": 1.6896345615386963, "learning_rate": 0.00020974834794255308, "loss": 5.1408, "step": 430950 }, { "epoch": 2.598859155099432, "grad_norm": 1.5082416534423828, "learning_rate": 0.00020972968239565967, "loss": 5.261, "step": 431000 }, { "epoch": 2.599160646880766, "grad_norm": 2.760913610458374, "learning_rate": 0.00020971101574952672, "loss": 4.9822, "step": 431050 }, { "epoch": 2.5994621386621, "grad_norm": 1.5691756010055542, "learning_rate": 0.0002096923480044977, "loss": 5.1184, "step": 431100 }, { "epoch": 2.599763630443434, "grad_norm": 1.8666983842849731, "learning_rate": 0.0002096736791609162, "loss": 5.0553, "step": 431150 }, { "epoch": 2.600065122224768, "grad_norm": 2.44781756401062, "learning_rate": 0.00020965500921912584, "loss": 4.7931, "step": 431200 }, { "epoch": 2.600366614006102, "grad_norm": 1.642311453819275, "learning_rate": 0.00020963633817947016, "loss": 4.7879, "step": 431250 }, { "epoch": 2.6006681057874363, "grad_norm": 2.484342575073242, "learning_rate": 0.0002096176660422927, "loss": 5.0337, "step": 431300 }, { "epoch": 2.6009695975687706, "grad_norm": 1.6953564882278442, "learning_rate": 0.00020959899280793726, "loss": 4.8519, "step": 431350 }, { "epoch": 2.6012710893501043, "grad_norm": 1.7375868558883667, "learning_rate": 0.0002095803184767474, "loss": 5.2554, "step": 431400 }, { "epoch": 2.6015725811314385, "grad_norm": 1.6857019662857056, "learning_rate": 0.00020956164304906678, "loss": 4.5984, "step": 431450 }, { "epoch": 2.6018740729127723, "grad_norm": 1.5467935800552368, "learning_rate": 0.00020954296652523914, "loss": 5.138, "step": 431500 }, { "epoch": 2.6021755646941065, "grad_norm": 1.1351580619812012, "learning_rate": 0.00020952428890560814, "loss": 4.5315, "step": 431550 }, { "epoch": 2.6024770564754407, "grad_norm": 1.5669673681259155, "learning_rate": 0.00020950561019051757, "loss": 5.1492, "step": 431600 }, { "epoch": 2.6027785482567745, "grad_norm": 1.4264525175094604, "learning_rate": 0.0002094869303803112, "loss": 4.8775, "step": 431650 }, { "epoch": 2.6030800400381087, "grad_norm": 1.6845742464065552, "learning_rate": 0.00020946824947533273, "loss": 5.4188, "step": 431700 }, { "epoch": 2.6033815318194424, "grad_norm": 1.8952301740646362, "learning_rate": 0.00020944956747592598, "loss": 4.8869, "step": 431750 }, { "epoch": 2.6036830236007766, "grad_norm": 0.7989301681518555, "learning_rate": 0.00020943088438243478, "loss": 4.4758, "step": 431800 }, { "epoch": 2.603984515382111, "grad_norm": 2.144437313079834, "learning_rate": 0.000209412200195203, "loss": 4.9837, "step": 431850 }, { "epoch": 2.6042860071634446, "grad_norm": 1.541782259941101, "learning_rate": 0.00020939351491457443, "loss": 4.7371, "step": 431900 }, { "epoch": 2.604587498944779, "grad_norm": 1.9340089559555054, "learning_rate": 0.00020937482854089303, "loss": 5.1243, "step": 431950 }, { "epoch": 2.6048889907261126, "grad_norm": 2.2603085041046143, "learning_rate": 0.0002093561410745026, "loss": 5.1007, "step": 432000 }, { "epoch": 2.6048889907261126, "eval_loss": 5.425371170043945, "eval_runtime": 39.1497, "eval_samples_per_second": 13.078, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.360583747076463, "step": 432000 }, { "epoch": 2.605190482507447, "grad_norm": 2.2861459255218506, "learning_rate": 0.0002093374525157471, "loss": 5.288, "step": 432050 }, { "epoch": 2.605491974288781, "grad_norm": 3.2650046348571777, "learning_rate": 0.00020931876286497044, "loss": 5.1968, "step": 432100 }, { "epoch": 2.605793466070115, "grad_norm": 1.7541558742523193, "learning_rate": 0.00020930007212251666, "loss": 4.8831, "step": 432150 }, { "epoch": 2.606094957851449, "grad_norm": 0.9393763542175293, "learning_rate": 0.00020928138028872962, "loss": 4.9023, "step": 432200 }, { "epoch": 2.606396449632783, "grad_norm": 2.223578929901123, "learning_rate": 0.00020926268736395342, "loss": 5.2651, "step": 432250 }, { "epoch": 2.606697941414117, "grad_norm": 2.097384214401245, "learning_rate": 0.00020924399334853203, "loss": 5.2175, "step": 432300 }, { "epoch": 2.606999433195451, "grad_norm": 3.3701059818267822, "learning_rate": 0.00020922529824280947, "loss": 5.1438, "step": 432350 }, { "epoch": 2.6073009249767853, "grad_norm": 1.760046124458313, "learning_rate": 0.00020920660204712982, "loss": 5.3296, "step": 432400 }, { "epoch": 2.607602416758119, "grad_norm": 3.0383071899414062, "learning_rate": 0.00020918790476183715, "loss": 5.1989, "step": 432450 }, { "epoch": 2.6079039085394533, "grad_norm": 2.6272122859954834, "learning_rate": 0.00020916920638727556, "loss": 4.825, "step": 432500 }, { "epoch": 2.608205400320787, "grad_norm": 2.2375571727752686, "learning_rate": 0.00020915050692378915, "loss": 5.0486, "step": 432550 }, { "epoch": 2.6085068921021213, "grad_norm": 1.6614986658096313, "learning_rate": 0.0002091318063717221, "loss": 5.1555, "step": 432600 }, { "epoch": 2.6088083838834555, "grad_norm": 1.748497486114502, "learning_rate": 0.00020911310473141853, "loss": 4.5829, "step": 432650 }, { "epoch": 2.6091098756647892, "grad_norm": 1.8807804584503174, "learning_rate": 0.0002090944020032226, "loss": 5.4033, "step": 432700 }, { "epoch": 2.6094113674461235, "grad_norm": 1.700453758239746, "learning_rate": 0.00020907569818747853, "loss": 5.2127, "step": 432750 }, { "epoch": 2.609712859227457, "grad_norm": 2.2429285049438477, "learning_rate": 0.00020905699328453058, "loss": 4.8887, "step": 432800 }, { "epoch": 2.6100143510087914, "grad_norm": 1.4461594820022583, "learning_rate": 0.00020903828729472296, "loss": 4.7748, "step": 432850 }, { "epoch": 2.6103158427901256, "grad_norm": 1.645449161529541, "learning_rate": 0.00020901958021839987, "loss": 4.881, "step": 432900 }, { "epoch": 2.61061733457146, "grad_norm": 2.4435620307922363, "learning_rate": 0.00020900087205590563, "loss": 4.5329, "step": 432950 }, { "epoch": 2.6109188263527936, "grad_norm": 2.153395891189575, "learning_rate": 0.00020898216280758458, "loss": 4.5496, "step": 433000 }, { "epoch": 2.611220318134128, "grad_norm": 1.100813865661621, "learning_rate": 0.000208963452473781, "loss": 4.8459, "step": 433050 }, { "epoch": 2.6115218099154616, "grad_norm": 2.4159796237945557, "learning_rate": 0.00020894474105483918, "loss": 5.1228, "step": 433100 }, { "epoch": 2.6118233016967958, "grad_norm": 0.8893303871154785, "learning_rate": 0.00020892602855110355, "loss": 5.2362, "step": 433150 }, { "epoch": 2.61212479347813, "grad_norm": 1.873366355895996, "learning_rate": 0.00020890731496291841, "loss": 4.4452, "step": 433200 }, { "epoch": 2.6124262852594637, "grad_norm": 1.5206475257873535, "learning_rate": 0.00020888860029062825, "loss": 5.2049, "step": 433250 }, { "epoch": 2.612727777040798, "grad_norm": 1.417188048362732, "learning_rate": 0.00020886988453457743, "loss": 5.2932, "step": 433300 }, { "epoch": 2.6130292688221317, "grad_norm": 0.6361641883850098, "learning_rate": 0.00020885116769511037, "loss": 4.8038, "step": 433350 }, { "epoch": 2.613330760603466, "grad_norm": 3.2825655937194824, "learning_rate": 0.00020883244977257156, "loss": 4.625, "step": 433400 }, { "epoch": 2.6136322523848, "grad_norm": 1.4553327560424805, "learning_rate": 0.00020881373076730555, "loss": 5.0211, "step": 433450 }, { "epoch": 2.613933744166134, "grad_norm": 1.2398653030395508, "learning_rate": 0.00020879501067965666, "loss": 4.9262, "step": 433500 }, { "epoch": 2.614235235947468, "grad_norm": 1.139600396156311, "learning_rate": 0.00020877628950996955, "loss": 5.3272, "step": 433550 }, { "epoch": 2.614536727728802, "grad_norm": 2.075014352798462, "learning_rate": 0.0002087575672585887, "loss": 4.9282, "step": 433600 }, { "epoch": 2.614838219510136, "grad_norm": 1.9019057750701904, "learning_rate": 0.00020873884392585865, "loss": 5.0926, "step": 433650 }, { "epoch": 2.6151397112914703, "grad_norm": 3.2111706733703613, "learning_rate": 0.00020872011951212404, "loss": 4.9031, "step": 433700 }, { "epoch": 2.6154412030728045, "grad_norm": 1.6960424184799194, "learning_rate": 0.00020870139401772943, "loss": 4.7496, "step": 433750 }, { "epoch": 2.6157426948541382, "grad_norm": 1.7110929489135742, "learning_rate": 0.00020868266744301938, "loss": 4.9441, "step": 433800 }, { "epoch": 2.6160441866354724, "grad_norm": 2.4193389415740967, "learning_rate": 0.00020866393978833864, "loss": 4.7912, "step": 433850 }, { "epoch": 2.616345678416806, "grad_norm": 1.544533371925354, "learning_rate": 0.00020864521105403181, "loss": 5.2314, "step": 433900 }, { "epoch": 2.6166471701981404, "grad_norm": 1.353302001953125, "learning_rate": 0.0002086264812404435, "loss": 5.0442, "step": 433950 }, { "epoch": 2.6169486619794746, "grad_norm": 6.024861812591553, "learning_rate": 0.00020860775034791851, "loss": 4.6456, "step": 434000 }, { "epoch": 2.6172501537608084, "grad_norm": 2.4069414138793945, "learning_rate": 0.00020858901837680154, "loss": 5.0557, "step": 434050 }, { "epoch": 2.6175516455421426, "grad_norm": 1.828302025794983, "learning_rate": 0.00020857028532743723, "loss": 5.084, "step": 434100 }, { "epoch": 2.6178531373234764, "grad_norm": 1.7828880548477173, "learning_rate": 0.00020855155120017043, "loss": 5.3094, "step": 434150 }, { "epoch": 2.6181546291048106, "grad_norm": 1.7308353185653687, "learning_rate": 0.0002085328159953459, "loss": 4.9117, "step": 434200 }, { "epoch": 2.6184561208861448, "grad_norm": 2.4066383838653564, "learning_rate": 0.00020851407971330844, "loss": 4.957, "step": 434250 }, { "epoch": 2.6187576126674785, "grad_norm": 1.5799798965454102, "learning_rate": 0.00020849534235440283, "loss": 4.9974, "step": 434300 }, { "epoch": 2.6190591044488127, "grad_norm": 1.624779224395752, "learning_rate": 0.0002084766039189739, "loss": 4.9696, "step": 434350 }, { "epoch": 2.6193605962301465, "grad_norm": 1.2158584594726562, "learning_rate": 0.00020845786440736654, "loss": 4.9903, "step": 434400 }, { "epoch": 2.6196620880114807, "grad_norm": 1.7384653091430664, "learning_rate": 0.0002084391238199256, "loss": 4.9323, "step": 434450 }, { "epoch": 2.619963579792815, "grad_norm": 1.4097511768341064, "learning_rate": 0.000208420382156996, "loss": 5.0134, "step": 434500 }, { "epoch": 2.620265071574149, "grad_norm": 2.144205331802368, "learning_rate": 0.00020840163941892263, "loss": 5.2259, "step": 434550 }, { "epoch": 2.620566563355483, "grad_norm": 2.6244373321533203, "learning_rate": 0.0002083828956060504, "loss": 4.79, "step": 434600 }, { "epoch": 2.620868055136817, "grad_norm": 1.3598973751068115, "learning_rate": 0.00020836415071872434, "loss": 5.0553, "step": 434650 }, { "epoch": 2.621169546918151, "grad_norm": 2.2391724586486816, "learning_rate": 0.00020834540475728937, "loss": 5.2399, "step": 434700 }, { "epoch": 2.621471038699485, "grad_norm": 2.2870688438415527, "learning_rate": 0.0002083266577220905, "loss": 4.927, "step": 434750 }, { "epoch": 2.6217725304808193, "grad_norm": 1.7174073457717896, "learning_rate": 0.00020830790961347267, "loss": 4.953, "step": 434800 }, { "epoch": 2.622074022262153, "grad_norm": 1.4166580438613892, "learning_rate": 0.00020828916043178104, "loss": 4.8832, "step": 434850 }, { "epoch": 2.6223755140434872, "grad_norm": 1.4358099699020386, "learning_rate": 0.00020827041017736057, "loss": 4.476, "step": 434900 }, { "epoch": 2.622677005824821, "grad_norm": 1.8181419372558594, "learning_rate": 0.00020825165885055634, "loss": 4.7953, "step": 434950 }, { "epoch": 2.622978497606155, "grad_norm": 2.0605037212371826, "learning_rate": 0.00020823290645171347, "loss": 5.1411, "step": 435000 }, { "epoch": 2.622978497606155, "eval_loss": 5.413544178009033, "eval_runtime": 39.2429, "eval_samples_per_second": 13.047, "eval_steps_per_second": 6.523, "eval_tts_loss": 7.452078957205646, "step": 435000 }, { "epoch": 2.6232799893874894, "grad_norm": 2.040555238723755, "learning_rate": 0.00020821415298117712, "loss": 5.3951, "step": 435050 }, { "epoch": 2.6235814811688236, "grad_norm": 1.8736711740493774, "learning_rate": 0.00020819539843929227, "loss": 4.7225, "step": 435100 }, { "epoch": 2.6238829729501574, "grad_norm": 2.0353593826293945, "learning_rate": 0.0002081766428264042, "loss": 5.2561, "step": 435150 }, { "epoch": 2.6241844647314916, "grad_norm": 0.966487467288971, "learning_rate": 0.000208157886142858, "loss": 4.996, "step": 435200 }, { "epoch": 2.6244859565128253, "grad_norm": 2.207716464996338, "learning_rate": 0.00020813912838899896, "loss": 5.2829, "step": 435250 }, { "epoch": 2.6247874482941596, "grad_norm": 2.0088531970977783, "learning_rate": 0.0002081203695651722, "loss": 4.9321, "step": 435300 }, { "epoch": 2.6250889400754938, "grad_norm": 1.3959473371505737, "learning_rate": 0.00020810160967172302, "loss": 4.7156, "step": 435350 }, { "epoch": 2.6253904318568275, "grad_norm": 2.122359275817871, "learning_rate": 0.00020808284870899653, "loss": 5.1594, "step": 435400 }, { "epoch": 2.6256919236381617, "grad_norm": 2.545201539993286, "learning_rate": 0.0002080640866773382, "loss": 5.3054, "step": 435450 }, { "epoch": 2.6259934154194955, "grad_norm": 1.4627712965011597, "learning_rate": 0.0002080453235770932, "loss": 5.1414, "step": 435500 }, { "epoch": 2.6262949072008297, "grad_norm": 1.748610258102417, "learning_rate": 0.00020802655940860683, "loss": 4.9873, "step": 435550 }, { "epoch": 2.626596398982164, "grad_norm": 2.4231655597686768, "learning_rate": 0.00020800779417222445, "loss": 5.2974, "step": 435600 }, { "epoch": 2.6268978907634977, "grad_norm": 1.6743963956832886, "learning_rate": 0.00020798902786829142, "loss": 4.9479, "step": 435650 }, { "epoch": 2.627199382544832, "grad_norm": 1.477051854133606, "learning_rate": 0.00020797026049715304, "loss": 4.6842, "step": 435700 }, { "epoch": 2.6275008743261656, "grad_norm": 1.626962661743164, "learning_rate": 0.0002079514920591548, "loss": 5.0504, "step": 435750 }, { "epoch": 2.6278023661075, "grad_norm": 3.175668716430664, "learning_rate": 0.000207932722554642, "loss": 5.1772, "step": 435800 }, { "epoch": 2.628103857888834, "grad_norm": 2.448606252670288, "learning_rate": 0.00020791395198396014, "loss": 5.0295, "step": 435850 }, { "epoch": 2.6284053496701683, "grad_norm": 1.3269016742706299, "learning_rate": 0.00020789518034745465, "loss": 4.712, "step": 435900 }, { "epoch": 2.628706841451502, "grad_norm": 1.7262400388717651, "learning_rate": 0.00020787640764547097, "loss": 5.2232, "step": 435950 }, { "epoch": 2.6290083332328362, "grad_norm": 1.9395537376403809, "learning_rate": 0.0002078576338783546, "loss": 4.6178, "step": 436000 }, { "epoch": 2.62930982501417, "grad_norm": 0.8662488460540771, "learning_rate": 0.00020783885904645106, "loss": 4.76, "step": 436050 }, { "epoch": 2.629611316795504, "grad_norm": 1.1233994960784912, "learning_rate": 0.00020782008315010589, "loss": 4.9376, "step": 436100 }, { "epoch": 2.6299128085768384, "grad_norm": 0.5171986818313599, "learning_rate": 0.0002078013061896645, "loss": 4.8604, "step": 436150 }, { "epoch": 2.630214300358172, "grad_norm": 2.5376498699188232, "learning_rate": 0.00020778252816547264, "loss": 4.8046, "step": 436200 }, { "epoch": 2.6305157921395064, "grad_norm": 2.5275232791900635, "learning_rate": 0.0002077637490778758, "loss": 5.0223, "step": 436250 }, { "epoch": 2.63081728392084, "grad_norm": 1.6206902265548706, "learning_rate": 0.00020774496892721957, "loss": 5.172, "step": 436300 }, { "epoch": 2.6311187757021743, "grad_norm": 2.5376386642456055, "learning_rate": 0.0002077261877138496, "loss": 5.0918, "step": 436350 }, { "epoch": 2.6314202674835085, "grad_norm": 0.8666787147521973, "learning_rate": 0.0002077074054381115, "loss": 4.9532, "step": 436400 }, { "epoch": 2.6317217592648423, "grad_norm": 2.053304433822632, "learning_rate": 0.00020768862210035103, "loss": 4.6549, "step": 436450 }, { "epoch": 2.6320232510461765, "grad_norm": 1.9525700807571411, "learning_rate": 0.00020766983770091374, "loss": 4.7747, "step": 436500 }, { "epoch": 2.6323247428275103, "grad_norm": 2.378406047821045, "learning_rate": 0.00020765105224014536, "loss": 4.8755, "step": 436550 }, { "epoch": 2.6326262346088445, "grad_norm": 2.0334086418151855, "learning_rate": 0.0002076322657183917, "loss": 4.8279, "step": 436600 }, { "epoch": 2.6329277263901787, "grad_norm": 1.5121841430664062, "learning_rate": 0.00020761347813599845, "loss": 5.1749, "step": 436650 }, { "epoch": 2.633229218171513, "grad_norm": 1.691491961479187, "learning_rate": 0.00020759468949331126, "loss": 4.873, "step": 436700 }, { "epoch": 2.6335307099528467, "grad_norm": 1.877246618270874, "learning_rate": 0.00020757589979067607, "loss": 4.6503, "step": 436750 }, { "epoch": 2.633832201734181, "grad_norm": 1.6248620748519897, "learning_rate": 0.0002075571090284386, "loss": 4.7113, "step": 436800 }, { "epoch": 2.6341336935155146, "grad_norm": 2.115753412246704, "learning_rate": 0.00020753831720694466, "loss": 5.1494, "step": 436850 }, { "epoch": 2.634435185296849, "grad_norm": 2.076021432876587, "learning_rate": 0.00020751952432654014, "loss": 5.1019, "step": 436900 }, { "epoch": 2.634736677078183, "grad_norm": 2.3812689781188965, "learning_rate": 0.00020750073038757085, "loss": 4.8178, "step": 436950 }, { "epoch": 2.635038168859517, "grad_norm": 1.9511040449142456, "learning_rate": 0.0002074819353903826, "loss": 5.0861, "step": 437000 }, { "epoch": 2.635339660640851, "grad_norm": 1.5136700868606567, "learning_rate": 0.00020746313933532148, "loss": 5.2612, "step": 437050 }, { "epoch": 2.6356411524221848, "grad_norm": 2.315138101577759, "learning_rate": 0.00020744434222273324, "loss": 4.9498, "step": 437100 }, { "epoch": 2.635942644203519, "grad_norm": 1.6310371160507202, "learning_rate": 0.0002074255440529638, "loss": 4.9275, "step": 437150 }, { "epoch": 2.636244135984853, "grad_norm": 1.7384064197540283, "learning_rate": 0.0002074067448263592, "loss": 5.1603, "step": 437200 }, { "epoch": 2.636545627766187, "grad_norm": 2.101942300796509, "learning_rate": 0.00020738794454326545, "loss": 5.3989, "step": 437250 }, { "epoch": 2.636847119547521, "grad_norm": 2.4644954204559326, "learning_rate": 0.0002073691432040284, "loss": 4.8545, "step": 437300 }, { "epoch": 2.637148611328855, "grad_norm": 1.9715440273284912, "learning_rate": 0.00020735034080899414, "loss": 4.9173, "step": 437350 }, { "epoch": 2.637450103110189, "grad_norm": 1.5183241367340088, "learning_rate": 0.00020733153735850876, "loss": 4.5849, "step": 437400 }, { "epoch": 2.6377515948915233, "grad_norm": 1.4209235906600952, "learning_rate": 0.00020731273285291824, "loss": 4.6941, "step": 437450 }, { "epoch": 2.6380530866728575, "grad_norm": 1.4371623992919922, "learning_rate": 0.00020729392729256863, "loss": 5.025, "step": 437500 }, { "epoch": 2.6383545784541913, "grad_norm": 1.9862120151519775, "learning_rate": 0.00020727512067780606, "loss": 5.3691, "step": 437550 }, { "epoch": 2.6386560702355255, "grad_norm": 2.009369373321533, "learning_rate": 0.0002072563130089766, "loss": 5.2377, "step": 437600 }, { "epoch": 2.6389575620168593, "grad_norm": 2.0585856437683105, "learning_rate": 0.00020723750428642648, "loss": 4.7877, "step": 437650 }, { "epoch": 2.6392590537981935, "grad_norm": 2.4469103813171387, "learning_rate": 0.00020721869451050174, "loss": 4.9301, "step": 437700 }, { "epoch": 2.6395605455795277, "grad_norm": 2.045037031173706, "learning_rate": 0.00020719988368154854, "loss": 5.1646, "step": 437750 }, { "epoch": 2.6398620373608614, "grad_norm": 1.379736065864563, "learning_rate": 0.00020718107179991318, "loss": 4.66, "step": 437800 }, { "epoch": 2.6401635291421957, "grad_norm": 0.5246308445930481, "learning_rate": 0.00020716225886594178, "loss": 4.2869, "step": 437850 }, { "epoch": 2.6404650209235294, "grad_norm": 2.3156678676605225, "learning_rate": 0.00020714344487998056, "loss": 4.7536, "step": 437900 }, { "epoch": 2.6407665127048636, "grad_norm": 1.260223627090454, "learning_rate": 0.00020712462984237576, "loss": 5.0997, "step": 437950 }, { "epoch": 2.641068004486198, "grad_norm": 1.623124122619629, "learning_rate": 0.00020710581375347372, "loss": 5.1729, "step": 438000 }, { "epoch": 2.641068004486198, "eval_loss": 5.421302795410156, "eval_runtime": 39.1481, "eval_samples_per_second": 13.079, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.407740970202304, "step": 438000 }, { "epoch": 2.6413694962675316, "grad_norm": 2.8143558502197266, "learning_rate": 0.00020708699661362067, "loss": 5.0457, "step": 438050 }, { "epoch": 2.641670988048866, "grad_norm": 1.2756474018096924, "learning_rate": 0.00020706817842316289, "loss": 5.04, "step": 438100 }, { "epoch": 2.6419724798301996, "grad_norm": 1.8730255365371704, "learning_rate": 0.0002070493591824467, "loss": 4.9364, "step": 438150 }, { "epoch": 2.6422739716115338, "grad_norm": 1.5092415809631348, "learning_rate": 0.00020703053889181849, "loss": 5.1606, "step": 438200 }, { "epoch": 2.642575463392868, "grad_norm": 1.0843733549118042, "learning_rate": 0.00020701171755162465, "loss": 4.5234, "step": 438250 }, { "epoch": 2.642876955174202, "grad_norm": 1.4480210542678833, "learning_rate": 0.00020699289516221145, "loss": 5.0243, "step": 438300 }, { "epoch": 2.643178446955536, "grad_norm": 2.3378357887268066, "learning_rate": 0.00020697407172392537, "loss": 5.1273, "step": 438350 }, { "epoch": 2.64347993873687, "grad_norm": 1.475003957748413, "learning_rate": 0.0002069552472371128, "loss": 5.1491, "step": 438400 }, { "epoch": 2.643781430518204, "grad_norm": 1.846811056137085, "learning_rate": 0.0002069364217021202, "loss": 4.9199, "step": 438450 }, { "epoch": 2.644082922299538, "grad_norm": 1.4832494258880615, "learning_rate": 0.000206917595119294, "loss": 4.9611, "step": 438500 }, { "epoch": 2.6443844140808723, "grad_norm": 1.6905028820037842, "learning_rate": 0.00020689876748898066, "loss": 4.8575, "step": 438550 }, { "epoch": 2.644685905862206, "grad_norm": 1.9310983419418335, "learning_rate": 0.00020687993881152674, "loss": 4.8531, "step": 438600 }, { "epoch": 2.6449873976435403, "grad_norm": 1.7790565490722656, "learning_rate": 0.00020686110908727873, "loss": 5.128, "step": 438650 }, { "epoch": 2.645288889424874, "grad_norm": 0.5224432945251465, "learning_rate": 0.0002068422783165831, "loss": 5.0013, "step": 438700 }, { "epoch": 2.6455903812062083, "grad_norm": 1.724401593208313, "learning_rate": 0.0002068234464997865, "loss": 5.089, "step": 438750 }, { "epoch": 2.6458918729875425, "grad_norm": 1.2562521696090698, "learning_rate": 0.00020680461363723545, "loss": 5.5035, "step": 438800 }, { "epoch": 2.6461933647688767, "grad_norm": 1.4221951961517334, "learning_rate": 0.00020678577972927658, "loss": 5.1441, "step": 438850 }, { "epoch": 2.6464948565502104, "grad_norm": 2.234593391418457, "learning_rate": 0.0002067669447762564, "loss": 4.8884, "step": 438900 }, { "epoch": 2.6467963483315446, "grad_norm": 2.816098928451538, "learning_rate": 0.00020674810877852162, "loss": 4.8328, "step": 438950 }, { "epoch": 2.6470978401128784, "grad_norm": 2.2633237838745117, "learning_rate": 0.00020672927173641894, "loss": 4.7105, "step": 439000 }, { "epoch": 2.6473993318942126, "grad_norm": 1.8689980506896973, "learning_rate": 0.00020671043365029493, "loss": 5.0933, "step": 439050 }, { "epoch": 2.647700823675547, "grad_norm": 1.7612513303756714, "learning_rate": 0.00020669159452049634, "loss": 5.1808, "step": 439100 }, { "epoch": 2.6480023154568806, "grad_norm": 2.02314829826355, "learning_rate": 0.00020667275434736986, "loss": 5.2779, "step": 439150 }, { "epoch": 2.648303807238215, "grad_norm": 1.7003791332244873, "learning_rate": 0.00020665391313126217, "loss": 4.9229, "step": 439200 }, { "epoch": 2.6486052990195486, "grad_norm": 2.68503737449646, "learning_rate": 0.00020663507087252008, "loss": 5.1216, "step": 439250 }, { "epoch": 2.6489067908008828, "grad_norm": 1.9271464347839355, "learning_rate": 0.00020661622757149037, "loss": 4.9034, "step": 439300 }, { "epoch": 2.649208282582217, "grad_norm": 2.2735376358032227, "learning_rate": 0.0002065973832285197, "loss": 5.0086, "step": 439350 }, { "epoch": 2.6495097743635507, "grad_norm": 1.3906512260437012, "learning_rate": 0.00020657853784395504, "loss": 4.5096, "step": 439400 }, { "epoch": 2.649811266144885, "grad_norm": 1.4948866367340088, "learning_rate": 0.00020655969141814316, "loss": 5.0163, "step": 439450 }, { "epoch": 2.6501127579262187, "grad_norm": 1.637844204902649, "learning_rate": 0.00020654084395143078, "loss": 5.2593, "step": 439500 }, { "epoch": 2.650414249707553, "grad_norm": 1.3856858015060425, "learning_rate": 0.00020652199544416493, "loss": 4.8274, "step": 439550 }, { "epoch": 2.650715741488887, "grad_norm": 2.621222496032715, "learning_rate": 0.00020650314589669238, "loss": 4.8549, "step": 439600 }, { "epoch": 2.6510172332702213, "grad_norm": 1.410722255706787, "learning_rate": 0.00020648429530936005, "loss": 5.2658, "step": 439650 }, { "epoch": 2.651318725051555, "grad_norm": 1.7718340158462524, "learning_rate": 0.00020646544368251493, "loss": 4.7622, "step": 439700 }, { "epoch": 2.6516202168328893, "grad_norm": 2.3199262619018555, "learning_rate": 0.0002064465910165038, "loss": 5.0853, "step": 439750 }, { "epoch": 2.651921708614223, "grad_norm": 1.9683341979980469, "learning_rate": 0.00020642773731167379, "loss": 5.0774, "step": 439800 }, { "epoch": 2.6522232003955573, "grad_norm": 1.7358086109161377, "learning_rate": 0.0002064088825683718, "loss": 4.6757, "step": 439850 }, { "epoch": 2.6525246921768915, "grad_norm": 1.7782351970672607, "learning_rate": 0.0002063900267869448, "loss": 5.1175, "step": 439900 }, { "epoch": 2.6528261839582252, "grad_norm": 1.8443175554275513, "learning_rate": 0.00020637116996773987, "loss": 4.7071, "step": 439950 }, { "epoch": 2.6531276757395594, "grad_norm": 1.3502905368804932, "learning_rate": 0.00020635231211110395, "loss": 4.8091, "step": 440000 }, { "epoch": 2.653429167520893, "grad_norm": 1.4978843927383423, "learning_rate": 0.00020633345321738417, "loss": 5.5682, "step": 440050 }, { "epoch": 2.6537306593022274, "grad_norm": 2.3619062900543213, "learning_rate": 0.0002063145932869276, "loss": 5.4215, "step": 440100 }, { "epoch": 2.6540321510835616, "grad_norm": 1.6098183393478394, "learning_rate": 0.0002062957323200813, "loss": 5.0836, "step": 440150 }, { "epoch": 2.6543336428648954, "grad_norm": 0.4866611957550049, "learning_rate": 0.00020627687031719234, "loss": 4.8475, "step": 440200 }, { "epoch": 2.6546351346462296, "grad_norm": 1.485914707183838, "learning_rate": 0.00020625800727860796, "loss": 5.1974, "step": 440250 }, { "epoch": 2.6549366264275633, "grad_norm": 1.5539668798446655, "learning_rate": 0.00020623914320467518, "loss": 4.8658, "step": 440300 }, { "epoch": 2.6552381182088975, "grad_norm": 1.3107876777648926, "learning_rate": 0.00020622027809574122, "loss": 5.1761, "step": 440350 }, { "epoch": 2.6555396099902318, "grad_norm": 1.7061867713928223, "learning_rate": 0.0002062014119521533, "loss": 5.1805, "step": 440400 }, { "epoch": 2.655841101771566, "grad_norm": 1.0166499614715576, "learning_rate": 0.00020618254477425865, "loss": 4.7558, "step": 440450 }, { "epoch": 2.6561425935528997, "grad_norm": 1.8664777278900146, "learning_rate": 0.00020616367656240434, "loss": 5.2372, "step": 440500 }, { "epoch": 2.656444085334234, "grad_norm": 1.8719890117645264, "learning_rate": 0.00020614480731693779, "loss": 5.1317, "step": 440550 }, { "epoch": 2.6567455771155677, "grad_norm": 1.5341250896453857, "learning_rate": 0.00020612593703820612, "loss": 4.718, "step": 440600 }, { "epoch": 2.657047068896902, "grad_norm": 1.2110400199890137, "learning_rate": 0.00020610706572655674, "loss": 5.1585, "step": 440650 }, { "epoch": 2.657348560678236, "grad_norm": 1.9953607320785522, "learning_rate": 0.00020608819338233686, "loss": 5.0524, "step": 440700 }, { "epoch": 2.65765005245957, "grad_norm": 1.4692119359970093, "learning_rate": 0.00020606932000589382, "loss": 5.0006, "step": 440750 }, { "epoch": 2.657951544240904, "grad_norm": 2.6103761196136475, "learning_rate": 0.00020605044559757492, "loss": 5.1632, "step": 440800 }, { "epoch": 2.658253036022238, "grad_norm": 1.6560620069503784, "learning_rate": 0.00020603157015772764, "loss": 5.255, "step": 440850 }, { "epoch": 2.658554527803572, "grad_norm": 1.6236096620559692, "learning_rate": 0.00020601269368669925, "loss": 5.0673, "step": 440900 }, { "epoch": 2.6588560195849062, "grad_norm": 1.6683170795440674, "learning_rate": 0.00020599381618483707, "loss": 4.9203, "step": 440950 }, { "epoch": 2.65915751136624, "grad_norm": 1.4719107151031494, "learning_rate": 0.00020597493765248868, "loss": 5.1997, "step": 441000 }, { "epoch": 2.65915751136624, "eval_loss": 5.431072235107422, "eval_runtime": 39.0366, "eval_samples_per_second": 13.116, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.377847815757062, "step": 441000 }, { "epoch": 2.659459003147574, "grad_norm": 1.7286090850830078, "learning_rate": 0.0002059560580900015, "loss": 5.2549, "step": 441050 }, { "epoch": 2.659760494928908, "grad_norm": 1.6187355518341064, "learning_rate": 0.00020593717749772285, "loss": 5.3772, "step": 441100 }, { "epoch": 2.660061986710242, "grad_norm": 1.4400075674057007, "learning_rate": 0.0002059182958760003, "loss": 4.2992, "step": 441150 }, { "epoch": 2.6603634784915764, "grad_norm": 1.8349392414093018, "learning_rate": 0.00020589941322518132, "loss": 5.1771, "step": 441200 }, { "epoch": 2.6606649702729106, "grad_norm": 2.5374653339385986, "learning_rate": 0.0002058805295456134, "loss": 4.8357, "step": 441250 }, { "epoch": 2.6609664620542444, "grad_norm": 1.2808914184570312, "learning_rate": 0.00020586164483764408, "loss": 5.2021, "step": 441300 }, { "epoch": 2.6612679538355786, "grad_norm": 2.898959159851074, "learning_rate": 0.00020584275910162086, "loss": 4.8308, "step": 441350 }, { "epoch": 2.6615694456169123, "grad_norm": 1.14915132522583, "learning_rate": 0.00020582387233789143, "loss": 5.055, "step": 441400 }, { "epoch": 2.6618709373982465, "grad_norm": 1.754184603691101, "learning_rate": 0.00020580498454680332, "loss": 4.9202, "step": 441450 }, { "epoch": 2.6621724291795807, "grad_norm": 3.0785627365112305, "learning_rate": 0.000205786095728704, "loss": 4.5466, "step": 441500 }, { "epoch": 2.6624739209609145, "grad_norm": 1.2602195739746094, "learning_rate": 0.00020576720588394125, "loss": 4.7202, "step": 441550 }, { "epoch": 2.6627754127422487, "grad_norm": 2.730139970779419, "learning_rate": 0.00020574831501286267, "loss": 5.2271, "step": 441600 }, { "epoch": 2.6630769045235825, "grad_norm": 2.266667604446411, "learning_rate": 0.00020572942311581588, "loss": 4.7588, "step": 441650 }, { "epoch": 2.6633783963049167, "grad_norm": 1.9081997871398926, "learning_rate": 0.00020571053019314862, "loss": 4.9857, "step": 441700 }, { "epoch": 2.663679888086251, "grad_norm": 3.824484348297119, "learning_rate": 0.00020569163624520857, "loss": 4.9403, "step": 441750 }, { "epoch": 2.6639813798675847, "grad_norm": 2.7153871059417725, "learning_rate": 0.0002056727412723434, "loss": 5.3524, "step": 441800 }, { "epoch": 2.664282871648919, "grad_norm": 1.4515808820724487, "learning_rate": 0.00020565384527490089, "loss": 5.1066, "step": 441850 }, { "epoch": 2.6645843634302526, "grad_norm": 1.7689220905303955, "learning_rate": 0.00020563494825322874, "loss": 4.8194, "step": 441900 }, { "epoch": 2.664885855211587, "grad_norm": 3.590125799179077, "learning_rate": 0.0002056160502076748, "loss": 5.3164, "step": 441950 }, { "epoch": 2.665187346992921, "grad_norm": 1.806290626525879, "learning_rate": 0.00020559715113858682, "loss": 4.7873, "step": 442000 }, { "epoch": 2.6654888387742552, "grad_norm": 2.1659083366394043, "learning_rate": 0.00020557825104631266, "loss": 4.719, "step": 442050 }, { "epoch": 2.665790330555589, "grad_norm": 1.4914212226867676, "learning_rate": 0.0002055593499312, "loss": 5.2345, "step": 442100 }, { "epoch": 2.666091822336923, "grad_norm": 1.4541923999786377, "learning_rate": 0.00020554044779359687, "loss": 4.7971, "step": 442150 }, { "epoch": 2.666393314118257, "grad_norm": 2.29225492477417, "learning_rate": 0.00020552154463385103, "loss": 5.2745, "step": 442200 }, { "epoch": 2.666694805899591, "grad_norm": 1.5002127885818481, "learning_rate": 0.0002055026404523104, "loss": 4.7283, "step": 442250 }, { "epoch": 2.6669962976809254, "grad_norm": 0.4798014760017395, "learning_rate": 0.0002054837352493229, "loss": 4.9076, "step": 442300 }, { "epoch": 2.667297789462259, "grad_norm": 2.487337589263916, "learning_rate": 0.00020546482902523638, "loss": 4.838, "step": 442350 }, { "epoch": 2.6675992812435934, "grad_norm": 1.3185136318206787, "learning_rate": 0.00020544592178039887, "loss": 4.6106, "step": 442400 }, { "epoch": 2.667900773024927, "grad_norm": 1.5074098110198975, "learning_rate": 0.0002054270135151583, "loss": 4.9432, "step": 442450 }, { "epoch": 2.6682022648062613, "grad_norm": 1.8410011529922485, "learning_rate": 0.00020540810422986264, "loss": 4.5815, "step": 442500 }, { "epoch": 2.6685037565875955, "grad_norm": 1.4850258827209473, "learning_rate": 0.00020538919392485983, "loss": 4.8387, "step": 442550 }, { "epoch": 2.6688052483689297, "grad_norm": 1.9171204566955566, "learning_rate": 0.00020537028260049804, "loss": 5.1556, "step": 442600 }, { "epoch": 2.6691067401502635, "grad_norm": 1.6761391162872314, "learning_rate": 0.00020535137025712522, "loss": 4.9933, "step": 442650 }, { "epoch": 2.6694082319315977, "grad_norm": 1.5018337965011597, "learning_rate": 0.00020533245689508935, "loss": 5.0784, "step": 442700 }, { "epoch": 2.6697097237129315, "grad_norm": 2.0123064517974854, "learning_rate": 0.00020531354251473864, "loss": 4.7978, "step": 442750 }, { "epoch": 2.6700112154942657, "grad_norm": 1.6263973712921143, "learning_rate": 0.00020529462711642107, "loss": 5.1542, "step": 442800 }, { "epoch": 2.6703127072756, "grad_norm": 1.508156657218933, "learning_rate": 0.0002052757107004848, "loss": 4.7306, "step": 442850 }, { "epoch": 2.6706141990569336, "grad_norm": 1.7001501321792603, "learning_rate": 0.000205256793267278, "loss": 4.8543, "step": 442900 }, { "epoch": 2.670915690838268, "grad_norm": 3.4908645153045654, "learning_rate": 0.00020523787481714872, "loss": 5.1756, "step": 442950 }, { "epoch": 2.6712171826196016, "grad_norm": 1.6638734340667725, "learning_rate": 0.00020521895535044522, "loss": 4.9882, "step": 443000 }, { "epoch": 2.671518674400936, "grad_norm": 1.492790699005127, "learning_rate": 0.00020520003486751567, "loss": 4.978, "step": 443050 }, { "epoch": 2.67182016618227, "grad_norm": 2.8221096992492676, "learning_rate": 0.0002051811133687082, "loss": 4.8229, "step": 443100 }, { "epoch": 2.672121657963604, "grad_norm": 1.8752222061157227, "learning_rate": 0.00020516219085437112, "loss": 4.9527, "step": 443150 }, { "epoch": 2.672423149744938, "grad_norm": 1.3290542364120483, "learning_rate": 0.0002051432673248526, "loss": 4.9817, "step": 443200 }, { "epoch": 2.6727246415262718, "grad_norm": 0.7521706819534302, "learning_rate": 0.00020512434278050097, "loss": 5.0451, "step": 443250 }, { "epoch": 2.673026133307606, "grad_norm": 1.7619494199752808, "learning_rate": 0.0002051054172216645, "loss": 5.2969, "step": 443300 }, { "epoch": 2.67332762508894, "grad_norm": 1.379699945449829, "learning_rate": 0.00020508649064869143, "loss": 4.8706, "step": 443350 }, { "epoch": 2.6736291168702744, "grad_norm": 0.4581584334373474, "learning_rate": 0.00020506756306193014, "loss": 4.9524, "step": 443400 }, { "epoch": 2.673930608651608, "grad_norm": 2.1737470626831055, "learning_rate": 0.00020504863446172892, "loss": 4.7073, "step": 443450 }, { "epoch": 2.6742321004329423, "grad_norm": 1.4398455619812012, "learning_rate": 0.00020502970484843612, "loss": 5.0548, "step": 443500 }, { "epoch": 2.674533592214276, "grad_norm": 2.1643600463867188, "learning_rate": 0.00020501077422240013, "loss": 5.1163, "step": 443550 }, { "epoch": 2.6748350839956103, "grad_norm": 1.3310295343399048, "learning_rate": 0.00020499184258396938, "loss": 5.4516, "step": 443600 }, { "epoch": 2.6751365757769445, "grad_norm": 1.7541489601135254, "learning_rate": 0.00020497290993349229, "loss": 4.949, "step": 443650 }, { "epoch": 2.6754380675582783, "grad_norm": 2.4426560401916504, "learning_rate": 0.00020495397627131714, "loss": 4.8505, "step": 443700 }, { "epoch": 2.6757395593396125, "grad_norm": 1.5112850666046143, "learning_rate": 0.00020493504159779254, "loss": 5.1133, "step": 443750 }, { "epoch": 2.6760410511209463, "grad_norm": 1.3915029764175415, "learning_rate": 0.00020491610591326685, "loss": 5.1767, "step": 443800 }, { "epoch": 2.6763425429022805, "grad_norm": 1.29791259765625, "learning_rate": 0.00020489716921808865, "loss": 4.9066, "step": 443850 }, { "epoch": 2.6766440346836147, "grad_norm": 1.0276962518692017, "learning_rate": 0.0002048782315126064, "loss": 5.1019, "step": 443900 }, { "epoch": 2.6769455264649484, "grad_norm": 2.0142910480499268, "learning_rate": 0.00020485929279716856, "loss": 5.2755, "step": 443950 }, { "epoch": 2.6772470182462826, "grad_norm": 1.4759243726730347, "learning_rate": 0.0002048403530721238, "loss": 5.031, "step": 444000 }, { "epoch": 2.6772470182462826, "eval_loss": 5.427376747131348, "eval_runtime": 39.1649, "eval_samples_per_second": 13.073, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.417894826910515, "step": 444000 }, { "epoch": 2.6775485100276164, "grad_norm": 1.2170132398605347, "learning_rate": 0.00020482141233782057, "loss": 5.1203, "step": 444050 }, { "epoch": 2.6778500018089506, "grad_norm": 1.5896555185317993, "learning_rate": 0.00020480247059460744, "loss": 4.498, "step": 444100 }, { "epoch": 2.678151493590285, "grad_norm": 1.8991340398788452, "learning_rate": 0.00020478352784283308, "loss": 5.1661, "step": 444150 }, { "epoch": 2.678452985371619, "grad_norm": 1.664775013923645, "learning_rate": 0.00020476458408284607, "loss": 5.3048, "step": 444200 }, { "epoch": 2.678754477152953, "grad_norm": 1.8978996276855469, "learning_rate": 0.00020474563931499506, "loss": 4.8448, "step": 444250 }, { "epoch": 2.679055968934287, "grad_norm": 1.6925206184387207, "learning_rate": 0.0002047266935396286, "loss": 5.097, "step": 444300 }, { "epoch": 2.6793574607156208, "grad_norm": 2.56742787361145, "learning_rate": 0.00020470774675709552, "loss": 4.9626, "step": 444350 }, { "epoch": 2.679658952496955, "grad_norm": 2.1543567180633545, "learning_rate": 0.00020468879896774444, "loss": 5.2786, "step": 444400 }, { "epoch": 2.679960444278289, "grad_norm": 1.797709584236145, "learning_rate": 0.00020466985017192403, "loss": 5.4341, "step": 444450 }, { "epoch": 2.680261936059623, "grad_norm": 1.0887266397476196, "learning_rate": 0.00020465090036998307, "loss": 4.7149, "step": 444500 }, { "epoch": 2.680563427840957, "grad_norm": 1.635082483291626, "learning_rate": 0.00020463194956227027, "loss": 4.5708, "step": 444550 }, { "epoch": 2.680864919622291, "grad_norm": 1.498836874961853, "learning_rate": 0.0002046129977491344, "loss": 4.9453, "step": 444600 }, { "epoch": 2.681166411403625, "grad_norm": 2.867337226867676, "learning_rate": 0.00020459404493092424, "loss": 4.7474, "step": 444650 }, { "epoch": 2.6814679031849593, "grad_norm": 1.7167936563491821, "learning_rate": 0.00020457509110798856, "loss": 5.0878, "step": 444700 }, { "epoch": 2.681769394966293, "grad_norm": 1.6091290712356567, "learning_rate": 0.00020455613628067625, "loss": 5.4218, "step": 444750 }, { "epoch": 2.6820708867476273, "grad_norm": 1.360585331916809, "learning_rate": 0.00020453718044933613, "loss": 4.6982, "step": 444800 }, { "epoch": 2.682372378528961, "grad_norm": 3.009791135787964, "learning_rate": 0.000204518223614317, "loss": 4.7984, "step": 444850 }, { "epoch": 2.6826738703102952, "grad_norm": 1.8363901376724243, "learning_rate": 0.00020449926577596777, "loss": 5.0576, "step": 444900 }, { "epoch": 2.6829753620916295, "grad_norm": 1.7550568580627441, "learning_rate": 0.00020448030693463734, "loss": 5.0346, "step": 444950 }, { "epoch": 2.6832768538729637, "grad_norm": 2.1379551887512207, "learning_rate": 0.00020446134709067458, "loss": 4.7253, "step": 445000 }, { "epoch": 2.6835783456542974, "grad_norm": 1.7107871770858765, "learning_rate": 0.0002044423862444285, "loss": 4.8859, "step": 445050 }, { "epoch": 2.6838798374356316, "grad_norm": 2.112229108810425, "learning_rate": 0.00020442342439624795, "loss": 5.1032, "step": 445100 }, { "epoch": 2.6841813292169654, "grad_norm": 1.2373663187026978, "learning_rate": 0.0002044044615464819, "loss": 5.1721, "step": 445150 }, { "epoch": 2.6844828209982996, "grad_norm": 1.3873467445373535, "learning_rate": 0.0002043854976954795, "loss": 5.0883, "step": 445200 }, { "epoch": 2.684784312779634, "grad_norm": 1.6341044902801514, "learning_rate": 0.00020436653284358955, "loss": 4.8626, "step": 445250 }, { "epoch": 2.6850858045609676, "grad_norm": 1.8806787729263306, "learning_rate": 0.00020434756699116112, "loss": 4.7509, "step": 445300 }, { "epoch": 2.6853872963423018, "grad_norm": 2.8138554096221924, "learning_rate": 0.0002043286001385433, "loss": 5.2329, "step": 445350 }, { "epoch": 2.6856887881236355, "grad_norm": 1.5586529970169067, "learning_rate": 0.00020430963228608515, "loss": 4.5286, "step": 445400 }, { "epoch": 2.6859902799049697, "grad_norm": 1.9954493045806885, "learning_rate": 0.00020429066343413568, "loss": 4.833, "step": 445450 }, { "epoch": 2.686291771686304, "grad_norm": 1.4891936779022217, "learning_rate": 0.00020427169358304406, "loss": 4.8873, "step": 445500 }, { "epoch": 2.6865932634676377, "grad_norm": 2.0128095149993896, "learning_rate": 0.00020425272273315938, "loss": 4.5916, "step": 445550 }, { "epoch": 2.686894755248972, "grad_norm": 0.7817835807800293, "learning_rate": 0.00020423375088483075, "loss": 5.0064, "step": 445600 }, { "epoch": 2.6871962470303057, "grad_norm": 1.5391061305999756, "learning_rate": 0.00020421477803840734, "loss": 4.9205, "step": 445650 }, { "epoch": 2.68749773881164, "grad_norm": 2.2178494930267334, "learning_rate": 0.00020419580419423828, "loss": 5.0555, "step": 445700 }, { "epoch": 2.687799230592974, "grad_norm": 1.5508586168289185, "learning_rate": 0.00020417682935267275, "loss": 4.9803, "step": 445750 }, { "epoch": 2.6881007223743083, "grad_norm": 1.8528653383255005, "learning_rate": 0.00020415785351406008, "loss": 5.0266, "step": 445800 }, { "epoch": 2.688402214155642, "grad_norm": 1.751908302307129, "learning_rate": 0.00020413887667874937, "loss": 5.259, "step": 445850 }, { "epoch": 2.6887037059369763, "grad_norm": 1.9905540943145752, "learning_rate": 0.00020411989884708983, "loss": 4.981, "step": 445900 }, { "epoch": 2.68900519771831, "grad_norm": 0.5513720512390137, "learning_rate": 0.00020410092001943087, "loss": 4.5172, "step": 445950 }, { "epoch": 2.6893066894996442, "grad_norm": 2.390381336212158, "learning_rate": 0.00020408194019612164, "loss": 5.4036, "step": 446000 }, { "epoch": 2.6896081812809784, "grad_norm": 1.8756022453308105, "learning_rate": 0.00020406295937751151, "loss": 4.9166, "step": 446050 }, { "epoch": 2.689909673062312, "grad_norm": 1.3791128396987915, "learning_rate": 0.00020404397756394976, "loss": 4.9685, "step": 446100 }, { "epoch": 2.6902111648436464, "grad_norm": 1.4820120334625244, "learning_rate": 0.00020402499475578574, "loss": 4.9407, "step": 446150 }, { "epoch": 2.69051265662498, "grad_norm": 3.24166202545166, "learning_rate": 0.00020400601095336876, "loss": 5.0416, "step": 446200 }, { "epoch": 2.6908141484063144, "grad_norm": 1.7042206525802612, "learning_rate": 0.00020398702615704822, "loss": 5.0355, "step": 446250 }, { "epoch": 2.6911156401876486, "grad_norm": 1.4768624305725098, "learning_rate": 0.00020396804036717348, "loss": 4.9217, "step": 446300 }, { "epoch": 2.691417131968983, "grad_norm": 1.6977734565734863, "learning_rate": 0.00020394905358409398, "loss": 5.2924, "step": 446350 }, { "epoch": 2.6917186237503166, "grad_norm": 2.3423712253570557, "learning_rate": 0.0002039300658081592, "loss": 5.0202, "step": 446400 }, { "epoch": 2.6920201155316508, "grad_norm": 1.132155179977417, "learning_rate": 0.0002039110770397185, "loss": 4.844, "step": 446450 }, { "epoch": 2.6923216073129845, "grad_norm": 0.9833331108093262, "learning_rate": 0.00020389208727912133, "loss": 5.0091, "step": 446500 }, { "epoch": 2.6926230990943187, "grad_norm": 1.433908462524414, "learning_rate": 0.0002038730965267172, "loss": 4.689, "step": 446550 }, { "epoch": 2.692924590875653, "grad_norm": 1.5158636569976807, "learning_rate": 0.00020385410478285565, "loss": 5.2284, "step": 446600 }, { "epoch": 2.6932260826569867, "grad_norm": 0.991759717464447, "learning_rate": 0.00020383511204788616, "loss": 4.5795, "step": 446650 }, { "epoch": 2.693527574438321, "grad_norm": 1.5520384311676025, "learning_rate": 0.00020381611832215823, "loss": 4.5862, "step": 446700 }, { "epoch": 2.6938290662196547, "grad_norm": 1.8612253665924072, "learning_rate": 0.0002037971236060214, "loss": 4.8484, "step": 446750 }, { "epoch": 2.694130558000989, "grad_norm": 1.969244122505188, "learning_rate": 0.00020377812789982536, "loss": 4.8567, "step": 446800 }, { "epoch": 2.694432049782323, "grad_norm": 1.661899209022522, "learning_rate": 0.00020375913120391958, "loss": 4.771, "step": 446850 }, { "epoch": 2.694733541563657, "grad_norm": 1.4937608242034912, "learning_rate": 0.00020374013351865373, "loss": 5.0908, "step": 446900 }, { "epoch": 2.695035033344991, "grad_norm": 1.811357021331787, "learning_rate": 0.0002037211348443774, "loss": 5.2697, "step": 446950 }, { "epoch": 2.695336525126325, "grad_norm": 1.9100817441940308, "learning_rate": 0.00020370213518144026, "loss": 5.3681, "step": 447000 }, { "epoch": 2.695336525126325, "eval_loss": 5.399447917938232, "eval_runtime": 39.1203, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.382261116974472, "step": 447000 }, { "epoch": 2.695638016907659, "grad_norm": 2.563528537750244, "learning_rate": 0.00020368313453019194, "loss": 4.9396, "step": 447050 }, { "epoch": 2.6959395086889932, "grad_norm": 2.3642189502716064, "learning_rate": 0.00020366413289098215, "loss": 5.2935, "step": 447100 }, { "epoch": 2.6962410004703274, "grad_norm": 1.655181884765625, "learning_rate": 0.00020364513026416056, "loss": 5.0997, "step": 447150 }, { "epoch": 2.696542492251661, "grad_norm": 2.39983868598938, "learning_rate": 0.00020362612665007695, "loss": 5.3679, "step": 447200 }, { "epoch": 2.6968439840329954, "grad_norm": 2.229175090789795, "learning_rate": 0.00020360712204908096, "loss": 5.2076, "step": 447250 }, { "epoch": 2.697145475814329, "grad_norm": 2.3927102088928223, "learning_rate": 0.0002035881164615224, "loss": 5.0099, "step": 447300 }, { "epoch": 2.6974469675956634, "grad_norm": 1.4795362949371338, "learning_rate": 0.00020356910988775101, "loss": 4.4581, "step": 447350 }, { "epoch": 2.6977484593769976, "grad_norm": 1.6423618793487549, "learning_rate": 0.00020355010232811667, "loss": 5.2643, "step": 447400 }, { "epoch": 2.6980499511583313, "grad_norm": 1.9595777988433838, "learning_rate": 0.0002035310937829691, "loss": 4.9129, "step": 447450 }, { "epoch": 2.6983514429396656, "grad_norm": 1.432995080947876, "learning_rate": 0.00020351208425265806, "loss": 5.3839, "step": 447500 }, { "epoch": 2.6986529347209993, "grad_norm": 1.7826075553894043, "learning_rate": 0.0002034930737375335, "loss": 4.9809, "step": 447550 }, { "epoch": 2.6989544265023335, "grad_norm": 1.5512921810150146, "learning_rate": 0.0002034740622379453, "loss": 5.4488, "step": 447600 }, { "epoch": 2.6992559182836677, "grad_norm": 1.960693359375, "learning_rate": 0.00020345504975424332, "loss": 5.1537, "step": 447650 }, { "epoch": 2.6995574100650015, "grad_norm": 1.5546517372131348, "learning_rate": 0.00020343603628677738, "loss": 4.9178, "step": 447700 }, { "epoch": 2.6998589018463357, "grad_norm": 2.0684142112731934, "learning_rate": 0.00020341702183589746, "loss": 5.0513, "step": 447750 }, { "epoch": 2.7001603936276695, "grad_norm": 2.460618734359741, "learning_rate": 0.00020339800640195352, "loss": 5.1535, "step": 447800 }, { "epoch": 2.7004618854090037, "grad_norm": 2.3066585063934326, "learning_rate": 0.00020337898998529542, "loss": 5.2168, "step": 447850 }, { "epoch": 2.700763377190338, "grad_norm": 1.9654204845428467, "learning_rate": 0.00020335997258627323, "loss": 5.0163, "step": 447900 }, { "epoch": 2.701064868971672, "grad_norm": 1.7021867036819458, "learning_rate": 0.00020334095420523684, "loss": 5.2227, "step": 447950 }, { "epoch": 2.701366360753006, "grad_norm": 0.9535350203514099, "learning_rate": 0.00020332193484253635, "loss": 4.7949, "step": 448000 }, { "epoch": 2.70166785253434, "grad_norm": 1.6216481924057007, "learning_rate": 0.00020330291449852174, "loss": 4.6677, "step": 448050 }, { "epoch": 2.701969344315674, "grad_norm": 1.4377752542495728, "learning_rate": 0.00020328389317354306, "loss": 4.7232, "step": 448100 }, { "epoch": 2.702270836097008, "grad_norm": 1.9369498491287231, "learning_rate": 0.00020326487086795036, "loss": 5.3411, "step": 448150 }, { "epoch": 2.7025723278783422, "grad_norm": 0.7166310548782349, "learning_rate": 0.0002032458475820937, "loss": 4.6957, "step": 448200 }, { "epoch": 2.702873819659676, "grad_norm": 2.3356502056121826, "learning_rate": 0.00020322682331632325, "loss": 4.6732, "step": 448250 }, { "epoch": 2.70317531144101, "grad_norm": 0.8398468494415283, "learning_rate": 0.00020320779807098905, "loss": 4.4899, "step": 448300 }, { "epoch": 2.703476803222344, "grad_norm": 1.9439665079116821, "learning_rate": 0.00020318877184644123, "loss": 5.2505, "step": 448350 }, { "epoch": 2.703778295003678, "grad_norm": 1.5322273969650269, "learning_rate": 0.00020316974464303002, "loss": 5.1198, "step": 448400 }, { "epoch": 2.7040797867850124, "grad_norm": 0.7436027526855469, "learning_rate": 0.00020315071646110554, "loss": 5.2906, "step": 448450 }, { "epoch": 2.704381278566346, "grad_norm": 3.7264020442962646, "learning_rate": 0.0002031316873010179, "loss": 5.1551, "step": 448500 }, { "epoch": 2.7046827703476803, "grad_norm": 1.5429482460021973, "learning_rate": 0.00020311265716311744, "loss": 5.1336, "step": 448550 }, { "epoch": 2.704984262129014, "grad_norm": 1.7741881608963013, "learning_rate": 0.00020309362604775433, "loss": 5.1728, "step": 448600 }, { "epoch": 2.7052857539103483, "grad_norm": 1.598806381225586, "learning_rate": 0.00020307459395527876, "loss": 4.9061, "step": 448650 }, { "epoch": 2.7055872456916825, "grad_norm": 1.595421552658081, "learning_rate": 0.00020305556088604104, "loss": 4.7837, "step": 448700 }, { "epoch": 2.7058887374730167, "grad_norm": 1.5587713718414307, "learning_rate": 0.00020303652684039146, "loss": 5.0776, "step": 448750 }, { "epoch": 2.7061902292543505, "grad_norm": 0.9931032657623291, "learning_rate": 0.00020301749181868026, "loss": 5.004, "step": 448800 }, { "epoch": 2.7064917210356847, "grad_norm": 1.5233702659606934, "learning_rate": 0.0002029984558212578, "loss": 4.8599, "step": 448850 }, { "epoch": 2.7067932128170185, "grad_norm": 2.2886571884155273, "learning_rate": 0.0002029794188484744, "loss": 4.8573, "step": 448900 }, { "epoch": 2.7070947045983527, "grad_norm": 1.7428209781646729, "learning_rate": 0.00020296038090068038, "loss": 4.4412, "step": 448950 }, { "epoch": 2.707396196379687, "grad_norm": 1.6368857622146606, "learning_rate": 0.00020294134197822617, "loss": 5.2817, "step": 449000 }, { "epoch": 2.7076976881610206, "grad_norm": 1.9776685237884521, "learning_rate": 0.00020292230208146208, "loss": 5.0059, "step": 449050 }, { "epoch": 2.707999179942355, "grad_norm": 2.018645763397217, "learning_rate": 0.00020290326121073852, "loss": 5.1807, "step": 449100 }, { "epoch": 2.7083006717236886, "grad_norm": 1.6111478805541992, "learning_rate": 0.00020288421936640594, "loss": 5.0663, "step": 449150 }, { "epoch": 2.708602163505023, "grad_norm": 1.5224919319152832, "learning_rate": 0.0002028651765488148, "loss": 5.1966, "step": 449200 }, { "epoch": 2.708903655286357, "grad_norm": 0.794322669506073, "learning_rate": 0.00020284613275831553, "loss": 4.8174, "step": 449250 }, { "epoch": 2.7092051470676908, "grad_norm": 1.2861120700836182, "learning_rate": 0.0002028270879952586, "loss": 4.9432, "step": 449300 }, { "epoch": 2.709506638849025, "grad_norm": 1.6872316598892212, "learning_rate": 0.0002028080422599945, "loss": 5.3409, "step": 449350 }, { "epoch": 2.7098081306303587, "grad_norm": 3.136969566345215, "learning_rate": 0.00020278899555287375, "loss": 4.8624, "step": 449400 }, { "epoch": 2.710109622411693, "grad_norm": 0.846822202205658, "learning_rate": 0.00020276994787424686, "loss": 4.6845, "step": 449450 }, { "epoch": 2.710411114193027, "grad_norm": 2.309692621231079, "learning_rate": 0.0002027508992244644, "loss": 4.8417, "step": 449500 }, { "epoch": 2.7107126059743614, "grad_norm": 1.5854737758636475, "learning_rate": 0.00020273184960387688, "loss": 4.9023, "step": 449550 }, { "epoch": 2.711014097755695, "grad_norm": 1.8830386400222778, "learning_rate": 0.00020271279901283497, "loss": 4.7686, "step": 449600 }, { "epoch": 2.7113155895370293, "grad_norm": 1.020954966545105, "learning_rate": 0.0002026937474516892, "loss": 4.7873, "step": 449650 }, { "epoch": 2.711617081318363, "grad_norm": 1.4215914011001587, "learning_rate": 0.00020267469492079017, "loss": 5.317, "step": 449700 }, { "epoch": 2.7119185730996973, "grad_norm": 1.5140907764434814, "learning_rate": 0.0002026556414204886, "loss": 5.0925, "step": 449750 }, { "epoch": 2.7122200648810315, "grad_norm": 1.6449707746505737, "learning_rate": 0.0002026365869511351, "loss": 5.1934, "step": 449800 }, { "epoch": 2.7125215566623653, "grad_norm": 1.8891783952713013, "learning_rate": 0.0002026175315130803, "loss": 4.6729, "step": 449850 }, { "epoch": 2.7128230484436995, "grad_norm": 1.5268771648406982, "learning_rate": 0.00020259847510667494, "loss": 4.6967, "step": 449900 }, { "epoch": 2.7131245402250332, "grad_norm": 2.1093645095825195, "learning_rate": 0.0002025794177322697, "loss": 5.0021, "step": 449950 }, { "epoch": 2.7134260320063674, "grad_norm": 2.1537673473358154, "learning_rate": 0.0002025603593902153, "loss": 4.9838, "step": 450000 }, { "epoch": 2.7134260320063674, "eval_loss": 5.4107866287231445, "eval_runtime": 39.1518, "eval_samples_per_second": 13.077, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.424520262878847, "step": 450000 }, { "epoch": 2.7137275237877017, "grad_norm": 1.3166481256484985, "learning_rate": 0.00020254130008086251, "loss": 4.523, "step": 450050 }, { "epoch": 2.714029015569036, "grad_norm": 1.9917404651641846, "learning_rate": 0.00020252223980456204, "loss": 4.9844, "step": 450100 }, { "epoch": 2.7143305073503696, "grad_norm": 2.5771121978759766, "learning_rate": 0.00020250317856166473, "loss": 5.0964, "step": 450150 }, { "epoch": 2.714631999131704, "grad_norm": 1.8888702392578125, "learning_rate": 0.00020248411635252138, "loss": 4.9392, "step": 450200 }, { "epoch": 2.7149334909130376, "grad_norm": 1.6237105131149292, "learning_rate": 0.00020246505317748272, "loss": 5.1534, "step": 450250 }, { "epoch": 2.715234982694372, "grad_norm": 1.8021275997161865, "learning_rate": 0.00020244598903689965, "loss": 5.2155, "step": 450300 }, { "epoch": 2.715536474475706, "grad_norm": 3.056645631790161, "learning_rate": 0.00020242692393112298, "loss": 5.2219, "step": 450350 }, { "epoch": 2.7158379662570398, "grad_norm": 1.5878256559371948, "learning_rate": 0.0002024078578605036, "loss": 5.3398, "step": 450400 }, { "epoch": 2.716139458038374, "grad_norm": 1.376305341720581, "learning_rate": 0.00020238879082539236, "loss": 4.8432, "step": 450450 }, { "epoch": 2.7164409498197077, "grad_norm": 0.4938049614429474, "learning_rate": 0.0002023697228261402, "loss": 5.2534, "step": 450500 }, { "epoch": 2.716742441601042, "grad_norm": 2.3673501014709473, "learning_rate": 0.00020235065386309804, "loss": 4.9987, "step": 450550 }, { "epoch": 2.717043933382376, "grad_norm": 1.7512445449829102, "learning_rate": 0.00020233158393661682, "loss": 4.7601, "step": 450600 }, { "epoch": 2.71734542516371, "grad_norm": 2.02176570892334, "learning_rate": 0.00020231251304704748, "loss": 4.9035, "step": 450650 }, { "epoch": 2.717646916945044, "grad_norm": 2.3697214126586914, "learning_rate": 0.00020229344119474093, "loss": 4.9804, "step": 450700 }, { "epoch": 2.717948408726378, "grad_norm": 1.787876009941101, "learning_rate": 0.0002022743683800483, "loss": 5.0796, "step": 450750 }, { "epoch": 2.718249900507712, "grad_norm": 2.103313446044922, "learning_rate": 0.0002022552946033205, "loss": 4.9884, "step": 450800 }, { "epoch": 2.7185513922890463, "grad_norm": 1.4655359983444214, "learning_rate": 0.0002022362198649085, "loss": 5.0431, "step": 450850 }, { "epoch": 2.7188528840703805, "grad_norm": 1.6496294736862183, "learning_rate": 0.00020221714416516348, "loss": 4.7163, "step": 450900 }, { "epoch": 2.7191543758517143, "grad_norm": 1.9297517538070679, "learning_rate": 0.00020219806750443645, "loss": 5.1557, "step": 450950 }, { "epoch": 2.7194558676330485, "grad_norm": 1.6239851713180542, "learning_rate": 0.00020217898988307846, "loss": 5.0878, "step": 451000 }, { "epoch": 2.7197573594143822, "grad_norm": 3.605574607849121, "learning_rate": 0.00020215991130144063, "loss": 4.6973, "step": 451050 }, { "epoch": 2.7200588511957164, "grad_norm": 1.9148168563842773, "learning_rate": 0.00020214083175987407, "loss": 4.9178, "step": 451100 }, { "epoch": 2.7203603429770506, "grad_norm": 2.4731996059417725, "learning_rate": 0.0002021217512587299, "loss": 5.0518, "step": 451150 }, { "epoch": 2.7206618347583844, "grad_norm": 1.5646471977233887, "learning_rate": 0.00020210266979835928, "loss": 5.0236, "step": 451200 }, { "epoch": 2.7209633265397186, "grad_norm": 1.1627588272094727, "learning_rate": 0.0002020835873791134, "loss": 4.7494, "step": 451250 }, { "epoch": 2.7212648183210524, "grad_norm": 3.263101100921631, "learning_rate": 0.0002020645040013434, "loss": 4.6931, "step": 451300 }, { "epoch": 2.7215663101023866, "grad_norm": 1.5273617506027222, "learning_rate": 0.0002020454196654005, "loss": 5.0238, "step": 451350 }, { "epoch": 2.721867801883721, "grad_norm": 3.747098445892334, "learning_rate": 0.00020202633437163594, "loss": 4.811, "step": 451400 }, { "epoch": 2.7221692936650546, "grad_norm": 3.2498891353607178, "learning_rate": 0.00020200724812040097, "loss": 4.9798, "step": 451450 }, { "epoch": 2.7224707854463888, "grad_norm": 0.5183735489845276, "learning_rate": 0.00020198816091204677, "loss": 4.8743, "step": 451500 }, { "epoch": 2.7227722772277225, "grad_norm": 1.2917020320892334, "learning_rate": 0.00020196907274692464, "loss": 4.4283, "step": 451550 }, { "epoch": 2.7230737690090567, "grad_norm": 1.4565054178237915, "learning_rate": 0.00020194998362538596, "loss": 5.1509, "step": 451600 }, { "epoch": 2.723375260790391, "grad_norm": 2.1733250617980957, "learning_rate": 0.00020193089354778192, "loss": 5.2504, "step": 451650 }, { "epoch": 2.723676752571725, "grad_norm": 1.5994951725006104, "learning_rate": 0.00020191180251446388, "loss": 4.9334, "step": 451700 }, { "epoch": 2.723978244353059, "grad_norm": 1.4671926498413086, "learning_rate": 0.0002018927105257832, "loss": 5.1896, "step": 451750 }, { "epoch": 2.724279736134393, "grad_norm": 1.1218868494033813, "learning_rate": 0.00020187361758209133, "loss": 5.1697, "step": 451800 }, { "epoch": 2.724581227915727, "grad_norm": 2.6666057109832764, "learning_rate": 0.00020185452368373942, "loss": 5.1764, "step": 451850 }, { "epoch": 2.724882719697061, "grad_norm": 0.7089901566505432, "learning_rate": 0.00020183542883107908, "loss": 4.9877, "step": 451900 }, { "epoch": 2.7251842114783953, "grad_norm": 1.8235141038894653, "learning_rate": 0.00020181633302446164, "loss": 5.1644, "step": 451950 }, { "epoch": 2.725485703259729, "grad_norm": 1.774613618850708, "learning_rate": 0.00020179723626423854, "loss": 4.9376, "step": 452000 }, { "epoch": 2.7257871950410633, "grad_norm": 1.7845473289489746, "learning_rate": 0.0002017781385507612, "loss": 5.0254, "step": 452050 }, { "epoch": 2.726088686822397, "grad_norm": 2.0269272327423096, "learning_rate": 0.00020175903988438115, "loss": 5.2192, "step": 452100 }, { "epoch": 2.7263901786037312, "grad_norm": 2.019245147705078, "learning_rate": 0.00020173994026544977, "loss": 4.9466, "step": 452150 }, { "epoch": 2.7266916703850654, "grad_norm": 2.592400074005127, "learning_rate": 0.0002017208396943187, "loss": 5.3021, "step": 452200 }, { "epoch": 2.726993162166399, "grad_norm": 2.4011828899383545, "learning_rate": 0.00020170173817133933, "loss": 4.6906, "step": 452250 }, { "epoch": 2.7272946539477334, "grad_norm": 1.7890441417694092, "learning_rate": 0.00020168263569686324, "loss": 5.1362, "step": 452300 }, { "epoch": 2.727596145729067, "grad_norm": 1.3830149173736572, "learning_rate": 0.00020166353227124207, "loss": 4.8003, "step": 452350 }, { "epoch": 2.7278976375104014, "grad_norm": 1.4411622285842896, "learning_rate": 0.0002016444278948273, "loss": 4.6675, "step": 452400 }, { "epoch": 2.7281991292917356, "grad_norm": 1.1712998151779175, "learning_rate": 0.00020162532256797047, "loss": 4.7306, "step": 452450 }, { "epoch": 2.72850062107307, "grad_norm": 1.992112398147583, "learning_rate": 0.0002016062162910233, "loss": 5.1736, "step": 452500 }, { "epoch": 2.7288021128544035, "grad_norm": 1.71371591091156, "learning_rate": 0.00020158710906433735, "loss": 4.8852, "step": 452550 }, { "epoch": 2.7291036046357378, "grad_norm": 2.0203330516815186, "learning_rate": 0.0002015680008882643, "loss": 5.0563, "step": 452600 }, { "epoch": 2.7294050964170715, "grad_norm": 2.049924850463867, "learning_rate": 0.00020154889176315574, "loss": 4.9292, "step": 452650 }, { "epoch": 2.7297065881984057, "grad_norm": 1.825777530670166, "learning_rate": 0.0002015297816893634, "loss": 5.0035, "step": 452700 }, { "epoch": 2.73000807997974, "grad_norm": 2.486621618270874, "learning_rate": 0.000201510670667239, "loss": 4.7818, "step": 452750 }, { "epoch": 2.7303095717610737, "grad_norm": 1.4086512327194214, "learning_rate": 0.0002014915586971342, "loss": 4.6798, "step": 452800 }, { "epoch": 2.730611063542408, "grad_norm": 1.5407963991165161, "learning_rate": 0.00020147244577940074, "loss": 4.7742, "step": 452850 }, { "epoch": 2.7309125553237417, "grad_norm": 0.6978301405906677, "learning_rate": 0.00020145333191439038, "loss": 4.7845, "step": 452900 }, { "epoch": 2.731214047105076, "grad_norm": 1.787386417388916, "learning_rate": 0.00020143421710245487, "loss": 4.7295, "step": 452950 }, { "epoch": 2.73151553888641, "grad_norm": 1.657888412475586, "learning_rate": 0.000201415101343946, "loss": 4.8318, "step": 453000 }, { "epoch": 2.73151553888641, "eval_loss": 5.40116548538208, "eval_runtime": 38.8236, "eval_samples_per_second": 13.188, "eval_steps_per_second": 6.594, "eval_tts_loss": 7.439000501237922, "step": 453000 }, { "epoch": 2.731817030667744, "grad_norm": 1.5493990182876587, "learning_rate": 0.0002013959846392155, "loss": 5.1104, "step": 453050 }, { "epoch": 2.732118522449078, "grad_norm": 1.8549036979675293, "learning_rate": 0.00020137686698861533, "loss": 5.2545, "step": 453100 }, { "epoch": 2.732420014230412, "grad_norm": 1.9101927280426025, "learning_rate": 0.00020135774839249722, "loss": 5.272, "step": 453150 }, { "epoch": 2.732721506011746, "grad_norm": 1.660157561302185, "learning_rate": 0.00020133862885121303, "loss": 4.9242, "step": 453200 }, { "epoch": 2.7330229977930802, "grad_norm": 1.5600529909133911, "learning_rate": 0.00020131950836511467, "loss": 4.9793, "step": 453250 }, { "epoch": 2.7333244895744144, "grad_norm": 1.5844366550445557, "learning_rate": 0.00020130038693455394, "loss": 4.9643, "step": 453300 }, { "epoch": 2.733625981355748, "grad_norm": 0.8248626589775085, "learning_rate": 0.00020128126455988283, "loss": 4.8454, "step": 453350 }, { "epoch": 2.7339274731370824, "grad_norm": 1.8325097560882568, "learning_rate": 0.00020126214124145328, "loss": 4.7852, "step": 453400 }, { "epoch": 2.734228964918416, "grad_norm": 2.2166876792907715, "learning_rate": 0.00020124301697961708, "loss": 5.2121, "step": 453450 }, { "epoch": 2.7345304566997504, "grad_norm": 2.6625282764434814, "learning_rate": 0.00020122389177472635, "loss": 5.1973, "step": 453500 }, { "epoch": 2.7348319484810846, "grad_norm": 1.663860559463501, "learning_rate": 0.00020120476562713295, "loss": 4.3638, "step": 453550 }, { "epoch": 2.7351334402624183, "grad_norm": 1.5550576448440552, "learning_rate": 0.00020118563853718894, "loss": 5.0811, "step": 453600 }, { "epoch": 2.7354349320437525, "grad_norm": 1.5413849353790283, "learning_rate": 0.0002011665105052463, "loss": 5.2512, "step": 453650 }, { "epoch": 2.7357364238250863, "grad_norm": 1.1223336458206177, "learning_rate": 0.00020114738153165703, "loss": 5.2117, "step": 453700 }, { "epoch": 2.7360379156064205, "grad_norm": 1.7785531282424927, "learning_rate": 0.0002011282516167732, "loss": 5.0285, "step": 453750 }, { "epoch": 2.7363394073877547, "grad_norm": 1.5468733310699463, "learning_rate": 0.00020110912076094688, "loss": 5.2255, "step": 453800 }, { "epoch": 2.736640899169089, "grad_norm": 2.259826183319092, "learning_rate": 0.0002010899889645301, "loss": 4.7853, "step": 453850 }, { "epoch": 2.7369423909504227, "grad_norm": 1.6273343563079834, "learning_rate": 0.00020107085622787498, "loss": 5.571, "step": 453900 }, { "epoch": 2.737243882731757, "grad_norm": 1.5784415006637573, "learning_rate": 0.00020105172255133365, "loss": 4.8663, "step": 453950 }, { "epoch": 2.7375453745130907, "grad_norm": 1.527209997177124, "learning_rate": 0.00020103258793525828, "loss": 5.0474, "step": 454000 }, { "epoch": 2.737846866294425, "grad_norm": 1.8217687606811523, "learning_rate": 0.00020101345238000089, "loss": 4.7162, "step": 454050 }, { "epoch": 2.738148358075759, "grad_norm": 2.701524019241333, "learning_rate": 0.0002009943158859137, "loss": 5.443, "step": 454100 }, { "epoch": 2.738449849857093, "grad_norm": 0.4666427969932556, "learning_rate": 0.00020097517845334893, "loss": 4.3968, "step": 454150 }, { "epoch": 2.738751341638427, "grad_norm": 2.1561810970306396, "learning_rate": 0.00020095604008265873, "loss": 5.0108, "step": 454200 }, { "epoch": 2.739052833419761, "grad_norm": 3.1466729640960693, "learning_rate": 0.00020093690077419533, "loss": 5.2677, "step": 454250 }, { "epoch": 2.739354325201095, "grad_norm": 2.046656847000122, "learning_rate": 0.00020091776052831098, "loss": 5.2044, "step": 454300 }, { "epoch": 2.739655816982429, "grad_norm": 2.500913381576538, "learning_rate": 0.00020089861934535785, "loss": 4.6566, "step": 454350 }, { "epoch": 2.739957308763763, "grad_norm": 2.0257601737976074, "learning_rate": 0.00020087947722568835, "loss": 5.1279, "step": 454400 }, { "epoch": 2.740258800545097, "grad_norm": 1.5149049758911133, "learning_rate": 0.00020086033416965465, "loss": 5.1741, "step": 454450 }, { "epoch": 2.740560292326431, "grad_norm": 1.578906536102295, "learning_rate": 0.00020084119017760904, "loss": 5.3344, "step": 454500 }, { "epoch": 2.740861784107765, "grad_norm": 2.6642043590545654, "learning_rate": 0.00020082204524990394, "loss": 5.1609, "step": 454550 }, { "epoch": 2.7411632758890994, "grad_norm": 1.6247215270996094, "learning_rate": 0.00020080289938689163, "loss": 4.9582, "step": 454600 }, { "epoch": 2.7414647676704336, "grad_norm": 1.6894910335540771, "learning_rate": 0.00020078375258892437, "loss": 5.2741, "step": 454650 }, { "epoch": 2.7417662594517673, "grad_norm": 1.6985448598861694, "learning_rate": 0.00020076460485635468, "loss": 5.0, "step": 454700 }, { "epoch": 2.7420677512331015, "grad_norm": 1.7425397634506226, "learning_rate": 0.0002007454561895349, "loss": 4.9613, "step": 454750 }, { "epoch": 2.7423692430144353, "grad_norm": 1.950602412223816, "learning_rate": 0.0002007263065888174, "loss": 4.7021, "step": 454800 }, { "epoch": 2.7426707347957695, "grad_norm": 1.6112604141235352, "learning_rate": 0.0002007071560545546, "loss": 4.7051, "step": 454850 }, { "epoch": 2.7429722265771037, "grad_norm": 1.9192174673080444, "learning_rate": 0.00020068800458709896, "loss": 4.9761, "step": 454900 }, { "epoch": 2.7432737183584375, "grad_norm": 1.4280012845993042, "learning_rate": 0.0002006688521868029, "loss": 5.3809, "step": 454950 }, { "epoch": 2.7435752101397717, "grad_norm": 1.2673091888427734, "learning_rate": 0.00020064969885401902, "loss": 4.8929, "step": 455000 }, { "epoch": 2.7438767019211054, "grad_norm": 2.5095977783203125, "learning_rate": 0.00020063054458909963, "loss": 5.2689, "step": 455050 }, { "epoch": 2.7441781937024396, "grad_norm": 1.5578949451446533, "learning_rate": 0.00020061138939239735, "loss": 5.38, "step": 455100 }, { "epoch": 2.744479685483774, "grad_norm": 1.3271307945251465, "learning_rate": 0.00020059223326426465, "loss": 4.7299, "step": 455150 }, { "epoch": 2.7447811772651076, "grad_norm": 1.963047742843628, "learning_rate": 0.00020057307620505413, "loss": 4.7624, "step": 455200 }, { "epoch": 2.745082669046442, "grad_norm": 1.0230371952056885, "learning_rate": 0.0002005539182151183, "loss": 4.9126, "step": 455250 }, { "epoch": 2.7453841608277756, "grad_norm": 1.3907092809677124, "learning_rate": 0.00020053475929480974, "loss": 4.9473, "step": 455300 }, { "epoch": 2.74568565260911, "grad_norm": 1.5692799091339111, "learning_rate": 0.00020051559944448105, "loss": 4.9571, "step": 455350 }, { "epoch": 2.745987144390444, "grad_norm": 1.629289150238037, "learning_rate": 0.00020049643866448485, "loss": 5.1417, "step": 455400 }, { "epoch": 2.746288636171778, "grad_norm": 1.4018985033035278, "learning_rate": 0.00020047727695517379, "loss": 5.1268, "step": 455450 }, { "epoch": 2.746590127953112, "grad_norm": 1.3916101455688477, "learning_rate": 0.0002004581143169004, "loss": 4.9963, "step": 455500 }, { "epoch": 2.746891619734446, "grad_norm": 2.9990477561950684, "learning_rate": 0.0002004389507500175, "loss": 5.2004, "step": 455550 }, { "epoch": 2.74719311151578, "grad_norm": 2.5733537673950195, "learning_rate": 0.00020041978625487769, "loss": 5.0499, "step": 455600 }, { "epoch": 2.747494603297114, "grad_norm": 0.7276865839958191, "learning_rate": 0.00020040062083183363, "loss": 4.9305, "step": 455650 }, { "epoch": 2.7477960950784484, "grad_norm": 3.6865766048431396, "learning_rate": 0.0002003814544812381, "loss": 4.8008, "step": 455700 }, { "epoch": 2.748097586859782, "grad_norm": 2.2855565547943115, "learning_rate": 0.00020036228720344377, "loss": 4.5672, "step": 455750 }, { "epoch": 2.7483990786411163, "grad_norm": 2.1196179389953613, "learning_rate": 0.00020034311899880342, "loss": 4.8627, "step": 455800 }, { "epoch": 2.74870057042245, "grad_norm": 1.6538017988204956, "learning_rate": 0.00020032394986766984, "loss": 4.9232, "step": 455850 }, { "epoch": 2.7490020622037843, "grad_norm": 1.959984540939331, "learning_rate": 0.00020030477981039575, "loss": 5.0956, "step": 455900 }, { "epoch": 2.7493035539851185, "grad_norm": 1.5302529335021973, "learning_rate": 0.00020028560882733395, "loss": 4.877, "step": 455950 }, { "epoch": 2.7496050457664523, "grad_norm": 2.0948398113250732, "learning_rate": 0.00020026643691883732, "loss": 5.0385, "step": 456000 }, { "epoch": 2.7496050457664523, "eval_loss": 5.414810657501221, "eval_runtime": 39.0267, "eval_samples_per_second": 13.119, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.390161469191185, "step": 456000 }, { "epoch": 2.7499065375477865, "grad_norm": 2.2157557010650635, "learning_rate": 0.00020024726408525868, "loss": 5.1647, "step": 456050 }, { "epoch": 2.7502080293291202, "grad_norm": 3.4126384258270264, "learning_rate": 0.00020022809032695077, "loss": 5.2316, "step": 456100 }, { "epoch": 2.7505095211104544, "grad_norm": 2.1786627769470215, "learning_rate": 0.00020020891564426658, "loss": 5.1146, "step": 456150 }, { "epoch": 2.7508110128917886, "grad_norm": 2.2519431114196777, "learning_rate": 0.00020018974003755901, "loss": 4.6145, "step": 456200 }, { "epoch": 2.751112504673123, "grad_norm": 1.3473002910614014, "learning_rate": 0.0002001705635071808, "loss": 5.1694, "step": 456250 }, { "epoch": 2.7514139964544566, "grad_norm": 2.1016488075256348, "learning_rate": 0.000200151386053485, "loss": 5.031, "step": 456300 }, { "epoch": 2.751715488235791, "grad_norm": 1.9019198417663574, "learning_rate": 0.00020013220767682455, "loss": 4.6779, "step": 456350 }, { "epoch": 2.7520169800171246, "grad_norm": 1.928911566734314, "learning_rate": 0.00020011302837755232, "loss": 5.1093, "step": 456400 }, { "epoch": 2.752318471798459, "grad_norm": 2.8798859119415283, "learning_rate": 0.00020009384815602134, "loss": 5.2547, "step": 456450 }, { "epoch": 2.752619963579793, "grad_norm": 3.0856435298919678, "learning_rate": 0.00020007466701258452, "loss": 5.0689, "step": 456500 }, { "epoch": 2.7529214553611268, "grad_norm": 1.103474736213684, "learning_rate": 0.00020005548494759495, "loss": 4.9908, "step": 456550 }, { "epoch": 2.753222947142461, "grad_norm": 1.2037525177001953, "learning_rate": 0.00020003630196140565, "loss": 4.8252, "step": 456600 }, { "epoch": 2.7535244389237947, "grad_norm": 1.3881419897079468, "learning_rate": 0.00020001711805436954, "loss": 4.9457, "step": 456650 }, { "epoch": 2.753825930705129, "grad_norm": 1.6240195035934448, "learning_rate": 0.0001999979332268398, "loss": 4.9543, "step": 456700 }, { "epoch": 2.754127422486463, "grad_norm": 2.6053853034973145, "learning_rate": 0.00019997874747916945, "loss": 5.2322, "step": 456750 }, { "epoch": 2.754428914267797, "grad_norm": 1.991029977798462, "learning_rate": 0.00019995956081171156, "loss": 5.2697, "step": 456800 }, { "epoch": 2.754730406049131, "grad_norm": 2.8979814052581787, "learning_rate": 0.00019994037322481925, "loss": 4.6813, "step": 456850 }, { "epoch": 2.755031897830465, "grad_norm": 2.448270797729492, "learning_rate": 0.00019992118471884567, "loss": 5.1126, "step": 456900 }, { "epoch": 2.755333389611799, "grad_norm": 1.5729151964187622, "learning_rate": 0.00019990199529414392, "loss": 5.146, "step": 456950 }, { "epoch": 2.7556348813931333, "grad_norm": 2.8259778022766113, "learning_rate": 0.00019988280495106714, "loss": 5.0853, "step": 457000 }, { "epoch": 2.7559363731744675, "grad_norm": 1.7079858779907227, "learning_rate": 0.00019986361368996852, "loss": 5.0249, "step": 457050 }, { "epoch": 2.7562378649558013, "grad_norm": 2.229976177215576, "learning_rate": 0.00019984442151120123, "loss": 5.3265, "step": 457100 }, { "epoch": 2.7565393567371355, "grad_norm": 0.5638532638549805, "learning_rate": 0.00019982522841511852, "loss": 4.8463, "step": 457150 }, { "epoch": 2.7568408485184692, "grad_norm": 2.034048318862915, "learning_rate": 0.00019980603440207363, "loss": 5.1273, "step": 457200 }, { "epoch": 2.7571423402998034, "grad_norm": 2.485973358154297, "learning_rate": 0.00019978683947241967, "loss": 4.7078, "step": 457250 }, { "epoch": 2.7574438320811376, "grad_norm": 2.4216625690460205, "learning_rate": 0.00019976764362651003, "loss": 5.1669, "step": 457300 }, { "epoch": 2.7577453238624714, "grad_norm": 1.9344652891159058, "learning_rate": 0.00019974844686469794, "loss": 4.9945, "step": 457350 }, { "epoch": 2.7580468156438056, "grad_norm": 1.7647978067398071, "learning_rate": 0.00019972924918733662, "loss": 4.778, "step": 457400 }, { "epoch": 2.7583483074251394, "grad_norm": 0.6426975727081299, "learning_rate": 0.0001997100505947795, "loss": 4.454, "step": 457450 }, { "epoch": 2.7586497992064736, "grad_norm": 0.5789716243743896, "learning_rate": 0.0001996908510873798, "loss": 4.9244, "step": 457500 }, { "epoch": 2.758951290987808, "grad_norm": 0.6898590922355652, "learning_rate": 0.00019967165066549092, "loss": 4.9419, "step": 457550 }, { "epoch": 2.759252782769142, "grad_norm": 1.4390745162963867, "learning_rate": 0.0001996524493294662, "loss": 5.2557, "step": 457600 }, { "epoch": 2.7595542745504757, "grad_norm": 0.6386452913284302, "learning_rate": 0.000199633247079659, "loss": 4.9384, "step": 457650 }, { "epoch": 2.75985576633181, "grad_norm": 2.5157339572906494, "learning_rate": 0.00019961404391642267, "loss": 5.2648, "step": 457700 }, { "epoch": 2.7601572581131437, "grad_norm": 2.040323495864868, "learning_rate": 0.00019959483984011075, "loss": 4.7383, "step": 457750 }, { "epoch": 2.760458749894478, "grad_norm": 1.4971867799758911, "learning_rate": 0.00019957563485107653, "loss": 4.9003, "step": 457800 }, { "epoch": 2.760760241675812, "grad_norm": 1.738012671470642, "learning_rate": 0.00019955642894967348, "loss": 5.4721, "step": 457850 }, { "epoch": 2.761061733457146, "grad_norm": 1.8916194438934326, "learning_rate": 0.0001995372221362551, "loss": 4.9199, "step": 457900 }, { "epoch": 2.76136322523848, "grad_norm": 1.9802134037017822, "learning_rate": 0.00019951801441117487, "loss": 5.2851, "step": 457950 }, { "epoch": 2.761664717019814, "grad_norm": 2.0389022827148438, "learning_rate": 0.00019949880577478622, "loss": 5.2796, "step": 458000 }, { "epoch": 2.761966208801148, "grad_norm": 1.3266185522079468, "learning_rate": 0.00019947959622744268, "loss": 4.9503, "step": 458050 }, { "epoch": 2.7622677005824823, "grad_norm": 1.7482104301452637, "learning_rate": 0.00019946038576949777, "loss": 4.3805, "step": 458100 }, { "epoch": 2.762569192363816, "grad_norm": 1.322411298751831, "learning_rate": 0.00019944117440130507, "loss": 4.7613, "step": 458150 }, { "epoch": 2.7628706841451502, "grad_norm": 2.288862466812134, "learning_rate": 0.00019942196212321815, "loss": 4.9372, "step": 458200 }, { "epoch": 2.763172175926484, "grad_norm": 0.8703647255897522, "learning_rate": 0.00019940274893559046, "loss": 4.4574, "step": 458250 }, { "epoch": 2.763473667707818, "grad_norm": 1.4987354278564453, "learning_rate": 0.00019938353483877572, "loss": 5.0922, "step": 458300 }, { "epoch": 2.7637751594891524, "grad_norm": 1.9265655279159546, "learning_rate": 0.0001993643198331275, "loss": 5.1142, "step": 458350 }, { "epoch": 2.7640766512704866, "grad_norm": 2.6487996578216553, "learning_rate": 0.0001993451039189994, "loss": 5.0529, "step": 458400 }, { "epoch": 2.7643781430518204, "grad_norm": 1.8007479906082153, "learning_rate": 0.0001993258870967451, "loss": 4.8723, "step": 458450 }, { "epoch": 2.7646796348331546, "grad_norm": 2.618647575378418, "learning_rate": 0.00019930666936671828, "loss": 4.8063, "step": 458500 }, { "epoch": 2.7649811266144884, "grad_norm": 2.783189058303833, "learning_rate": 0.00019928745072927254, "loss": 4.9858, "step": 458550 }, { "epoch": 2.7652826183958226, "grad_norm": 1.4118058681488037, "learning_rate": 0.0001992682311847616, "loss": 4.818, "step": 458600 }, { "epoch": 2.7655841101771568, "grad_norm": 1.7567675113677979, "learning_rate": 0.00019924901073353917, "loss": 4.9404, "step": 458650 }, { "epoch": 2.7658856019584905, "grad_norm": 1.9892531633377075, "learning_rate": 0.00019922978937595895, "loss": 4.5987, "step": 458700 }, { "epoch": 2.7661870937398247, "grad_norm": 2.0747933387756348, "learning_rate": 0.00019921056711237476, "loss": 4.7676, "step": 458750 }, { "epoch": 2.7664885855211585, "grad_norm": 1.5440841913223267, "learning_rate": 0.0001991913439431403, "loss": 4.9533, "step": 458800 }, { "epoch": 2.7667900773024927, "grad_norm": 2.284054756164551, "learning_rate": 0.00019917211986860935, "loss": 4.8095, "step": 458850 }, { "epoch": 2.767091569083827, "grad_norm": 1.5165280103683472, "learning_rate": 0.00019915289488913571, "loss": 4.7418, "step": 458900 }, { "epoch": 2.7673930608651607, "grad_norm": 1.1950578689575195, "learning_rate": 0.00019913366900507315, "loss": 5.1678, "step": 458950 }, { "epoch": 2.767694552646495, "grad_norm": 1.9598878622055054, "learning_rate": 0.00019911444221677559, "loss": 4.7405, "step": 459000 }, { "epoch": 2.767694552646495, "eval_loss": 5.3985066413879395, "eval_runtime": 39.0499, "eval_samples_per_second": 13.111, "eval_steps_per_second": 6.556, "eval_tts_loss": 7.467903544342052, "step": 459000 }, { "epoch": 2.7679960444278287, "grad_norm": 2.3936994075775146, "learning_rate": 0.00019909521452459675, "loss": 4.4672, "step": 459050 }, { "epoch": 2.768297536209163, "grad_norm": 2.1828980445861816, "learning_rate": 0.0001990759859288906, "loss": 5.1429, "step": 459100 }, { "epoch": 2.768599027990497, "grad_norm": 2.5644032955169678, "learning_rate": 0.00019905675643001094, "loss": 4.7681, "step": 459150 }, { "epoch": 2.7689005197718313, "grad_norm": 1.6772922277450562, "learning_rate": 0.0001990375260283117, "loss": 4.7359, "step": 459200 }, { "epoch": 2.769202011553165, "grad_norm": 2.062356948852539, "learning_rate": 0.00019901829472414675, "loss": 5.1559, "step": 459250 }, { "epoch": 2.7695035033344992, "grad_norm": 1.8054461479187012, "learning_rate": 0.00019899906251787005, "loss": 4.7314, "step": 459300 }, { "epoch": 2.769804995115833, "grad_norm": 1.5994125604629517, "learning_rate": 0.00019897982940983552, "loss": 4.6026, "step": 459350 }, { "epoch": 2.770106486897167, "grad_norm": 1.7733192443847656, "learning_rate": 0.00019896059540039718, "loss": 5.0853, "step": 459400 }, { "epoch": 2.7704079786785014, "grad_norm": 1.3434486389160156, "learning_rate": 0.0001989413604899089, "loss": 5.1145, "step": 459450 }, { "epoch": 2.770709470459835, "grad_norm": 2.451122760772705, "learning_rate": 0.0001989221246787247, "loss": 4.9593, "step": 459500 }, { "epoch": 2.7710109622411694, "grad_norm": 0.6097370982170105, "learning_rate": 0.00019890288796719865, "loss": 4.8247, "step": 459550 }, { "epoch": 2.771312454022503, "grad_norm": 2.190502643585205, "learning_rate": 0.00019888365035568477, "loss": 4.9269, "step": 459600 }, { "epoch": 2.7716139458038374, "grad_norm": 2.6696271896362305, "learning_rate": 0.000198864411844537, "loss": 5.0738, "step": 459650 }, { "epoch": 2.7719154375851716, "grad_norm": 1.622506022453308, "learning_rate": 0.00019884517243410948, "loss": 4.9277, "step": 459700 }, { "epoch": 2.7722169293665053, "grad_norm": 2.892160654067993, "learning_rate": 0.00019882593212475628, "loss": 5.2357, "step": 459750 }, { "epoch": 2.7725184211478395, "grad_norm": 2.3896265029907227, "learning_rate": 0.0001988066909168315, "loss": 4.8616, "step": 459800 }, { "epoch": 2.7728199129291733, "grad_norm": 1.9759312868118286, "learning_rate": 0.00019878744881068915, "loss": 4.6689, "step": 459850 }, { "epoch": 2.7731214047105075, "grad_norm": 1.534446358680725, "learning_rate": 0.00019876820580668348, "loss": 4.6833, "step": 459900 }, { "epoch": 2.7734228964918417, "grad_norm": 0.40564554929733276, "learning_rate": 0.00019874896190516858, "loss": 4.9074, "step": 459950 }, { "epoch": 2.773724388273176, "grad_norm": 0.7248961329460144, "learning_rate": 0.0001987297171064986, "loss": 5.1218, "step": 460000 }, { "epoch": 2.7740258800545097, "grad_norm": 1.756866216659546, "learning_rate": 0.00019871047141102773, "loss": 5.3142, "step": 460050 }, { "epoch": 2.774327371835844, "grad_norm": 2.1495542526245117, "learning_rate": 0.00019869122481911016, "loss": 4.8856, "step": 460100 }, { "epoch": 2.7746288636171776, "grad_norm": 1.952327013015747, "learning_rate": 0.00019867197733110004, "loss": 5.0059, "step": 460150 }, { "epoch": 2.774930355398512, "grad_norm": 1.87051260471344, "learning_rate": 0.00019865272894735165, "loss": 4.9715, "step": 460200 }, { "epoch": 2.775231847179846, "grad_norm": 1.6854300498962402, "learning_rate": 0.00019863347966821921, "loss": 4.871, "step": 460250 }, { "epoch": 2.77553333896118, "grad_norm": 2.7221648693084717, "learning_rate": 0.00019861422949405698, "loss": 4.9605, "step": 460300 }, { "epoch": 2.775834830742514, "grad_norm": 1.8508780002593994, "learning_rate": 0.00019859497842521925, "loss": 5.0504, "step": 460350 }, { "epoch": 2.776136322523848, "grad_norm": 0.6387856006622314, "learning_rate": 0.0001985757264620603, "loss": 4.8438, "step": 460400 }, { "epoch": 2.776437814305182, "grad_norm": 1.5976754426956177, "learning_rate": 0.00019855647360493442, "loss": 5.1673, "step": 460450 }, { "epoch": 2.776739306086516, "grad_norm": 1.292983889579773, "learning_rate": 0.0001985372198541959, "loss": 4.9624, "step": 460500 }, { "epoch": 2.77704079786785, "grad_norm": 1.8826771974563599, "learning_rate": 0.00019851796521019914, "loss": 4.9815, "step": 460550 }, { "epoch": 2.777342289649184, "grad_norm": 2.4618821144104004, "learning_rate": 0.00019849870967329848, "loss": 5.0758, "step": 460600 }, { "epoch": 2.777643781430518, "grad_norm": 1.786635160446167, "learning_rate": 0.0001984794532438483, "loss": 4.8587, "step": 460650 }, { "epoch": 2.777945273211852, "grad_norm": 3.995833158493042, "learning_rate": 0.0001984601959222029, "loss": 4.8829, "step": 460700 }, { "epoch": 2.7782467649931863, "grad_norm": 2.2147984504699707, "learning_rate": 0.00019844093770871678, "loss": 5.0685, "step": 460750 }, { "epoch": 2.7785482567745206, "grad_norm": 2.178452491760254, "learning_rate": 0.00019842167860374435, "loss": 4.988, "step": 460800 }, { "epoch": 2.7788497485558543, "grad_norm": 2.108867883682251, "learning_rate": 0.00019840241860764, "loss": 5.0913, "step": 460850 }, { "epoch": 2.7791512403371885, "grad_norm": 0.9423072934150696, "learning_rate": 0.00019838315772075822, "loss": 5.1552, "step": 460900 }, { "epoch": 2.7794527321185223, "grad_norm": 1.4368114471435547, "learning_rate": 0.00019836389594345347, "loss": 5.1343, "step": 460950 }, { "epoch": 2.7797542238998565, "grad_norm": 1.6871592998504639, "learning_rate": 0.00019834463327608027, "loss": 5.1639, "step": 461000 }, { "epoch": 2.7800557156811907, "grad_norm": 1.5805563926696777, "learning_rate": 0.000198325369718993, "loss": 5.19, "step": 461050 }, { "epoch": 2.7803572074625245, "grad_norm": 2.023087739944458, "learning_rate": 0.00019830610527254636, "loss": 4.4833, "step": 461100 }, { "epoch": 2.7806586992438587, "grad_norm": 1.7646499872207642, "learning_rate": 0.00019828683993709472, "loss": 5.0064, "step": 461150 }, { "epoch": 2.7809601910251924, "grad_norm": 0.9861526489257812, "learning_rate": 0.00019826757371299277, "loss": 4.9412, "step": 461200 }, { "epoch": 2.7812616828065266, "grad_norm": 1.6163301467895508, "learning_rate": 0.00019824830660059492, "loss": 4.8708, "step": 461250 }, { "epoch": 2.781563174587861, "grad_norm": 0.9076203107833862, "learning_rate": 0.00019822903860025593, "loss": 4.8233, "step": 461300 }, { "epoch": 2.781864666369195, "grad_norm": 1.6415303945541382, "learning_rate": 0.00019820976971233022, "loss": 5.0832, "step": 461350 }, { "epoch": 2.782166158150529, "grad_norm": 1.6640948057174683, "learning_rate": 0.00019819049993717258, "loss": 5.2808, "step": 461400 }, { "epoch": 2.782467649931863, "grad_norm": 0.9712314605712891, "learning_rate": 0.0001981712292751375, "loss": 4.7433, "step": 461450 }, { "epoch": 2.782769141713197, "grad_norm": 1.50376558303833, "learning_rate": 0.0001981519577265797, "loss": 4.6357, "step": 461500 }, { "epoch": 2.783070633494531, "grad_norm": 2.1088335514068604, "learning_rate": 0.00019813268529185381, "loss": 5.213, "step": 461550 }, { "epoch": 2.783372125275865, "grad_norm": 2.7973906993865967, "learning_rate": 0.0001981134119713146, "loss": 5.0843, "step": 461600 }, { "epoch": 2.783673617057199, "grad_norm": 1.7906948328018188, "learning_rate": 0.00019809413776531666, "loss": 5.1961, "step": 461650 }, { "epoch": 2.783975108838533, "grad_norm": 2.262406826019287, "learning_rate": 0.00019807486267421477, "loss": 4.786, "step": 461700 }, { "epoch": 2.784276600619867, "grad_norm": 2.4347116947174072, "learning_rate": 0.00019805558669836365, "loss": 5.0146, "step": 461750 }, { "epoch": 2.784578092401201, "grad_norm": 1.5735349655151367, "learning_rate": 0.000198036309838118, "loss": 5.0816, "step": 461800 }, { "epoch": 2.7848795841825353, "grad_norm": 3.134788751602173, "learning_rate": 0.00019801703209383265, "loss": 4.9692, "step": 461850 }, { "epoch": 2.785181075963869, "grad_norm": 1.8239126205444336, "learning_rate": 0.00019799775346586234, "loss": 5.0902, "step": 461900 }, { "epoch": 2.7854825677452033, "grad_norm": 2.2143282890319824, "learning_rate": 0.00019797847395456187, "loss": 5.1247, "step": 461950 }, { "epoch": 2.785784059526537, "grad_norm": 1.2365515232086182, "learning_rate": 0.0001979591935602861, "loss": 5.0598, "step": 462000 }, { "epoch": 2.785784059526537, "eval_loss": 5.396196365356445, "eval_runtime": 38.9939, "eval_samples_per_second": 13.13, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.409294491653213, "step": 462000 }, { "epoch": 2.7860855513078713, "grad_norm": 1.3413782119750977, "learning_rate": 0.00019793991228338973, "loss": 4.9766, "step": 462050 }, { "epoch": 2.7863870430892055, "grad_norm": 1.8724420070648193, "learning_rate": 0.00019792063012422773, "loss": 4.661, "step": 462100 }, { "epoch": 2.7866885348705397, "grad_norm": 1.8930284976959229, "learning_rate": 0.00019790134708315492, "loss": 4.9912, "step": 462150 }, { "epoch": 2.7869900266518735, "grad_norm": 2.8510429859161377, "learning_rate": 0.0001978820631605262, "loss": 4.9235, "step": 462200 }, { "epoch": 2.7872915184332077, "grad_norm": 1.9185762405395508, "learning_rate": 0.00019786277835669638, "loss": 4.8339, "step": 462250 }, { "epoch": 2.7875930102145414, "grad_norm": 1.4876285791397095, "learning_rate": 0.00019784349267202049, "loss": 4.8664, "step": 462300 }, { "epoch": 2.7878945019958756, "grad_norm": 1.165963053703308, "learning_rate": 0.00019782420610685335, "loss": 5.2649, "step": 462350 }, { "epoch": 2.78819599377721, "grad_norm": 2.8829147815704346, "learning_rate": 0.00019780491866154998, "loss": 5.0889, "step": 462400 }, { "epoch": 2.7884974855585436, "grad_norm": 1.8570637702941895, "learning_rate": 0.00019778563033646528, "loss": 4.8713, "step": 462450 }, { "epoch": 2.788798977339878, "grad_norm": 1.7412141561508179, "learning_rate": 0.00019776634113195422, "loss": 4.6311, "step": 462500 }, { "epoch": 2.7891004691212116, "grad_norm": 1.1346780061721802, "learning_rate": 0.0001977470510483719, "loss": 4.821, "step": 462550 }, { "epoch": 2.7894019609025458, "grad_norm": 1.6396081447601318, "learning_rate": 0.00019772776008607318, "loss": 5.2309, "step": 462600 }, { "epoch": 2.78970345268388, "grad_norm": 2.489706516265869, "learning_rate": 0.00019770846824541312, "loss": 4.716, "step": 462650 }, { "epoch": 2.7900049444652137, "grad_norm": 1.8090903759002686, "learning_rate": 0.00019768917552674683, "loss": 5.1374, "step": 462700 }, { "epoch": 2.790306436246548, "grad_norm": 1.5389113426208496, "learning_rate": 0.0001976698819304293, "loss": 5.2906, "step": 462750 }, { "epoch": 2.7906079280278817, "grad_norm": 1.5809299945831299, "learning_rate": 0.00019765058745681565, "loss": 4.9728, "step": 462800 }, { "epoch": 2.790909419809216, "grad_norm": 2.2716948986053467, "learning_rate": 0.00019763129210626092, "loss": 4.9437, "step": 462850 }, { "epoch": 2.79121091159055, "grad_norm": 1.1441104412078857, "learning_rate": 0.00019761199587912023, "loss": 5.3748, "step": 462900 }, { "epoch": 2.7915124033718843, "grad_norm": 1.676908254623413, "learning_rate": 0.00019759269877574876, "loss": 4.6189, "step": 462950 }, { "epoch": 2.791813895153218, "grad_norm": 2.447565793991089, "learning_rate": 0.00019757340079650152, "loss": 4.8369, "step": 463000 }, { "epoch": 2.7921153869345523, "grad_norm": 3.0306031703948975, "learning_rate": 0.00019755410194173377, "loss": 4.3635, "step": 463050 }, { "epoch": 2.792416878715886, "grad_norm": 0.9664621949195862, "learning_rate": 0.00019753480221180056, "loss": 4.9825, "step": 463100 }, { "epoch": 2.7927183704972203, "grad_norm": 1.9277151823043823, "learning_rate": 0.0001975155016070572, "loss": 4.6187, "step": 463150 }, { "epoch": 2.7930198622785545, "grad_norm": 1.4432224035263062, "learning_rate": 0.00019749620012785886, "loss": 4.9057, "step": 463200 }, { "epoch": 2.7933213540598882, "grad_norm": 1.2613712549209595, "learning_rate": 0.00019747689777456075, "loss": 5.0611, "step": 463250 }, { "epoch": 2.7936228458412224, "grad_norm": 2.3804690837860107, "learning_rate": 0.00019745759454751805, "loss": 4.7123, "step": 463300 }, { "epoch": 2.793924337622556, "grad_norm": 1.7513902187347412, "learning_rate": 0.00019743829044708606, "loss": 4.778, "step": 463350 }, { "epoch": 2.7942258294038904, "grad_norm": 1.7655911445617676, "learning_rate": 0.00019741898547362004, "loss": 5.1604, "step": 463400 }, { "epoch": 2.7945273211852246, "grad_norm": 1.4082399606704712, "learning_rate": 0.00019739967962747524, "loss": 5.0577, "step": 463450 }, { "epoch": 2.7948288129665584, "grad_norm": 3.3426811695098877, "learning_rate": 0.00019738037290900696, "loss": 4.8655, "step": 463500 }, { "epoch": 2.7951303047478926, "grad_norm": 1.7095271348953247, "learning_rate": 0.00019736106531857063, "loss": 4.9329, "step": 463550 }, { "epoch": 2.7954317965292264, "grad_norm": 1.8701353073120117, "learning_rate": 0.00019734175685652143, "loss": 5.0688, "step": 463600 }, { "epoch": 2.7957332883105606, "grad_norm": 2.518496036529541, "learning_rate": 0.00019732244752321469, "loss": 5.1994, "step": 463650 }, { "epoch": 2.7960347800918948, "grad_norm": 1.6147061586380005, "learning_rate": 0.00019730313731900588, "loss": 5.0258, "step": 463700 }, { "epoch": 2.796336271873229, "grad_norm": 1.0200350284576416, "learning_rate": 0.00019728382624425037, "loss": 4.6921, "step": 463750 }, { "epoch": 2.7966377636545627, "grad_norm": 1.3928370475769043, "learning_rate": 0.00019726451429930348, "loss": 5.1792, "step": 463800 }, { "epoch": 2.796939255435897, "grad_norm": 1.6017448902130127, "learning_rate": 0.00019724520148452066, "loss": 4.8605, "step": 463850 }, { "epoch": 2.7972407472172307, "grad_norm": 2.9868595600128174, "learning_rate": 0.00019722588780025736, "loss": 4.8248, "step": 463900 }, { "epoch": 2.797542238998565, "grad_norm": 1.9076348543167114, "learning_rate": 0.00019720657324686895, "loss": 4.7289, "step": 463950 }, { "epoch": 2.797843730779899, "grad_norm": 1.7042595148086548, "learning_rate": 0.00019718725782471094, "loss": 5.3079, "step": 464000 }, { "epoch": 2.798145222561233, "grad_norm": 0.8510984778404236, "learning_rate": 0.0001971679415341388, "loss": 4.9186, "step": 464050 }, { "epoch": 2.798446714342567, "grad_norm": 2.941509962081909, "learning_rate": 0.00019714862437550796, "loss": 4.2356, "step": 464100 }, { "epoch": 2.798748206123901, "grad_norm": 2.2494051456451416, "learning_rate": 0.00019712930634917404, "loss": 5.4542, "step": 464150 }, { "epoch": 2.799049697905235, "grad_norm": 2.1952571868896484, "learning_rate": 0.00019710998745549248, "loss": 4.9006, "step": 464200 }, { "epoch": 2.7993511896865693, "grad_norm": 1.0585049390792847, "learning_rate": 0.00019709066769481879, "loss": 4.4467, "step": 464250 }, { "epoch": 2.799652681467903, "grad_norm": 1.5079368352890015, "learning_rate": 0.0001970713470675086, "loss": 5.0465, "step": 464300 }, { "epoch": 2.7999541732492372, "grad_norm": 1.8129920959472656, "learning_rate": 0.00019705202557391744, "loss": 5.2991, "step": 464350 }, { "epoch": 2.800255665030571, "grad_norm": 1.819122076034546, "learning_rate": 0.00019703270321440086, "loss": 4.7139, "step": 464400 }, { "epoch": 2.800557156811905, "grad_norm": 0.6648308038711548, "learning_rate": 0.00019701337998931454, "loss": 4.7402, "step": 464450 }, { "epoch": 2.8008586485932394, "grad_norm": 1.4434560537338257, "learning_rate": 0.00019699405589901405, "loss": 4.7308, "step": 464500 }, { "epoch": 2.8011601403745736, "grad_norm": 0.751947283744812, "learning_rate": 0.00019697473094385502, "loss": 4.5761, "step": 464550 }, { "epoch": 2.8014616321559074, "grad_norm": 1.4210336208343506, "learning_rate": 0.00019695540512419308, "loss": 4.4567, "step": 464600 }, { "epoch": 2.8017631239372416, "grad_norm": 2.0489611625671387, "learning_rate": 0.0001969360784403839, "loss": 4.7786, "step": 464650 }, { "epoch": 2.8020646157185753, "grad_norm": 1.6590256690979004, "learning_rate": 0.0001969167508927832, "loss": 4.9962, "step": 464700 }, { "epoch": 2.8023661074999096, "grad_norm": 1.4699602127075195, "learning_rate": 0.0001968974224817467, "loss": 4.8662, "step": 464750 }, { "epoch": 2.8026675992812438, "grad_norm": 2.327765703201294, "learning_rate": 0.00019687809320763, "loss": 5.2724, "step": 464800 }, { "epoch": 2.8029690910625775, "grad_norm": 1.9495247602462769, "learning_rate": 0.00019685876307078893, "loss": 4.5265, "step": 464850 }, { "epoch": 2.8032705828439117, "grad_norm": 2.047973871231079, "learning_rate": 0.0001968394320715792, "loss": 4.9342, "step": 464900 }, { "epoch": 2.8035720746252455, "grad_norm": 2.674164056777954, "learning_rate": 0.00019682010021035654, "loss": 5.2718, "step": 464950 }, { "epoch": 2.8038735664065797, "grad_norm": 2.0195682048797607, "learning_rate": 0.00019680076748747677, "loss": 5.0216, "step": 465000 }, { "epoch": 2.8038735664065797, "eval_loss": 5.399842739105225, "eval_runtime": 39.0955, "eval_samples_per_second": 13.096, "eval_steps_per_second": 6.548, "eval_tts_loss": 7.398221706306304, "step": 465000 }, { "epoch": 2.804175058187914, "grad_norm": 0.6195836067199707, "learning_rate": 0.00019678143390329566, "loss": 4.973, "step": 465050 }, { "epoch": 2.804476549969248, "grad_norm": 1.8278272151947021, "learning_rate": 0.00019676209945816898, "loss": 4.7973, "step": 465100 }, { "epoch": 2.804778041750582, "grad_norm": 1.917458415031433, "learning_rate": 0.00019674276415245268, "loss": 5.0656, "step": 465150 }, { "epoch": 2.805079533531916, "grad_norm": 1.728176236152649, "learning_rate": 0.00019672342798650246, "loss": 5.0134, "step": 465200 }, { "epoch": 2.80538102531325, "grad_norm": 1.338276982307434, "learning_rate": 0.00019670409096067418, "loss": 4.9513, "step": 465250 }, { "epoch": 2.805682517094584, "grad_norm": 1.8247896432876587, "learning_rate": 0.00019668475307532384, "loss": 5.1727, "step": 465300 }, { "epoch": 2.8059840088759183, "grad_norm": 1.4304035902023315, "learning_rate": 0.00019666541433080721, "loss": 5.0032, "step": 465350 }, { "epoch": 2.806285500657252, "grad_norm": 2.365902900695801, "learning_rate": 0.00019664607472748023, "loss": 4.3807, "step": 465400 }, { "epoch": 2.8065869924385862, "grad_norm": 1.760999321937561, "learning_rate": 0.0001966267342656988, "loss": 4.6958, "step": 465450 }, { "epoch": 2.80688848421992, "grad_norm": 1.5760389566421509, "learning_rate": 0.00019660739294581885, "loss": 4.6767, "step": 465500 }, { "epoch": 2.807189976001254, "grad_norm": 0.5488294959068298, "learning_rate": 0.00019658805076819635, "loss": 5.0428, "step": 465550 }, { "epoch": 2.8074914677825884, "grad_norm": 1.2515382766723633, "learning_rate": 0.0001965687077331873, "loss": 4.7396, "step": 465600 }, { "epoch": 2.807792959563922, "grad_norm": 1.5754210948944092, "learning_rate": 0.00019654936384114763, "loss": 4.6843, "step": 465650 }, { "epoch": 2.8080944513452564, "grad_norm": 2.0343551635742188, "learning_rate": 0.0001965300190924333, "loss": 4.9616, "step": 465700 }, { "epoch": 2.80839594312659, "grad_norm": 2.4890754222869873, "learning_rate": 0.00019651067348740043, "loss": 5.0536, "step": 465750 }, { "epoch": 2.8086974349079243, "grad_norm": 3.6035032272338867, "learning_rate": 0.00019649132702640495, "loss": 4.7828, "step": 465800 }, { "epoch": 2.8089989266892585, "grad_norm": 1.7236692905426025, "learning_rate": 0.00019647197970980296, "loss": 4.6376, "step": 465850 }, { "epoch": 2.8093004184705928, "grad_norm": 1.499353289604187, "learning_rate": 0.00019645263153795047, "loss": 4.8008, "step": 465900 }, { "epoch": 2.8096019102519265, "grad_norm": 2.1033976078033447, "learning_rate": 0.00019643328251120362, "loss": 5.095, "step": 465950 }, { "epoch": 2.8099034020332607, "grad_norm": 2.512061357498169, "learning_rate": 0.0001964139326299185, "loss": 4.9221, "step": 466000 }, { "epoch": 2.8102048938145945, "grad_norm": 2.27772855758667, "learning_rate": 0.00019639458189445113, "loss": 5.0438, "step": 466050 }, { "epoch": 2.8105063855959287, "grad_norm": 1.120812177658081, "learning_rate": 0.00019637523030515772, "loss": 4.8866, "step": 466100 }, { "epoch": 2.810807877377263, "grad_norm": 1.3795241117477417, "learning_rate": 0.00019635587786239435, "loss": 5.039, "step": 466150 }, { "epoch": 2.8111093691585967, "grad_norm": 2.190532922744751, "learning_rate": 0.00019633652456651723, "loss": 4.974, "step": 466200 }, { "epoch": 2.811410860939931, "grad_norm": 2.008037567138672, "learning_rate": 0.0001963171704178825, "loss": 4.6724, "step": 466250 }, { "epoch": 2.8117123527212646, "grad_norm": 1.7060716152191162, "learning_rate": 0.0001962978154168463, "loss": 5.0956, "step": 466300 }, { "epoch": 2.812013844502599, "grad_norm": 3.1353931427001953, "learning_rate": 0.00019627845956376495, "loss": 4.5909, "step": 466350 }, { "epoch": 2.812315336283933, "grad_norm": 1.2680513858795166, "learning_rate": 0.00019625910285899458, "loss": 4.7349, "step": 466400 }, { "epoch": 2.812616828065267, "grad_norm": 2.6042139530181885, "learning_rate": 0.00019623974530289136, "loss": 4.8993, "step": 466450 }, { "epoch": 2.812918319846601, "grad_norm": 1.9054566621780396, "learning_rate": 0.00019622038689581168, "loss": 4.8069, "step": 466500 }, { "epoch": 2.8132198116279348, "grad_norm": 2.2665371894836426, "learning_rate": 0.00019620102763811173, "loss": 5.1591, "step": 466550 }, { "epoch": 2.813521303409269, "grad_norm": 1.8276485204696655, "learning_rate": 0.0001961816675301478, "loss": 5.2344, "step": 466600 }, { "epoch": 2.813822795190603, "grad_norm": 1.5926213264465332, "learning_rate": 0.00019616230657227622, "loss": 5.3105, "step": 466650 }, { "epoch": 2.8141242869719374, "grad_norm": 1.9201174974441528, "learning_rate": 0.00019614294476485316, "loss": 4.8679, "step": 466700 }, { "epoch": 2.814425778753271, "grad_norm": 1.6534258127212524, "learning_rate": 0.00019612358210823516, "loss": 5.2222, "step": 466750 }, { "epoch": 2.8147272705346054, "grad_norm": 1.8267642259597778, "learning_rate": 0.0001961042186027784, "loss": 5.412, "step": 466800 }, { "epoch": 2.815028762315939, "grad_norm": 1.0177404880523682, "learning_rate": 0.0001960848542488393, "loss": 4.7694, "step": 466850 }, { "epoch": 2.8153302540972733, "grad_norm": 1.5323551893234253, "learning_rate": 0.00019606548904677423, "loss": 4.9518, "step": 466900 }, { "epoch": 2.8156317458786075, "grad_norm": 2.30100154876709, "learning_rate": 0.00019604612299693958, "loss": 4.5268, "step": 466950 }, { "epoch": 2.8159332376599413, "grad_norm": 2.8155405521392822, "learning_rate": 0.00019602675609969175, "loss": 5.0985, "step": 467000 }, { "epoch": 2.8162347294412755, "grad_norm": 2.7134459018707275, "learning_rate": 0.00019600738835538715, "loss": 4.9932, "step": 467050 }, { "epoch": 2.8165362212226093, "grad_norm": 1.8382315635681152, "learning_rate": 0.0001959880197643822, "loss": 4.7763, "step": 467100 }, { "epoch": 2.8168377130039435, "grad_norm": 2.1687943935394287, "learning_rate": 0.00019596865032703339, "loss": 5.1367, "step": 467150 }, { "epoch": 2.8171392047852777, "grad_norm": 2.9535024166107178, "learning_rate": 0.00019594928004369717, "loss": 5.1021, "step": 467200 }, { "epoch": 2.8174406965666114, "grad_norm": 2.2190027236938477, "learning_rate": 0.00019592990891473006, "loss": 4.7914, "step": 467250 }, { "epoch": 2.8177421883479457, "grad_norm": 1.7613718509674072, "learning_rate": 0.00019591053694048845, "loss": 5.1191, "step": 467300 }, { "epoch": 2.8180436801292794, "grad_norm": 1.6074093580245972, "learning_rate": 0.00019589116412132898, "loss": 4.9147, "step": 467350 }, { "epoch": 2.8183451719106136, "grad_norm": 0.6556107997894287, "learning_rate": 0.00019587179045760814, "loss": 4.7995, "step": 467400 }, { "epoch": 2.818646663691948, "grad_norm": 1.4421862363815308, "learning_rate": 0.00019585241594968238, "loss": 5.024, "step": 467450 }, { "epoch": 2.818948155473282, "grad_norm": 1.406404733657837, "learning_rate": 0.00019583304059790838, "loss": 4.9667, "step": 467500 }, { "epoch": 2.819249647254616, "grad_norm": 2.0665876865386963, "learning_rate": 0.00019581366440264272, "loss": 4.8464, "step": 467550 }, { "epoch": 2.81955113903595, "grad_norm": 2.4770641326904297, "learning_rate": 0.0001957942873642419, "loss": 5.1191, "step": 467600 }, { "epoch": 2.8198526308172838, "grad_norm": 1.330724835395813, "learning_rate": 0.0001957749094830626, "loss": 4.8445, "step": 467650 }, { "epoch": 2.820154122598618, "grad_norm": 1.9264559745788574, "learning_rate": 0.00019575553075946138, "loss": 4.8809, "step": 467700 }, { "epoch": 2.820455614379952, "grad_norm": 1.7116671800613403, "learning_rate": 0.00019573615119379494, "loss": 5.0707, "step": 467750 }, { "epoch": 2.820757106161286, "grad_norm": 1.6353354454040527, "learning_rate": 0.0001957167707864199, "loss": 4.8292, "step": 467800 }, { "epoch": 2.82105859794262, "grad_norm": 1.5297921895980835, "learning_rate": 0.00019569738953769295, "loss": 5.6336, "step": 467850 }, { "epoch": 2.821360089723954, "grad_norm": 1.8086544275283813, "learning_rate": 0.0001956780074479707, "loss": 5.0666, "step": 467900 }, { "epoch": 2.821661581505288, "grad_norm": 1.6476154327392578, "learning_rate": 0.00019565862451761, "loss": 5.3449, "step": 467950 }, { "epoch": 2.8219630732866223, "grad_norm": 2.2404158115386963, "learning_rate": 0.00019563924074696742, "loss": 5.2049, "step": 468000 }, { "epoch": 2.8219630732866223, "eval_loss": 5.401710510253906, "eval_runtime": 39.2825, "eval_samples_per_second": 13.034, "eval_steps_per_second": 6.517, "eval_tts_loss": 7.407531946336335, "step": 468000 }, { "epoch": 2.822264565067956, "grad_norm": 1.8456650972366333, "learning_rate": 0.00019561985613639971, "loss": 4.9729, "step": 468050 }, { "epoch": 2.8225660568492903, "grad_norm": 1.4683157205581665, "learning_rate": 0.0001956004706862637, "loss": 4.6857, "step": 468100 }, { "epoch": 2.822867548630624, "grad_norm": 1.3044735193252563, "learning_rate": 0.00019558108439691612, "loss": 4.7221, "step": 468150 }, { "epoch": 2.8231690404119583, "grad_norm": 1.1612169742584229, "learning_rate": 0.00019556169726871368, "loss": 5.0816, "step": 468200 }, { "epoch": 2.8234705321932925, "grad_norm": 2.330173969268799, "learning_rate": 0.00019554230930201325, "loss": 4.9336, "step": 468250 }, { "epoch": 2.8237720239746267, "grad_norm": 1.825264573097229, "learning_rate": 0.0001955229204971716, "loss": 4.7909, "step": 468300 }, { "epoch": 2.8240735157559604, "grad_norm": 1.5112483501434326, "learning_rate": 0.00019550353085454557, "loss": 4.8839, "step": 468350 }, { "epoch": 2.8243750075372946, "grad_norm": 1.7192494869232178, "learning_rate": 0.000195484140374492, "loss": 5.3239, "step": 468400 }, { "epoch": 2.8246764993186284, "grad_norm": 2.309828281402588, "learning_rate": 0.00019546474905736768, "loss": 4.9821, "step": 468450 }, { "epoch": 2.8249779910999626, "grad_norm": 1.5918104648590088, "learning_rate": 0.0001954453569035296, "loss": 5.0642, "step": 468500 }, { "epoch": 2.825279482881297, "grad_norm": 1.696441411972046, "learning_rate": 0.0001954259639133346, "loss": 5.126, "step": 468550 }, { "epoch": 2.8255809746626306, "grad_norm": 1.9247394800186157, "learning_rate": 0.0001954065700871395, "loss": 4.6098, "step": 468600 }, { "epoch": 2.825882466443965, "grad_norm": 2.1662535667419434, "learning_rate": 0.00019538717542530128, "loss": 5.1514, "step": 468650 }, { "epoch": 2.8261839582252986, "grad_norm": 0.7370722889900208, "learning_rate": 0.0001953677799281769, "loss": 5.0316, "step": 468700 }, { "epoch": 2.8264854500066328, "grad_norm": 2.084470510482788, "learning_rate": 0.00019534838359612324, "loss": 4.6336, "step": 468750 }, { "epoch": 2.826786941787967, "grad_norm": 0.8066112995147705, "learning_rate": 0.00019532898642949731, "loss": 4.7505, "step": 468800 }, { "epoch": 2.827088433569301, "grad_norm": 1.6332868337631226, "learning_rate": 0.00019530958842865613, "loss": 5.2535, "step": 468850 }, { "epoch": 2.827389925350635, "grad_norm": 2.0056519508361816, "learning_rate": 0.00019529018959395654, "loss": 4.7803, "step": 468900 }, { "epoch": 2.827691417131969, "grad_norm": 1.7474428415298462, "learning_rate": 0.0001952707899257557, "loss": 5.1034, "step": 468950 }, { "epoch": 2.827992908913303, "grad_norm": 2.947114944458008, "learning_rate": 0.00019525138942441054, "loss": 5.1383, "step": 469000 }, { "epoch": 2.828294400694637, "grad_norm": 0.863851010799408, "learning_rate": 0.00019523198809027814, "loss": 5.0108, "step": 469050 }, { "epoch": 2.8285958924759713, "grad_norm": 1.9822648763656616, "learning_rate": 0.00019521258592371555, "loss": 5.3364, "step": 469100 }, { "epoch": 2.828897384257305, "grad_norm": 0.8086576461791992, "learning_rate": 0.0001951931829250799, "loss": 5.1872, "step": 469150 }, { "epoch": 2.8291988760386393, "grad_norm": 1.863338589668274, "learning_rate": 0.00019517377909472817, "loss": 4.8474, "step": 469200 }, { "epoch": 2.829500367819973, "grad_norm": 2.1334357261657715, "learning_rate": 0.00019515437443301748, "loss": 4.9925, "step": 469250 }, { "epoch": 2.8298018596013073, "grad_norm": 1.9630078077316284, "learning_rate": 0.000195134968940305, "loss": 5.0867, "step": 469300 }, { "epoch": 2.8301033513826415, "grad_norm": 1.8535816669464111, "learning_rate": 0.0001951155626169478, "loss": 5.2446, "step": 469350 }, { "epoch": 2.8304048431639752, "grad_norm": 1.5023236274719238, "learning_rate": 0.00019509615546330308, "loss": 5.25, "step": 469400 }, { "epoch": 2.8307063349453094, "grad_norm": 2.3563125133514404, "learning_rate": 0.00019507674747972796, "loss": 4.9673, "step": 469450 }, { "epoch": 2.831007826726643, "grad_norm": 1.9986774921417236, "learning_rate": 0.0001950573386665796, "loss": 5.2027, "step": 469500 }, { "epoch": 2.8313093185079774, "grad_norm": 2.1828646659851074, "learning_rate": 0.0001950379290242153, "loss": 5.0198, "step": 469550 }, { "epoch": 2.8316108102893116, "grad_norm": 2.0212724208831787, "learning_rate": 0.0001950185185529921, "loss": 5.4644, "step": 469600 }, { "epoch": 2.831912302070646, "grad_norm": 1.5607017278671265, "learning_rate": 0.00019499910725326734, "loss": 4.829, "step": 469650 }, { "epoch": 2.8322137938519796, "grad_norm": 1.6935901641845703, "learning_rate": 0.00019497969512539826, "loss": 4.7479, "step": 469700 }, { "epoch": 2.832515285633314, "grad_norm": 1.9420166015625, "learning_rate": 0.0001949602821697421, "loss": 4.9298, "step": 469750 }, { "epoch": 2.8328167774146475, "grad_norm": 2.216987371444702, "learning_rate": 0.00019494086838665601, "loss": 4.8128, "step": 469800 }, { "epoch": 2.8331182691959818, "grad_norm": 1.5589096546173096, "learning_rate": 0.00019492145377649744, "loss": 4.8829, "step": 469850 }, { "epoch": 2.833419760977316, "grad_norm": 1.8002575635910034, "learning_rate": 0.0001949020383396236, "loss": 5.17, "step": 469900 }, { "epoch": 2.8337212527586497, "grad_norm": 2.068063497543335, "learning_rate": 0.00019488262207639185, "loss": 5.0201, "step": 469950 }, { "epoch": 2.834022744539984, "grad_norm": 1.7819522619247437, "learning_rate": 0.0001948632049871595, "loss": 4.6168, "step": 470000 }, { "epoch": 2.8343242363213177, "grad_norm": 1.5868326425552368, "learning_rate": 0.0001948437870722838, "loss": 4.2452, "step": 470050 }, { "epoch": 2.834625728102652, "grad_norm": 1.4685903787612915, "learning_rate": 0.00019482436833212228, "loss": 5.1626, "step": 470100 }, { "epoch": 2.834927219883986, "grad_norm": 1.826907753944397, "learning_rate": 0.00019480494876703225, "loss": 5.0592, "step": 470150 }, { "epoch": 2.83522871166532, "grad_norm": 2.6646320819854736, "learning_rate": 0.000194785528377371, "loss": 4.7037, "step": 470200 }, { "epoch": 2.835530203446654, "grad_norm": 1.9434771537780762, "learning_rate": 0.00019476610716349604, "loss": 5.2358, "step": 470250 }, { "epoch": 2.835831695227988, "grad_norm": 2.7905564308166504, "learning_rate": 0.00019474668512576475, "loss": 5.198, "step": 470300 }, { "epoch": 2.836133187009322, "grad_norm": 1.3540351390838623, "learning_rate": 0.00019472726226453461, "loss": 4.8464, "step": 470350 }, { "epoch": 2.8364346787906562, "grad_norm": 1.0596897602081299, "learning_rate": 0.000194707838580163, "loss": 5.2394, "step": 470400 }, { "epoch": 2.8367361705719905, "grad_norm": 1.897362470626831, "learning_rate": 0.00019468841407300745, "loss": 4.8917, "step": 470450 }, { "epoch": 2.837037662353324, "grad_norm": 1.9580841064453125, "learning_rate": 0.00019466898874342538, "loss": 5.24, "step": 470500 }, { "epoch": 2.8373391541346584, "grad_norm": 2.0145645141601562, "learning_rate": 0.00019464956259177438, "loss": 4.6556, "step": 470550 }, { "epoch": 2.837640645915992, "grad_norm": 0.47705528140068054, "learning_rate": 0.0001946301356184119, "loss": 4.6623, "step": 470600 }, { "epoch": 2.8379421376973264, "grad_norm": 0.5626952648162842, "learning_rate": 0.0001946107078236954, "loss": 4.7417, "step": 470650 }, { "epoch": 2.8382436294786606, "grad_norm": 1.6847999095916748, "learning_rate": 0.0001945912792079825, "loss": 5.0117, "step": 470700 }, { "epoch": 2.8385451212599944, "grad_norm": 1.6644567251205444, "learning_rate": 0.0001945718497716308, "loss": 5.1015, "step": 470750 }, { "epoch": 2.8388466130413286, "grad_norm": 1.9549788236618042, "learning_rate": 0.00019455241951499774, "loss": 5.1735, "step": 470800 }, { "epoch": 2.8391481048226623, "grad_norm": 2.4037985801696777, "learning_rate": 0.000194532988438441, "loss": 4.8935, "step": 470850 }, { "epoch": 2.8394495966039965, "grad_norm": 0.9953063130378723, "learning_rate": 0.00019451355654231817, "loss": 4.7284, "step": 470900 }, { "epoch": 2.8397510883853307, "grad_norm": 1.9626022577285767, "learning_rate": 0.00019449412382698683, "loss": 5.3664, "step": 470950 }, { "epoch": 2.8400525801666645, "grad_norm": 1.5634663105010986, "learning_rate": 0.0001944746902928047, "loss": 4.9417, "step": 471000 }, { "epoch": 2.8400525801666645, "eval_loss": 5.394231796264648, "eval_runtime": 39.0759, "eval_samples_per_second": 13.103, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.373326017257156, "step": 471000 }, { "epoch": 2.8403540719479987, "grad_norm": 1.895806908607483, "learning_rate": 0.00019445525594012931, "loss": 4.6992, "step": 471050 }, { "epoch": 2.8406555637293325, "grad_norm": 2.3217999935150146, "learning_rate": 0.00019443582076931836, "loss": 5.0402, "step": 471100 }, { "epoch": 2.8409570555106667, "grad_norm": 1.671241044998169, "learning_rate": 0.0001944163847807296, "loss": 5.1736, "step": 471150 }, { "epoch": 2.841258547292001, "grad_norm": 1.7170509099960327, "learning_rate": 0.00019439694797472064, "loss": 5.2905, "step": 471200 }, { "epoch": 2.841560039073335, "grad_norm": 1.4296671152114868, "learning_rate": 0.00019437751035164918, "loss": 4.8233, "step": 471250 }, { "epoch": 2.841861530854669, "grad_norm": 1.7524505853652954, "learning_rate": 0.000194358071911873, "loss": 4.9899, "step": 471300 }, { "epoch": 2.842163022636003, "grad_norm": 1.872069001197815, "learning_rate": 0.00019433863265574985, "loss": 4.523, "step": 471350 }, { "epoch": 2.842464514417337, "grad_norm": 2.619621515274048, "learning_rate": 0.00019431919258363733, "loss": 4.712, "step": 471400 }, { "epoch": 2.842766006198671, "grad_norm": 2.0912880897521973, "learning_rate": 0.00019429975169589343, "loss": 4.8803, "step": 471450 }, { "epoch": 2.8430674979800052, "grad_norm": 1.515156865119934, "learning_rate": 0.00019428030999287573, "loss": 4.8236, "step": 471500 }, { "epoch": 2.843368989761339, "grad_norm": 2.781249761581421, "learning_rate": 0.0001942608674749422, "loss": 4.9841, "step": 471550 }, { "epoch": 2.843670481542673, "grad_norm": 2.1160340309143066, "learning_rate": 0.00019424142414245052, "loss": 5.4029, "step": 471600 }, { "epoch": 2.843971973324007, "grad_norm": 1.043164849281311, "learning_rate": 0.00019422197999575852, "loss": 5.1406, "step": 471650 }, { "epoch": 2.844273465105341, "grad_norm": 1.1219474077224731, "learning_rate": 0.00019420253503522414, "loss": 4.9844, "step": 471700 }, { "epoch": 2.8445749568866754, "grad_norm": 1.5144063234329224, "learning_rate": 0.00019418308926120523, "loss": 5.3492, "step": 471750 }, { "epoch": 2.844876448668009, "grad_norm": 2.815365791320801, "learning_rate": 0.00019416364267405955, "loss": 4.9446, "step": 471800 }, { "epoch": 2.8451779404493434, "grad_norm": 1.7267259359359741, "learning_rate": 0.00019414419527414507, "loss": 4.959, "step": 471850 }, { "epoch": 2.845479432230677, "grad_norm": 2.4548046588897705, "learning_rate": 0.00019412474706181966, "loss": 5.022, "step": 471900 }, { "epoch": 2.8457809240120113, "grad_norm": 2.3426640033721924, "learning_rate": 0.00019410529803744124, "loss": 5.2422, "step": 471950 }, { "epoch": 2.8460824157933455, "grad_norm": 1.7735979557037354, "learning_rate": 0.0001940858482013678, "loss": 5.3486, "step": 472000 }, { "epoch": 2.8463839075746797, "grad_norm": 2.029370069503784, "learning_rate": 0.00019406639755395718, "loss": 4.6359, "step": 472050 }, { "epoch": 2.8466853993560135, "grad_norm": 1.788291573524475, "learning_rate": 0.00019404694609556747, "loss": 4.801, "step": 472100 }, { "epoch": 2.8469868911373477, "grad_norm": 2.4668641090393066, "learning_rate": 0.00019402749382655651, "loss": 4.7582, "step": 472150 }, { "epoch": 2.8472883829186815, "grad_norm": 1.3363800048828125, "learning_rate": 0.0001940080407472824, "loss": 4.8948, "step": 472200 }, { "epoch": 2.8475898747000157, "grad_norm": 3.3136343955993652, "learning_rate": 0.00019398858685810305, "loss": 5.0245, "step": 472250 }, { "epoch": 2.84789136648135, "grad_norm": 1.3376410007476807, "learning_rate": 0.00019396913215937657, "loss": 5.0931, "step": 472300 }, { "epoch": 2.8481928582626836, "grad_norm": 2.220317840576172, "learning_rate": 0.000193949676651461, "loss": 5.1104, "step": 472350 }, { "epoch": 2.848494350044018, "grad_norm": 1.615308403968811, "learning_rate": 0.00019393022033471427, "loss": 5.1478, "step": 472400 }, { "epoch": 2.8487958418253516, "grad_norm": 1.480654239654541, "learning_rate": 0.00019391076320949454, "loss": 4.8547, "step": 472450 }, { "epoch": 2.849097333606686, "grad_norm": 0.9803133010864258, "learning_rate": 0.0001938913052761599, "loss": 4.7557, "step": 472500 }, { "epoch": 2.84939882538802, "grad_norm": 1.5904042720794678, "learning_rate": 0.00019387184653506847, "loss": 5.2951, "step": 472550 }, { "epoch": 2.8497003171693542, "grad_norm": 2.0797996520996094, "learning_rate": 0.00019385238698657825, "loss": 5.3938, "step": 472600 }, { "epoch": 2.850001808950688, "grad_norm": 2.012218475341797, "learning_rate": 0.00019383292663104746, "loss": 5.0302, "step": 472650 }, { "epoch": 2.850303300732022, "grad_norm": 1.6424082517623901, "learning_rate": 0.00019381346546883414, "loss": 5.3232, "step": 472700 }, { "epoch": 2.850604792513356, "grad_norm": 2.719108819961548, "learning_rate": 0.0001937940035002966, "loss": 4.8844, "step": 472750 }, { "epoch": 2.85090628429469, "grad_norm": 1.6994765996932983, "learning_rate": 0.00019377454072579288, "loss": 4.922, "step": 472800 }, { "epoch": 2.8512077760760244, "grad_norm": 1.9038591384887695, "learning_rate": 0.0001937550771456812, "loss": 5.2885, "step": 472850 }, { "epoch": 2.851509267857358, "grad_norm": 1.1233576536178589, "learning_rate": 0.0001937356127603198, "loss": 4.6533, "step": 472900 }, { "epoch": 2.8518107596386923, "grad_norm": 1.323151707649231, "learning_rate": 0.00019371614757006687, "loss": 5.0158, "step": 472950 }, { "epoch": 2.852112251420026, "grad_norm": 2.632918357849121, "learning_rate": 0.0001936966815752806, "loss": 5.0838, "step": 473000 }, { "epoch": 2.8524137432013603, "grad_norm": 1.7978202104568481, "learning_rate": 0.00019367721477631927, "loss": 4.999, "step": 473050 }, { "epoch": 2.8527152349826945, "grad_norm": 1.6906875371932983, "learning_rate": 0.00019365774717354113, "loss": 5.0033, "step": 473100 }, { "epoch": 2.8530167267640283, "grad_norm": 1.5790433883666992, "learning_rate": 0.00019363827876730448, "loss": 4.6873, "step": 473150 }, { "epoch": 2.8533182185453625, "grad_norm": 1.6186165809631348, "learning_rate": 0.00019361880955796757, "loss": 4.7168, "step": 473200 }, { "epoch": 2.8536197103266963, "grad_norm": 1.354364037513733, "learning_rate": 0.00019359933954588867, "loss": 5.2061, "step": 473250 }, { "epoch": 2.8539212021080305, "grad_norm": 1.7139695882797241, "learning_rate": 0.00019357986873142616, "loss": 4.6866, "step": 473300 }, { "epoch": 2.8542226938893647, "grad_norm": 1.019743800163269, "learning_rate": 0.00019356039711493844, "loss": 5.0049, "step": 473350 }, { "epoch": 2.854524185670699, "grad_norm": 1.7319034337997437, "learning_rate": 0.00019354092469678366, "loss": 4.9896, "step": 473400 }, { "epoch": 2.8548256774520326, "grad_norm": 2.549783945083618, "learning_rate": 0.0001935214514773204, "loss": 4.9525, "step": 473450 }, { "epoch": 2.855127169233367, "grad_norm": 1.5541554689407349, "learning_rate": 0.00019350197745690687, "loss": 5.3659, "step": 473500 }, { "epoch": 2.8554286610147006, "grad_norm": 3.2672667503356934, "learning_rate": 0.00019348250263590154, "loss": 4.9743, "step": 473550 }, { "epoch": 2.855730152796035, "grad_norm": 1.8165678977966309, "learning_rate": 0.0001934630270146628, "loss": 5.2293, "step": 473600 }, { "epoch": 2.856031644577369, "grad_norm": 2.2919578552246094, "learning_rate": 0.0001934435505935491, "loss": 5.4056, "step": 473650 }, { "epoch": 2.856333136358703, "grad_norm": 1.9576480388641357, "learning_rate": 0.00019342407337291883, "loss": 5.1961, "step": 473700 }, { "epoch": 2.856634628140037, "grad_norm": 2.0304248332977295, "learning_rate": 0.00019340459535313045, "loss": 5.2042, "step": 473750 }, { "epoch": 2.8569361199213708, "grad_norm": 1.2914626598358154, "learning_rate": 0.00019338511653454245, "loss": 5.17, "step": 473800 }, { "epoch": 2.857237611702705, "grad_norm": 1.919786810874939, "learning_rate": 0.00019336563691751325, "loss": 5.3237, "step": 473850 }, { "epoch": 2.857539103484039, "grad_norm": 2.109992504119873, "learning_rate": 0.00019334615650240145, "loss": 4.8958, "step": 473900 }, { "epoch": 2.857840595265373, "grad_norm": 1.154544472694397, "learning_rate": 0.00019332667528956552, "loss": 4.825, "step": 473950 }, { "epoch": 2.858142087046707, "grad_norm": 1.9027245044708252, "learning_rate": 0.00019330719327936388, "loss": 4.8043, "step": 474000 }, { "epoch": 2.858142087046707, "eval_loss": 5.398168563842773, "eval_runtime": 38.9072, "eval_samples_per_second": 13.16, "eval_steps_per_second": 6.58, "eval_tts_loss": 7.4764443565550085, "step": 474000 }, { "epoch": 2.858443578828041, "grad_norm": 2.8860039710998535, "learning_rate": 0.00019328771047215517, "loss": 5.033, "step": 474050 }, { "epoch": 2.858745070609375, "grad_norm": 1.4497008323669434, "learning_rate": 0.00019326822686829796, "loss": 5.0667, "step": 474100 }, { "epoch": 2.8590465623907093, "grad_norm": 3.242906332015991, "learning_rate": 0.00019324874246815077, "loss": 5.1952, "step": 474150 }, { "epoch": 2.8593480541720435, "grad_norm": 2.644341230392456, "learning_rate": 0.00019322925727207217, "loss": 4.8885, "step": 474200 }, { "epoch": 2.8596495459533773, "grad_norm": 1.3182265758514404, "learning_rate": 0.0001932097712804208, "loss": 5.0376, "step": 474250 }, { "epoch": 2.8599510377347115, "grad_norm": 1.5488239526748657, "learning_rate": 0.00019319028449355524, "loss": 5.2607, "step": 474300 }, { "epoch": 2.8602525295160453, "grad_norm": 2.8373923301696777, "learning_rate": 0.00019317079691183415, "loss": 4.9716, "step": 474350 }, { "epoch": 2.8605540212973795, "grad_norm": 1.645257592201233, "learning_rate": 0.00019315130853561616, "loss": 4.4119, "step": 474400 }, { "epoch": 2.8608555130787137, "grad_norm": 2.5519492626190186, "learning_rate": 0.00019313181936525984, "loss": 4.9509, "step": 474450 }, { "epoch": 2.8611570048600474, "grad_norm": 2.8405699729919434, "learning_rate": 0.00019311232940112398, "loss": 5.027, "step": 474500 }, { "epoch": 2.8614584966413816, "grad_norm": 2.441314220428467, "learning_rate": 0.0001930928386435673, "loss": 5.2053, "step": 474550 }, { "epoch": 2.8617599884227154, "grad_norm": 2.641195058822632, "learning_rate": 0.0001930733470929483, "loss": 4.8624, "step": 474600 }, { "epoch": 2.8620614802040496, "grad_norm": 1.9728834629058838, "learning_rate": 0.00019305385474962587, "loss": 5.0352, "step": 474650 }, { "epoch": 2.862362971985384, "grad_norm": 1.4432753324508667, "learning_rate": 0.00019303436161395872, "loss": 4.7902, "step": 474700 }, { "epoch": 2.8626644637667176, "grad_norm": 2.0544943809509277, "learning_rate": 0.00019301486768630552, "loss": 4.8832, "step": 474750 }, { "epoch": 2.8629659555480518, "grad_norm": 2.262230157852173, "learning_rate": 0.00019299537296702508, "loss": 4.974, "step": 474800 }, { "epoch": 2.8632674473293855, "grad_norm": 1.8517663478851318, "learning_rate": 0.00019297587745647615, "loss": 5.1109, "step": 474850 }, { "epoch": 2.8635689391107197, "grad_norm": 2.4139621257781982, "learning_rate": 0.0001929563811550175, "loss": 4.8906, "step": 474900 }, { "epoch": 2.863870430892054, "grad_norm": 1.4419121742248535, "learning_rate": 0.000192936884063008, "loss": 4.9011, "step": 474950 }, { "epoch": 2.864171922673388, "grad_norm": 1.8487756252288818, "learning_rate": 0.00019291738618080638, "loss": 5.2074, "step": 475000 }, { "epoch": 2.864473414454722, "grad_norm": 1.427270531654358, "learning_rate": 0.00019289788750877153, "loss": 5.1286, "step": 475050 }, { "epoch": 2.864774906236056, "grad_norm": 2.219270706176758, "learning_rate": 0.0001928783880472623, "loss": 4.825, "step": 475100 }, { "epoch": 2.86507639801739, "grad_norm": 2.5090060234069824, "learning_rate": 0.00019285888779663752, "loss": 4.9267, "step": 475150 }, { "epoch": 2.865377889798724, "grad_norm": 2.3157711029052734, "learning_rate": 0.00019283938675725606, "loss": 4.7781, "step": 475200 }, { "epoch": 2.8656793815800583, "grad_norm": 3.0881407260894775, "learning_rate": 0.0001928198849294768, "loss": 5.1685, "step": 475250 }, { "epoch": 2.865980873361392, "grad_norm": 2.274646520614624, "learning_rate": 0.00019280038231365873, "loss": 4.954, "step": 475300 }, { "epoch": 2.8662823651427263, "grad_norm": 0.8683888912200928, "learning_rate": 0.0001927808789101607, "loss": 4.5072, "step": 475350 }, { "epoch": 2.86658385692406, "grad_norm": 1.8207250833511353, "learning_rate": 0.0001927613747193416, "loss": 5.1485, "step": 475400 }, { "epoch": 2.8668853487053942, "grad_norm": 1.6341674327850342, "learning_rate": 0.0001927418697415604, "loss": 4.6992, "step": 475450 }, { "epoch": 2.8671868404867284, "grad_norm": 1.737189769744873, "learning_rate": 0.00019272236397717613, "loss": 5.1159, "step": 475500 }, { "epoch": 2.867488332268062, "grad_norm": 2.5933921337127686, "learning_rate": 0.00019270285742654773, "loss": 4.7591, "step": 475550 }, { "epoch": 2.8677898240493964, "grad_norm": 1.637286901473999, "learning_rate": 0.00019268335009003414, "loss": 4.727, "step": 475600 }, { "epoch": 2.86809131583073, "grad_norm": 2.4721670150756836, "learning_rate": 0.00019266384196799444, "loss": 5.0304, "step": 475650 }, { "epoch": 2.8683928076120644, "grad_norm": 2.2680294513702393, "learning_rate": 0.00019264433306078756, "loss": 5.0532, "step": 475700 }, { "epoch": 2.8686942993933986, "grad_norm": 1.3043252229690552, "learning_rate": 0.00019262482336877265, "loss": 4.6745, "step": 475750 }, { "epoch": 2.868995791174733, "grad_norm": 1.6976830959320068, "learning_rate": 0.00019260531289230865, "loss": 4.6124, "step": 475800 }, { "epoch": 2.8692972829560666, "grad_norm": 1.5751159191131592, "learning_rate": 0.0001925858016317547, "loss": 4.8171, "step": 475850 }, { "epoch": 2.8695987747374008, "grad_norm": 2.7163496017456055, "learning_rate": 0.0001925662895874698, "loss": 5.1242, "step": 475900 }, { "epoch": 2.8699002665187345, "grad_norm": 1.1597929000854492, "learning_rate": 0.00019254677675981314, "loss": 4.6712, "step": 475950 }, { "epoch": 2.8702017583000687, "grad_norm": 2.2651822566986084, "learning_rate": 0.00019252726314914376, "loss": 4.8586, "step": 476000 }, { "epoch": 2.870503250081403, "grad_norm": 1.1617302894592285, "learning_rate": 0.0001925077487558207, "loss": 5.1933, "step": 476050 }, { "epoch": 2.8708047418627367, "grad_norm": 1.996505856513977, "learning_rate": 0.0001924882335802033, "loss": 5.0773, "step": 476100 }, { "epoch": 2.871106233644071, "grad_norm": 1.9438215494155884, "learning_rate": 0.0001924687176226506, "loss": 5.0524, "step": 476150 }, { "epoch": 2.8714077254254047, "grad_norm": 2.7506701946258545, "learning_rate": 0.00019244920088352165, "loss": 4.8145, "step": 476200 }, { "epoch": 2.871709217206739, "grad_norm": 3.3534111976623535, "learning_rate": 0.00019242968336317582, "loss": 4.9471, "step": 476250 }, { "epoch": 2.872010708988073, "grad_norm": 2.2204091548919678, "learning_rate": 0.0001924101650619722, "loss": 4.9176, "step": 476300 }, { "epoch": 2.8723122007694073, "grad_norm": 1.9824039936065674, "learning_rate": 0.00019239064598027003, "loss": 5.3953, "step": 476350 }, { "epoch": 2.872613692550741, "grad_norm": 1.6848865747451782, "learning_rate": 0.0001923711261184285, "loss": 4.4306, "step": 476400 }, { "epoch": 2.8729151843320753, "grad_norm": 1.702459692955017, "learning_rate": 0.00019235160547680685, "loss": 5.1115, "step": 476450 }, { "epoch": 2.873216676113409, "grad_norm": 2.4669547080993652, "learning_rate": 0.00019233208405576436, "loss": 5.2539, "step": 476500 }, { "epoch": 2.8735181678947432, "grad_norm": 1.5676000118255615, "learning_rate": 0.00019231256185566026, "loss": 4.9678, "step": 476550 }, { "epoch": 2.8738196596760774, "grad_norm": 2.3769147396087646, "learning_rate": 0.00019229303887685384, "loss": 4.3643, "step": 476600 }, { "epoch": 2.874121151457411, "grad_norm": 1.7474054098129272, "learning_rate": 0.00019227351511970437, "loss": 4.6908, "step": 476650 }, { "epoch": 2.8744226432387454, "grad_norm": 1.4378328323364258, "learning_rate": 0.00019225399058457122, "loss": 4.9997, "step": 476700 }, { "epoch": 2.874724135020079, "grad_norm": 1.59762704372406, "learning_rate": 0.00019223446527181367, "loss": 5.3336, "step": 476750 }, { "epoch": 2.8750256268014134, "grad_norm": 1.9427677392959595, "learning_rate": 0.00019221493918179105, "loss": 4.9271, "step": 476800 }, { "epoch": 2.8753271185827476, "grad_norm": 1.6359440088272095, "learning_rate": 0.00019219541231486274, "loss": 4.8949, "step": 476850 }, { "epoch": 2.8756286103640814, "grad_norm": 1.7512835264205933, "learning_rate": 0.00019217588467138804, "loss": 5.1538, "step": 476900 }, { "epoch": 2.8759301021454156, "grad_norm": 1.9830390214920044, "learning_rate": 0.00019215635625172644, "loss": 5.1053, "step": 476950 }, { "epoch": 2.8762315939267493, "grad_norm": 2.50164532661438, "learning_rate": 0.00019213682705623722, "loss": 5.2921, "step": 477000 }, { "epoch": 2.8762315939267493, "eval_loss": 5.385182857513428, "eval_runtime": 38.9945, "eval_samples_per_second": 13.13, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.446491724381321, "step": 477000 }, { "epoch": 2.8765330857080835, "grad_norm": 2.106898784637451, "learning_rate": 0.00019211729708527977, "loss": 4.6816, "step": 477050 }, { "epoch": 2.8768345774894177, "grad_norm": 2.0909574031829834, "learning_rate": 0.00019209776633921367, "loss": 5.076, "step": 477100 }, { "epoch": 2.877136069270752, "grad_norm": 1.700556755065918, "learning_rate": 0.00019207823481839822, "loss": 4.9505, "step": 477150 }, { "epoch": 2.8774375610520857, "grad_norm": 2.6020262241363525, "learning_rate": 0.00019205870252319288, "loss": 4.9088, "step": 477200 }, { "epoch": 2.87773905283342, "grad_norm": 1.747480034828186, "learning_rate": 0.00019203916945395718, "loss": 4.979, "step": 477250 }, { "epoch": 2.8780405446147537, "grad_norm": 1.756088137626648, "learning_rate": 0.00019201963561105053, "loss": 5.2921, "step": 477300 }, { "epoch": 2.878342036396088, "grad_norm": 2.1908953189849854, "learning_rate": 0.00019200010099483248, "loss": 5.0993, "step": 477350 }, { "epoch": 2.878643528177422, "grad_norm": 1.9987667798995972, "learning_rate": 0.0001919805656056625, "loss": 5.0178, "step": 477400 }, { "epoch": 2.878945019958756, "grad_norm": 1.9269499778747559, "learning_rate": 0.0001919610294439001, "loss": 5.2933, "step": 477450 }, { "epoch": 2.87924651174009, "grad_norm": 1.7044168710708618, "learning_rate": 0.0001919414925099049, "loss": 4.7241, "step": 477500 }, { "epoch": 2.879548003521424, "grad_norm": 2.2069413661956787, "learning_rate": 0.0001919219548040363, "loss": 4.9896, "step": 477550 }, { "epoch": 2.879849495302758, "grad_norm": 1.773903250694275, "learning_rate": 0.00019190241632665399, "loss": 4.8019, "step": 477600 }, { "epoch": 2.8801509870840922, "grad_norm": 2.2015810012817383, "learning_rate": 0.00019188287707811748, "loss": 5.0883, "step": 477650 }, { "epoch": 2.880452478865426, "grad_norm": 1.5415898561477661, "learning_rate": 0.00019186333705878642, "loss": 5.4375, "step": 477700 }, { "epoch": 2.88075397064676, "grad_norm": 1.0090874433517456, "learning_rate": 0.00019184379626902036, "loss": 5.0609, "step": 477750 }, { "epoch": 2.881055462428094, "grad_norm": 1.715744137763977, "learning_rate": 0.00019182425470917896, "loss": 4.9987, "step": 477800 }, { "epoch": 2.881356954209428, "grad_norm": 2.2330322265625, "learning_rate": 0.00019180471237962183, "loss": 5.1202, "step": 477850 }, { "epoch": 2.8816584459907624, "grad_norm": 1.6315714120864868, "learning_rate": 0.00019178516928070863, "loss": 5.1018, "step": 477900 }, { "epoch": 2.8819599377720966, "grad_norm": 2.7325055599212646, "learning_rate": 0.00019176562541279904, "loss": 4.8114, "step": 477950 }, { "epoch": 2.8822614295534303, "grad_norm": 2.067218780517578, "learning_rate": 0.00019174608077625267, "loss": 4.7315, "step": 478000 }, { "epoch": 2.8825629213347645, "grad_norm": 1.8782119750976562, "learning_rate": 0.0001917265353714293, "loss": 5.1196, "step": 478050 }, { "epoch": 2.8828644131160983, "grad_norm": 1.7266430854797363, "learning_rate": 0.00019170698919868858, "loss": 4.8784, "step": 478100 }, { "epoch": 2.8831659048974325, "grad_norm": 3.646463394165039, "learning_rate": 0.00019168744225839024, "loss": 4.7589, "step": 478150 }, { "epoch": 2.8834673966787667, "grad_norm": 2.8479321002960205, "learning_rate": 0.000191667894550894, "loss": 5.0697, "step": 478200 }, { "epoch": 2.8837688884601005, "grad_norm": 1.7232869863510132, "learning_rate": 0.00019164834607655963, "loss": 4.9152, "step": 478250 }, { "epoch": 2.8840703802414347, "grad_norm": 1.6717517375946045, "learning_rate": 0.0001916287968357469, "loss": 4.7737, "step": 478300 }, { "epoch": 2.8843718720227685, "grad_norm": 2.7943971157073975, "learning_rate": 0.00019160924682881558, "loss": 4.7417, "step": 478350 }, { "epoch": 2.8846733638041027, "grad_norm": 1.038122534751892, "learning_rate": 0.00019158969605612543, "loss": 4.9802, "step": 478400 }, { "epoch": 2.884974855585437, "grad_norm": 1.8465229272842407, "learning_rate": 0.00019157014451803627, "loss": 5.1645, "step": 478450 }, { "epoch": 2.8852763473667706, "grad_norm": 2.747077226638794, "learning_rate": 0.00019155059221490794, "loss": 5.0322, "step": 478500 }, { "epoch": 2.885577839148105, "grad_norm": 1.8336166143417358, "learning_rate": 0.00019153103914710027, "loss": 5.1981, "step": 478550 }, { "epoch": 2.8858793309294386, "grad_norm": 1.9098206758499146, "learning_rate": 0.00019151148531497308, "loss": 4.7764, "step": 478600 }, { "epoch": 2.886180822710773, "grad_norm": 0.6024300456047058, "learning_rate": 0.0001914919307188862, "loss": 4.6946, "step": 478650 }, { "epoch": 2.886482314492107, "grad_norm": 0.826980471611023, "learning_rate": 0.00019147237535919965, "loss": 5.1266, "step": 478700 }, { "epoch": 2.886783806273441, "grad_norm": 0.6590956449508667, "learning_rate": 0.00019145281923627314, "loss": 5.0165, "step": 478750 }, { "epoch": 2.887085298054775, "grad_norm": 1.5554648637771606, "learning_rate": 0.00019143326235046665, "loss": 4.8501, "step": 478800 }, { "epoch": 2.887386789836109, "grad_norm": 1.603275179862976, "learning_rate": 0.00019141370470214007, "loss": 4.8219, "step": 478850 }, { "epoch": 2.887688281617443, "grad_norm": 2.143566370010376, "learning_rate": 0.00019139414629165344, "loss": 5.2399, "step": 478900 }, { "epoch": 2.887989773398777, "grad_norm": 3.3842663764953613, "learning_rate": 0.00019137458711936657, "loss": 4.6502, "step": 478950 }, { "epoch": 2.8882912651801114, "grad_norm": 0.9821599125862122, "learning_rate": 0.00019135502718563947, "loss": 4.7485, "step": 479000 }, { "epoch": 2.888592756961445, "grad_norm": 1.9016624689102173, "learning_rate": 0.00019133546649083212, "loss": 4.7571, "step": 479050 }, { "epoch": 2.8888942487427793, "grad_norm": 1.6233134269714355, "learning_rate": 0.0001913159050353045, "loss": 4.8662, "step": 479100 }, { "epoch": 2.889195740524113, "grad_norm": 2.4893746376037598, "learning_rate": 0.0001912963428194166, "loss": 5.2373, "step": 479150 }, { "epoch": 2.8894972323054473, "grad_norm": 1.6621942520141602, "learning_rate": 0.00019127677984352842, "loss": 5.2158, "step": 479200 }, { "epoch": 2.8897987240867815, "grad_norm": 3.1068315505981445, "learning_rate": 0.000191257216108, "loss": 5.1999, "step": 479250 }, { "epoch": 2.8901002158681153, "grad_norm": 2.1153879165649414, "learning_rate": 0.0001912376516131915, "loss": 5.1391, "step": 479300 }, { "epoch": 2.8904017076494495, "grad_norm": 0.9436885118484497, "learning_rate": 0.00019121808635946282, "loss": 4.9043, "step": 479350 }, { "epoch": 2.8907031994307832, "grad_norm": 1.4762805700302124, "learning_rate": 0.00019119852034717402, "loss": 5.2772, "step": 479400 }, { "epoch": 2.8910046912121175, "grad_norm": 1.7850704193115234, "learning_rate": 0.00019117895357668527, "loss": 5.1837, "step": 479450 }, { "epoch": 2.8913061829934517, "grad_norm": 1.4944790601730347, "learning_rate": 0.00019115938604835666, "loss": 5.0972, "step": 479500 }, { "epoch": 2.891607674774786, "grad_norm": 2.5575613975524902, "learning_rate": 0.00019113981776254832, "loss": 4.6797, "step": 479550 }, { "epoch": 2.8919091665561196, "grad_norm": 1.7089884281158447, "learning_rate": 0.0001911202487196203, "loss": 5.0386, "step": 479600 }, { "epoch": 2.892210658337454, "grad_norm": 2.052910804748535, "learning_rate": 0.0001911006789199328, "loss": 5.0064, "step": 479650 }, { "epoch": 2.8925121501187876, "grad_norm": 1.7359645366668701, "learning_rate": 0.00019108110836384595, "loss": 4.7864, "step": 479700 }, { "epoch": 2.892813641900122, "grad_norm": 2.0742623805999756, "learning_rate": 0.00019106153705171996, "loss": 5.0905, "step": 479750 }, { "epoch": 2.893115133681456, "grad_norm": 2.01440167427063, "learning_rate": 0.00019104196498391496, "loss": 4.8636, "step": 479800 }, { "epoch": 2.8934166254627898, "grad_norm": 3.6668176651000977, "learning_rate": 0.00019102239216079112, "loss": 5.1718, "step": 479850 }, { "epoch": 2.893718117244124, "grad_norm": 3.3173553943634033, "learning_rate": 0.00019100281858270876, "loss": 4.5744, "step": 479900 }, { "epoch": 2.8940196090254577, "grad_norm": 1.5287050008773804, "learning_rate": 0.000190983244250028, "loss": 5.1313, "step": 479950 }, { "epoch": 2.894321100806792, "grad_norm": 2.1818954944610596, "learning_rate": 0.00019096366916310911, "loss": 4.8883, "step": 480000 }, { "epoch": 2.894321100806792, "eval_loss": 5.400247573852539, "eval_runtime": 39.0565, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.410042257858688, "step": 480000 }, { "epoch": 2.894622592588126, "grad_norm": 2.082535743713379, "learning_rate": 0.00019094409332231238, "loss": 4.6168, "step": 480050 }, { "epoch": 2.8949240843694604, "grad_norm": 2.0688045024871826, "learning_rate": 0.000190924516727998, "loss": 5.0931, "step": 480100 }, { "epoch": 2.895225576150794, "grad_norm": 1.5303621292114258, "learning_rate": 0.00019090493938052634, "loss": 4.9926, "step": 480150 }, { "epoch": 2.8955270679321283, "grad_norm": 1.6067198514938354, "learning_rate": 0.00019088536128025762, "loss": 5.056, "step": 480200 }, { "epoch": 2.895828559713462, "grad_norm": 1.383793592453003, "learning_rate": 0.00019086578242755216, "loss": 4.6983, "step": 480250 }, { "epoch": 2.8961300514947963, "grad_norm": 1.5781927108764648, "learning_rate": 0.00019084620282277032, "loss": 5.1063, "step": 480300 }, { "epoch": 2.8964315432761305, "grad_norm": 2.534782886505127, "learning_rate": 0.0001908266224662724, "loss": 4.7564, "step": 480350 }, { "epoch": 2.8967330350574643, "grad_norm": 2.212531566619873, "learning_rate": 0.00019080704135841876, "loss": 4.974, "step": 480400 }, { "epoch": 2.8970345268387985, "grad_norm": 1.5196329355239868, "learning_rate": 0.00019078745949956973, "loss": 4.7815, "step": 480450 }, { "epoch": 2.8973360186201322, "grad_norm": 1.4207231998443604, "learning_rate": 0.00019076787689008574, "loss": 5.0907, "step": 480500 }, { "epoch": 2.8976375104014664, "grad_norm": 1.3665940761566162, "learning_rate": 0.00019074829353032716, "loss": 5.275, "step": 480550 }, { "epoch": 2.8979390021828006, "grad_norm": 2.820034980773926, "learning_rate": 0.00019072870942065436, "loss": 4.7484, "step": 480600 }, { "epoch": 2.8982404939641344, "grad_norm": 1.6713865995407104, "learning_rate": 0.00019070912456142783, "loss": 4.9449, "step": 480650 }, { "epoch": 2.8985419857454686, "grad_norm": 2.1215908527374268, "learning_rate": 0.00019068953895300794, "loss": 4.6757, "step": 480700 }, { "epoch": 2.8988434775268024, "grad_norm": 2.202455997467041, "learning_rate": 0.00019066995259575513, "loss": 4.741, "step": 480750 }, { "epoch": 2.8991449693081366, "grad_norm": 1.5203394889831543, "learning_rate": 0.00019065036549002992, "loss": 5.2045, "step": 480800 }, { "epoch": 2.899446461089471, "grad_norm": 1.184537649154663, "learning_rate": 0.00019063077763619267, "loss": 4.6554, "step": 480850 }, { "epoch": 2.899747952870805, "grad_norm": 2.5266005992889404, "learning_rate": 0.00019061118903460406, "loss": 5.1618, "step": 480900 }, { "epoch": 2.9000494446521388, "grad_norm": 1.8485585451126099, "learning_rate": 0.00019059159968562442, "loss": 4.6617, "step": 480950 }, { "epoch": 2.900350936433473, "grad_norm": 1.0900704860687256, "learning_rate": 0.00019057200958961422, "loss": 5.3023, "step": 481000 }, { "epoch": 2.9006524282148067, "grad_norm": 0.9786226153373718, "learning_rate": 0.00019055241874693418, "loss": 4.9674, "step": 481050 }, { "epoch": 2.900953919996141, "grad_norm": 2.078469753265381, "learning_rate": 0.00019053282715794474, "loss": 5.0224, "step": 481100 }, { "epoch": 2.901255411777475, "grad_norm": 2.4775586128234863, "learning_rate": 0.00019051323482300645, "loss": 5.0107, "step": 481150 }, { "epoch": 2.901556903558809, "grad_norm": 0.9055001139640808, "learning_rate": 0.0001904936417424799, "loss": 4.9225, "step": 481200 }, { "epoch": 2.901858395340143, "grad_norm": 3.4847254753112793, "learning_rate": 0.00019047404791672565, "loss": 4.9121, "step": 481250 }, { "epoch": 2.902159887121477, "grad_norm": 2.177241086959839, "learning_rate": 0.00019045445334610428, "loss": 4.947, "step": 481300 }, { "epoch": 2.902461378902811, "grad_norm": 1.6005479097366333, "learning_rate": 0.00019043485803097647, "loss": 4.752, "step": 481350 }, { "epoch": 2.9027628706841453, "grad_norm": 1.5152446031570435, "learning_rate": 0.00019041526197170275, "loss": 5.1494, "step": 481400 }, { "epoch": 2.903064362465479, "grad_norm": 1.5253952741622925, "learning_rate": 0.00019039566516864382, "loss": 4.897, "step": 481450 }, { "epoch": 2.9033658542468133, "grad_norm": 1.9745053052902222, "learning_rate": 0.0001903760676221604, "loss": 5.1529, "step": 481500 }, { "epoch": 2.903667346028147, "grad_norm": 1.4712119102478027, "learning_rate": 0.000190356469332613, "loss": 5.225, "step": 481550 }, { "epoch": 2.9039688378094812, "grad_norm": 1.4676750898361206, "learning_rate": 0.00019033687030036236, "loss": 4.7966, "step": 481600 }, { "epoch": 2.9042703295908154, "grad_norm": 0.9694355130195618, "learning_rate": 0.00019031727052576923, "loss": 4.8091, "step": 481650 }, { "epoch": 2.9045718213721496, "grad_norm": 2.3414547443389893, "learning_rate": 0.00019029767000919426, "loss": 5.1684, "step": 481700 }, { "epoch": 2.9048733131534834, "grad_norm": 1.5638476610183716, "learning_rate": 0.0001902780687509982, "loss": 4.8993, "step": 481750 }, { "epoch": 2.9051748049348176, "grad_norm": 1.8129833936691284, "learning_rate": 0.00019025846675154172, "loss": 5.0491, "step": 481800 }, { "epoch": 2.9054762967161514, "grad_norm": 1.917194128036499, "learning_rate": 0.00019023886401118559, "loss": 4.856, "step": 481850 }, { "epoch": 2.9057777884974856, "grad_norm": 2.065265417098999, "learning_rate": 0.0001902192605302907, "loss": 4.4538, "step": 481900 }, { "epoch": 2.90607928027882, "grad_norm": 0.8257079720497131, "learning_rate": 0.00019019965630921764, "loss": 4.7643, "step": 481950 }, { "epoch": 2.9063807720601536, "grad_norm": 2.303701162338257, "learning_rate": 0.00019018005134832728, "loss": 4.7446, "step": 482000 }, { "epoch": 2.9066822638414878, "grad_norm": 2.261807918548584, "learning_rate": 0.00019016044564798041, "loss": 4.7785, "step": 482050 }, { "epoch": 2.9069837556228215, "grad_norm": 2.426375150680542, "learning_rate": 0.00019014083920853789, "loss": 5.1659, "step": 482100 }, { "epoch": 2.9072852474041557, "grad_norm": 1.461344838142395, "learning_rate": 0.00019012123203036047, "loss": 5.0667, "step": 482150 }, { "epoch": 2.90758673918549, "grad_norm": 2.0153934955596924, "learning_rate": 0.00019010162411380903, "loss": 4.6717, "step": 482200 }, { "epoch": 2.9078882309668237, "grad_norm": 0.8898378014564514, "learning_rate": 0.00019008201545924443, "loss": 4.8616, "step": 482250 }, { "epoch": 2.908189722748158, "grad_norm": 1.497856616973877, "learning_rate": 0.00019006240606702756, "loss": 5.1354, "step": 482300 }, { "epoch": 2.9084912145294917, "grad_norm": 0.9623481035232544, "learning_rate": 0.00019004279593751928, "loss": 4.7462, "step": 482350 }, { "epoch": 2.908792706310826, "grad_norm": 1.9821149110794067, "learning_rate": 0.00019002318507108044, "loss": 4.7426, "step": 482400 }, { "epoch": 2.90909419809216, "grad_norm": 2.3722894191741943, "learning_rate": 0.000190003573468072, "loss": 5.1325, "step": 482450 }, { "epoch": 2.9093956898734943, "grad_norm": 1.7508102655410767, "learning_rate": 0.00018998396112885493, "loss": 4.8804, "step": 482500 }, { "epoch": 2.909697181654828, "grad_norm": 2.35786771774292, "learning_rate": 0.0001899643480537901, "loss": 5.0725, "step": 482550 }, { "epoch": 2.9099986734361623, "grad_norm": 1.854358434677124, "learning_rate": 0.00018994473424323842, "loss": 5.4055, "step": 482600 }, { "epoch": 2.910300165217496, "grad_norm": 1.3739396333694458, "learning_rate": 0.00018992511969756094, "loss": 5.2955, "step": 482650 }, { "epoch": 2.9106016569988302, "grad_norm": 2.7511820793151855, "learning_rate": 0.00018990550441711865, "loss": 4.6536, "step": 482700 }, { "epoch": 2.9109031487801644, "grad_norm": 1.2962377071380615, "learning_rate": 0.00018988588840227247, "loss": 4.7061, "step": 482750 }, { "epoch": 2.911204640561498, "grad_norm": 2.4324522018432617, "learning_rate": 0.00018986627165338342, "loss": 4.8024, "step": 482800 }, { "epoch": 2.9115061323428324, "grad_norm": 2.2515909671783447, "learning_rate": 0.00018984665417081254, "loss": 4.9005, "step": 482850 }, { "epoch": 2.911807624124166, "grad_norm": 2.294255018234253, "learning_rate": 0.00018982703595492087, "loss": 4.8869, "step": 482900 }, { "epoch": 2.9121091159055004, "grad_norm": 1.3432999849319458, "learning_rate": 0.00018980741700606944, "loss": 5.1478, "step": 482950 }, { "epoch": 2.9124106076868346, "grad_norm": 2.4148731231689453, "learning_rate": 0.0001897877973246193, "loss": 4.8489, "step": 483000 }, { "epoch": 2.9124106076868346, "eval_loss": 5.385049819946289, "eval_runtime": 39.0845, "eval_samples_per_second": 13.1, "eval_steps_per_second": 6.55, "eval_tts_loss": 7.488795788138605, "step": 483000 }, { "epoch": 2.9127120994681683, "grad_norm": 2.0022804737091064, "learning_rate": 0.00018976817691093148, "loss": 5.0416, "step": 483050 }, { "epoch": 2.9130135912495025, "grad_norm": 2.2647268772125244, "learning_rate": 0.0001897485557653672, "loss": 5.5904, "step": 483100 }, { "epoch": 2.9133150830308363, "grad_norm": 1.5706912279129028, "learning_rate": 0.00018972893388828746, "loss": 5.0634, "step": 483150 }, { "epoch": 2.9136165748121705, "grad_norm": 2.2178149223327637, "learning_rate": 0.00018970931128005334, "loss": 4.8046, "step": 483200 }, { "epoch": 2.9139180665935047, "grad_norm": 1.957242727279663, "learning_rate": 0.00018968968794102605, "loss": 4.8802, "step": 483250 }, { "epoch": 2.914219558374839, "grad_norm": 2.547891139984131, "learning_rate": 0.0001896700638715667, "loss": 5.027, "step": 483300 }, { "epoch": 2.9145210501561727, "grad_norm": 1.7500287294387817, "learning_rate": 0.0001896504390720364, "loss": 5.292, "step": 483350 }, { "epoch": 2.914822541937507, "grad_norm": 1.5089895725250244, "learning_rate": 0.00018963081354279637, "loss": 5.135, "step": 483400 }, { "epoch": 2.9151240337188407, "grad_norm": 0.7871612310409546, "learning_rate": 0.00018961118728420777, "loss": 4.6864, "step": 483450 }, { "epoch": 2.915425525500175, "grad_norm": 3.2102487087249756, "learning_rate": 0.00018959156029663184, "loss": 4.9002, "step": 483500 }, { "epoch": 2.915727017281509, "grad_norm": 1.4784631729125977, "learning_rate": 0.00018957193258042971, "loss": 4.8402, "step": 483550 }, { "epoch": 2.916028509062843, "grad_norm": 1.6630312204360962, "learning_rate": 0.0001895523041359626, "loss": 4.9535, "step": 483600 }, { "epoch": 2.916330000844177, "grad_norm": 2.1145870685577393, "learning_rate": 0.0001895326749635918, "loss": 4.9048, "step": 483650 }, { "epoch": 2.916631492625511, "grad_norm": 0.49863260984420776, "learning_rate": 0.00018951304506367858, "loss": 4.742, "step": 483700 }, { "epoch": 2.916932984406845, "grad_norm": 1.4409345388412476, "learning_rate": 0.00018949341443658411, "loss": 4.6338, "step": 483750 }, { "epoch": 2.917234476188179, "grad_norm": 1.6813561916351318, "learning_rate": 0.0001894737830826697, "loss": 4.5654, "step": 483800 }, { "epoch": 2.9175359679695134, "grad_norm": 2.018183469772339, "learning_rate": 0.00018945415100229666, "loss": 5.0326, "step": 483850 }, { "epoch": 2.917837459750847, "grad_norm": 1.6071871519088745, "learning_rate": 0.00018943451819582624, "loss": 4.7328, "step": 483900 }, { "epoch": 2.9181389515321814, "grad_norm": 1.090213418006897, "learning_rate": 0.00018941488466361984, "loss": 4.3207, "step": 483950 }, { "epoch": 2.918440443313515, "grad_norm": 1.9646947383880615, "learning_rate": 0.0001893952504060387, "loss": 5.0539, "step": 484000 }, { "epoch": 2.9187419350948494, "grad_norm": 1.744284749031067, "learning_rate": 0.0001893756154234442, "loss": 4.6378, "step": 484050 }, { "epoch": 2.9190434268761836, "grad_norm": 1.4990849494934082, "learning_rate": 0.00018935597971619775, "loss": 5.111, "step": 484100 }, { "epoch": 2.9193449186575173, "grad_norm": 2.388195514678955, "learning_rate": 0.0001893363432846606, "loss": 4.6944, "step": 484150 }, { "epoch": 2.9196464104388515, "grad_norm": 2.3664708137512207, "learning_rate": 0.00018931670612919418, "loss": 5.1212, "step": 484200 }, { "epoch": 2.9199479022201853, "grad_norm": 2.191455841064453, "learning_rate": 0.0001892970682501599, "loss": 4.8397, "step": 484250 }, { "epoch": 2.9202493940015195, "grad_norm": 2.257707357406616, "learning_rate": 0.00018927742964791922, "loss": 4.8656, "step": 484300 }, { "epoch": 2.9205508857828537, "grad_norm": 2.427840232849121, "learning_rate": 0.0001892577903228334, "loss": 4.7953, "step": 484350 }, { "epoch": 2.9208523775641875, "grad_norm": 1.811156153678894, "learning_rate": 0.00018923815027526401, "loss": 4.8302, "step": 484400 }, { "epoch": 2.9211538693455217, "grad_norm": 2.308962106704712, "learning_rate": 0.00018921850950557247, "loss": 5.2421, "step": 484450 }, { "epoch": 2.9214553611268554, "grad_norm": 2.355224609375, "learning_rate": 0.00018919886801412023, "loss": 4.683, "step": 484500 }, { "epoch": 2.9217568529081897, "grad_norm": 1.5458158254623413, "learning_rate": 0.00018917922580126878, "loss": 5.0352, "step": 484550 }, { "epoch": 2.922058344689524, "grad_norm": 1.5204042196273804, "learning_rate": 0.0001891595828673796, "loss": 4.8504, "step": 484600 }, { "epoch": 2.922359836470858, "grad_norm": 1.7129606008529663, "learning_rate": 0.0001891399392128141, "loss": 5.0775, "step": 484650 }, { "epoch": 2.922661328252192, "grad_norm": 1.4836210012435913, "learning_rate": 0.00018912029483793397, "loss": 4.6522, "step": 484700 }, { "epoch": 2.922962820033526, "grad_norm": 3.6354336738586426, "learning_rate": 0.0001891006497431006, "loss": 4.8878, "step": 484750 }, { "epoch": 2.92326431181486, "grad_norm": 1.938124656677246, "learning_rate": 0.00018908100392867554, "loss": 4.8562, "step": 484800 }, { "epoch": 2.923565803596194, "grad_norm": 1.341799020767212, "learning_rate": 0.00018906135739502042, "loss": 4.7068, "step": 484850 }, { "epoch": 2.923867295377528, "grad_norm": 1.0666002035140991, "learning_rate": 0.00018904171014249676, "loss": 4.6761, "step": 484900 }, { "epoch": 2.924168787158862, "grad_norm": 0.647466242313385, "learning_rate": 0.00018902206217146614, "loss": 5.1686, "step": 484950 }, { "epoch": 2.924470278940196, "grad_norm": 1.929855227470398, "learning_rate": 0.0001890024134822901, "loss": 5.0344, "step": 485000 }, { "epoch": 2.92477177072153, "grad_norm": 2.175678253173828, "learning_rate": 0.00018898276407533038, "loss": 4.7515, "step": 485050 }, { "epoch": 2.925073262502864, "grad_norm": 2.1115481853485107, "learning_rate": 0.00018896311395094845, "loss": 4.8262, "step": 485100 }, { "epoch": 2.9253747542841984, "grad_norm": 1.7033413648605347, "learning_rate": 0.00018894346310950605, "loss": 5.3911, "step": 485150 }, { "epoch": 2.925676246065532, "grad_norm": 0.9272875189781189, "learning_rate": 0.00018892381155136476, "loss": 5.1936, "step": 485200 }, { "epoch": 2.9259777378468663, "grad_norm": 2.392357110977173, "learning_rate": 0.00018890415927688627, "loss": 4.7252, "step": 485250 }, { "epoch": 2.9262792296282, "grad_norm": 2.751167058944702, "learning_rate": 0.0001888845062864323, "loss": 4.8774, "step": 485300 }, { "epoch": 2.9265807214095343, "grad_norm": 1.8779020309448242, "learning_rate": 0.00018886485258036442, "loss": 4.8262, "step": 485350 }, { "epoch": 2.9268822131908685, "grad_norm": 1.85697603225708, "learning_rate": 0.00018884519815904442, "loss": 4.6051, "step": 485400 }, { "epoch": 2.9271837049722027, "grad_norm": 2.4284555912017822, "learning_rate": 0.00018882554302283397, "loss": 4.8862, "step": 485450 }, { "epoch": 2.9274851967535365, "grad_norm": 1.3987573385238647, "learning_rate": 0.00018880588717209483, "loss": 5.0162, "step": 485500 }, { "epoch": 2.9277866885348707, "grad_norm": 1.6028248071670532, "learning_rate": 0.0001887862306071887, "loss": 4.826, "step": 485550 }, { "epoch": 2.9280881803162044, "grad_norm": 1.6885029077529907, "learning_rate": 0.00018876657332847735, "loss": 4.995, "step": 485600 }, { "epoch": 2.9283896720975386, "grad_norm": 1.5717148780822754, "learning_rate": 0.0001887469153363225, "loss": 4.7722, "step": 485650 }, { "epoch": 2.928691163878873, "grad_norm": 1.7408987283706665, "learning_rate": 0.00018872725663108604, "loss": 5.0808, "step": 485700 }, { "epoch": 2.9289926556602066, "grad_norm": 1.0260766744613647, "learning_rate": 0.00018870759721312968, "loss": 5.0344, "step": 485750 }, { "epoch": 2.929294147441541, "grad_norm": 2.260704278945923, "learning_rate": 0.00018868793708281517, "loss": 4.6657, "step": 485800 }, { "epoch": 2.9295956392228746, "grad_norm": 0.5136196613311768, "learning_rate": 0.00018866827624050442, "loss": 4.9678, "step": 485850 }, { "epoch": 2.929897131004209, "grad_norm": 2.649401903152466, "learning_rate": 0.00018864861468655928, "loss": 4.716, "step": 485900 }, { "epoch": 2.930198622785543, "grad_norm": 2.0867464542388916, "learning_rate": 0.00018862895242134143, "loss": 5.1427, "step": 485950 }, { "epoch": 2.9305001145668768, "grad_norm": 1.4744309186935425, "learning_rate": 0.00018860928944521292, "loss": 5.0735, "step": 486000 }, { "epoch": 2.9305001145668768, "eval_loss": 5.384421348571777, "eval_runtime": 39.2251, "eval_samples_per_second": 13.053, "eval_steps_per_second": 6.526, "eval_tts_loss": 7.372810278247256, "step": 486000 }, { "epoch": 2.930801606348211, "grad_norm": 2.5548908710479736, "learning_rate": 0.00018858962575853556, "loss": 5.0266, "step": 486050 }, { "epoch": 2.9311030981295447, "grad_norm": 2.3923256397247314, "learning_rate": 0.00018856996136167113, "loss": 4.4881, "step": 486100 }, { "epoch": 2.931404589910879, "grad_norm": 1.952491044998169, "learning_rate": 0.00018855029625498165, "loss": 5.0809, "step": 486150 }, { "epoch": 2.931706081692213, "grad_norm": 1.8679330348968506, "learning_rate": 0.00018853063043882898, "loss": 4.8956, "step": 486200 }, { "epoch": 2.9320075734735473, "grad_norm": 2.672300338745117, "learning_rate": 0.00018851096391357499, "loss": 5.1622, "step": 486250 }, { "epoch": 2.932309065254881, "grad_norm": 1.9964553117752075, "learning_rate": 0.00018849129667958178, "loss": 5.0258, "step": 486300 }, { "epoch": 2.9326105570362153, "grad_norm": 2.221071720123291, "learning_rate": 0.00018847162873721112, "loss": 4.9053, "step": 486350 }, { "epoch": 2.932912048817549, "grad_norm": 1.5855296850204468, "learning_rate": 0.000188451960086825, "loss": 4.8265, "step": 486400 }, { "epoch": 2.9332135405988833, "grad_norm": 1.6828289031982422, "learning_rate": 0.00018843229072878543, "loss": 4.6943, "step": 486450 }, { "epoch": 2.9335150323802175, "grad_norm": 1.8949888944625854, "learning_rate": 0.00018841262066345447, "loss": 4.9118, "step": 486500 }, { "epoch": 2.9338165241615513, "grad_norm": 1.6552726030349731, "learning_rate": 0.00018839294989119398, "loss": 4.8012, "step": 486550 }, { "epoch": 2.9341180159428855, "grad_norm": 1.9992390871047974, "learning_rate": 0.00018837327841236606, "loss": 4.7024, "step": 486600 }, { "epoch": 2.9344195077242192, "grad_norm": 1.6656324863433838, "learning_rate": 0.0001883536062273327, "loss": 5.3665, "step": 486650 }, { "epoch": 2.9347209995055534, "grad_norm": 0.5996415615081787, "learning_rate": 0.00018833393333645597, "loss": 4.835, "step": 486700 }, { "epoch": 2.9350224912868876, "grad_norm": 2.730095386505127, "learning_rate": 0.0001883142597400979, "loss": 4.4843, "step": 486750 }, { "epoch": 2.9353239830682214, "grad_norm": 2.0547773838043213, "learning_rate": 0.0001882945854386205, "loss": 5.033, "step": 486800 }, { "epoch": 2.9356254748495556, "grad_norm": 2.3592844009399414, "learning_rate": 0.00018827491043238597, "loss": 4.7816, "step": 486850 }, { "epoch": 2.9359269666308894, "grad_norm": 2.205765724182129, "learning_rate": 0.00018825523472175637, "loss": 4.7022, "step": 486900 }, { "epoch": 2.9362284584122236, "grad_norm": 1.696293830871582, "learning_rate": 0.00018823555830709365, "loss": 4.8788, "step": 486950 }, { "epoch": 2.936529950193558, "grad_norm": 2.1423497200012207, "learning_rate": 0.00018821588118876013, "loss": 4.9311, "step": 487000 }, { "epoch": 2.936831441974892, "grad_norm": 1.9746463298797607, "learning_rate": 0.00018819620336711784, "loss": 4.8033, "step": 487050 }, { "epoch": 2.9371329337562257, "grad_norm": 2.234518051147461, "learning_rate": 0.00018817652484252892, "loss": 5.0124, "step": 487100 }, { "epoch": 2.93743442553756, "grad_norm": 1.6610745191574097, "learning_rate": 0.00018815684561535555, "loss": 4.9606, "step": 487150 }, { "epoch": 2.9377359173188937, "grad_norm": 1.6995896100997925, "learning_rate": 0.0001881371656859599, "loss": 5.0083, "step": 487200 }, { "epoch": 2.938037409100228, "grad_norm": 1.2924972772598267, "learning_rate": 0.0001881174850547041, "loss": 5.1855, "step": 487250 }, { "epoch": 2.938338900881562, "grad_norm": 1.4959458112716675, "learning_rate": 0.00018809780372195046, "loss": 4.8094, "step": 487300 }, { "epoch": 2.938640392662896, "grad_norm": 1.1287888288497925, "learning_rate": 0.00018807812168806107, "loss": 4.8705, "step": 487350 }, { "epoch": 2.93894188444423, "grad_norm": 0.8478665947914124, "learning_rate": 0.00018805843895339813, "loss": 4.8288, "step": 487400 }, { "epoch": 2.939243376225564, "grad_norm": 2.300889015197754, "learning_rate": 0.000188038755518324, "loss": 4.9311, "step": 487450 }, { "epoch": 2.939544868006898, "grad_norm": 2.00642466545105, "learning_rate": 0.0001880190713832009, "loss": 4.9719, "step": 487500 }, { "epoch": 2.9398463597882323, "grad_norm": 2.1443252563476562, "learning_rate": 0.00018799938654839095, "loss": 4.7348, "step": 487550 }, { "epoch": 2.9401478515695665, "grad_norm": 1.8157026767730713, "learning_rate": 0.0001879797010142566, "loss": 4.834, "step": 487600 }, { "epoch": 2.9404493433509002, "grad_norm": 1.3425401449203491, "learning_rate": 0.00018796001478116003, "loss": 4.9257, "step": 487650 }, { "epoch": 2.9407508351322345, "grad_norm": 1.7326247692108154, "learning_rate": 0.0001879403278494635, "loss": 5.4146, "step": 487700 }, { "epoch": 2.941052326913568, "grad_norm": 2.3009755611419678, "learning_rate": 0.00018792064021952947, "loss": 4.7067, "step": 487750 }, { "epoch": 2.9413538186949024, "grad_norm": 2.992279291152954, "learning_rate": 0.0001879009518917201, "loss": 4.8652, "step": 487800 }, { "epoch": 2.9416553104762366, "grad_norm": 2.0740082263946533, "learning_rate": 0.00018788126286639782, "loss": 4.4122, "step": 487850 }, { "epoch": 2.9419568022575704, "grad_norm": 1.8049949407577515, "learning_rate": 0.000187861573143925, "loss": 4.7109, "step": 487900 }, { "epoch": 2.9422582940389046, "grad_norm": 2.5029969215393066, "learning_rate": 0.0001878418827246639, "loss": 4.8219, "step": 487950 }, { "epoch": 2.9425597858202384, "grad_norm": 2.2186014652252197, "learning_rate": 0.00018782219160897698, "loss": 4.8007, "step": 488000 }, { "epoch": 2.9428612776015726, "grad_norm": 1.7065647840499878, "learning_rate": 0.00018780249979722657, "loss": 5.1638, "step": 488050 }, { "epoch": 2.9431627693829068, "grad_norm": 1.5295652151107788, "learning_rate": 0.00018778280728977515, "loss": 5.0217, "step": 488100 }, { "epoch": 2.9434642611642405, "grad_norm": 2.340228319168091, "learning_rate": 0.00018776311408698503, "loss": 5.0132, "step": 488150 }, { "epoch": 2.9437657529455747, "grad_norm": 1.607174277305603, "learning_rate": 0.00018774342018921874, "loss": 5.2783, "step": 488200 }, { "epoch": 2.9440672447269085, "grad_norm": 2.3659369945526123, "learning_rate": 0.0001877237255968386, "loss": 4.7045, "step": 488250 }, { "epoch": 2.9443687365082427, "grad_norm": 1.5629881620407104, "learning_rate": 0.00018770403031020716, "loss": 4.7992, "step": 488300 }, { "epoch": 2.944670228289577, "grad_norm": 2.0309884548187256, "learning_rate": 0.00018768433432968685, "loss": 5.2114, "step": 488350 }, { "epoch": 2.944971720070911, "grad_norm": 1.5308544635772705, "learning_rate": 0.00018766463765564016, "loss": 4.8668, "step": 488400 }, { "epoch": 2.945273211852245, "grad_norm": 1.9208180904388428, "learning_rate": 0.00018764494028842952, "loss": 4.5243, "step": 488450 }, { "epoch": 2.945574703633579, "grad_norm": 1.7741751670837402, "learning_rate": 0.00018762524222841752, "loss": 4.8832, "step": 488500 }, { "epoch": 2.945876195414913, "grad_norm": 1.8668218851089478, "learning_rate": 0.00018760554347596657, "loss": 4.979, "step": 488550 }, { "epoch": 2.946177687196247, "grad_norm": 1.1370770931243896, "learning_rate": 0.00018758584403143932, "loss": 5.416, "step": 488600 }, { "epoch": 2.9464791789775813, "grad_norm": 1.8240931034088135, "learning_rate": 0.00018756614389519823, "loss": 4.8553, "step": 488650 }, { "epoch": 2.946780670758915, "grad_norm": 2.857642650604248, "learning_rate": 0.00018754644306760585, "loss": 4.6392, "step": 488700 }, { "epoch": 2.9470821625402492, "grad_norm": 1.7842975854873657, "learning_rate": 0.0001875267415490248, "loss": 4.7183, "step": 488750 }, { "epoch": 2.947383654321583, "grad_norm": 0.5367608070373535, "learning_rate": 0.00018750703933981762, "loss": 5.0066, "step": 488800 }, { "epoch": 2.947685146102917, "grad_norm": 1.2262375354766846, "learning_rate": 0.00018748733644034693, "loss": 4.6943, "step": 488850 }, { "epoch": 2.9479866378842514, "grad_norm": 2.1758296489715576, "learning_rate": 0.00018746763285097526, "loss": 5.1997, "step": 488900 }, { "epoch": 2.948288129665585, "grad_norm": 1.6052755117416382, "learning_rate": 0.00018744792857206528, "loss": 5.1587, "step": 488950 }, { "epoch": 2.9485896214469194, "grad_norm": 2.334592819213867, "learning_rate": 0.0001874282236039796, "loss": 5.2664, "step": 489000 }, { "epoch": 2.9485896214469194, "eval_loss": 5.380739212036133, "eval_runtime": 38.8635, "eval_samples_per_second": 13.174, "eval_steps_per_second": 6.587, "eval_tts_loss": 7.425174013293107, "step": 489000 }, { "epoch": 2.948891113228253, "grad_norm": 2.1524577140808105, "learning_rate": 0.00018740851794708094, "loss": 5.0183, "step": 489050 }, { "epoch": 2.9491926050095874, "grad_norm": 2.340575695037842, "learning_rate": 0.00018738881160173187, "loss": 5.0572, "step": 489100 }, { "epoch": 2.9494940967909216, "grad_norm": 0.6995661854743958, "learning_rate": 0.00018736910456829505, "loss": 4.9997, "step": 489150 }, { "epoch": 2.9497955885722558, "grad_norm": 1.9784549474716187, "learning_rate": 0.0001873493968471332, "loss": 5.0874, "step": 489200 }, { "epoch": 2.9500970803535895, "grad_norm": 0.7341315746307373, "learning_rate": 0.000187329688438609, "loss": 4.7104, "step": 489250 }, { "epoch": 2.9503985721349237, "grad_norm": 1.2109259366989136, "learning_rate": 0.0001873099793430852, "loss": 4.6768, "step": 489300 }, { "epoch": 2.9507000639162575, "grad_norm": 1.5176061391830444, "learning_rate": 0.00018729026956092442, "loss": 4.7174, "step": 489350 }, { "epoch": 2.9510015556975917, "grad_norm": 2.4323360919952393, "learning_rate": 0.00018727055909248947, "loss": 5.225, "step": 489400 }, { "epoch": 2.951303047478926, "grad_norm": 1.5510241985321045, "learning_rate": 0.00018725084793814304, "loss": 4.7683, "step": 489450 }, { "epoch": 2.9516045392602597, "grad_norm": 1.978031873703003, "learning_rate": 0.00018723113609824798, "loss": 4.8524, "step": 489500 }, { "epoch": 2.951906031041594, "grad_norm": 2.1510424613952637, "learning_rate": 0.00018721142357316697, "loss": 5.5653, "step": 489550 }, { "epoch": 2.9522075228229276, "grad_norm": 2.751823902130127, "learning_rate": 0.00018719171036326272, "loss": 5.0716, "step": 489600 }, { "epoch": 2.952509014604262, "grad_norm": 1.3511673212051392, "learning_rate": 0.00018717199646889823, "loss": 5.0864, "step": 489650 }, { "epoch": 2.952810506385596, "grad_norm": 1.678035020828247, "learning_rate": 0.00018715228189043615, "loss": 5.1029, "step": 489700 }, { "epoch": 2.95311199816693, "grad_norm": 1.5100332498550415, "learning_rate": 0.00018713256662823932, "loss": 5.5994, "step": 489750 }, { "epoch": 2.953413489948264, "grad_norm": 2.118985414505005, "learning_rate": 0.00018711285068267062, "loss": 4.7308, "step": 489800 }, { "epoch": 2.953714981729598, "grad_norm": 3.1586005687713623, "learning_rate": 0.00018709313405409288, "loss": 5.1915, "step": 489850 }, { "epoch": 2.954016473510932, "grad_norm": 1.941860556602478, "learning_rate": 0.0001870734167428689, "loss": 4.4632, "step": 489900 }, { "epoch": 2.954317965292266, "grad_norm": 1.7266087532043457, "learning_rate": 0.00018705369874936163, "loss": 4.7749, "step": 489950 }, { "epoch": 2.9546194570736004, "grad_norm": 2.214914560317993, "learning_rate": 0.0001870339800739339, "loss": 4.885, "step": 490000 }, { "epoch": 2.954920948854934, "grad_norm": 2.0960185527801514, "learning_rate": 0.00018701426071694857, "loss": 4.3688, "step": 490050 }, { "epoch": 2.9552224406362684, "grad_norm": 2.2868175506591797, "learning_rate": 0.0001869945406787687, "loss": 5.2621, "step": 490100 }, { "epoch": 2.955523932417602, "grad_norm": 2.320829391479492, "learning_rate": 0.00018697481995975704, "loss": 4.94, "step": 490150 }, { "epoch": 2.9558254241989363, "grad_norm": 1.4300296306610107, "learning_rate": 0.00018695509856027655, "loss": 4.8545, "step": 490200 }, { "epoch": 2.9561269159802706, "grad_norm": 1.7082023620605469, "learning_rate": 0.00018693537648069024, "loss": 4.6901, "step": 490250 }, { "epoch": 2.9564284077616043, "grad_norm": 0.9940247535705566, "learning_rate": 0.00018691565372136101, "loss": 4.4145, "step": 490300 }, { "epoch": 2.9567298995429385, "grad_norm": 1.2869545221328735, "learning_rate": 0.0001868959302826519, "loss": 5.1346, "step": 490350 }, { "epoch": 2.9570313913242723, "grad_norm": 2.4928383827209473, "learning_rate": 0.00018687620616492583, "loss": 4.7141, "step": 490400 }, { "epoch": 2.9573328831056065, "grad_norm": 3.442091941833496, "learning_rate": 0.00018685648136854583, "loss": 4.6943, "step": 490450 }, { "epoch": 2.9576343748869407, "grad_norm": 1.699309229850769, "learning_rate": 0.00018683675589387485, "loss": 5.0011, "step": 490500 }, { "epoch": 2.9579358666682745, "grad_norm": 0.8238717913627625, "learning_rate": 0.00018681702974127597, "loss": 5.1695, "step": 490550 }, { "epoch": 2.9582373584496087, "grad_norm": 1.7315499782562256, "learning_rate": 0.00018679730291111218, "loss": 4.8956, "step": 490600 }, { "epoch": 2.9585388502309424, "grad_norm": 2.8208515644073486, "learning_rate": 0.00018677757540374657, "loss": 5.4662, "step": 490650 }, { "epoch": 2.9588403420122766, "grad_norm": 2.932447671890259, "learning_rate": 0.00018675784721954218, "loss": 5.0367, "step": 490700 }, { "epoch": 2.959141833793611, "grad_norm": 2.2972664833068848, "learning_rate": 0.000186738118358862, "loss": 5.1061, "step": 490750 }, { "epoch": 2.959443325574945, "grad_norm": 1.9315645694732666, "learning_rate": 0.00018671838882206924, "loss": 4.9436, "step": 490800 }, { "epoch": 2.959744817356279, "grad_norm": 0.6536876559257507, "learning_rate": 0.0001866986586095269, "loss": 5.1252, "step": 490850 }, { "epoch": 2.960046309137613, "grad_norm": 3.2365033626556396, "learning_rate": 0.00018667892772159813, "loss": 4.5587, "step": 490900 }, { "epoch": 2.960347800918947, "grad_norm": 1.246462345123291, "learning_rate": 0.00018665919615864606, "loss": 5.3118, "step": 490950 }, { "epoch": 2.960649292700281, "grad_norm": 1.5913242101669312, "learning_rate": 0.0001866394639210338, "loss": 5.1896, "step": 491000 }, { "epoch": 2.960950784481615, "grad_norm": 1.690926194190979, "learning_rate": 0.00018661973100912447, "loss": 5.0031, "step": 491050 }, { "epoch": 2.961252276262949, "grad_norm": 1.8087538480758667, "learning_rate": 0.00018659999742328128, "loss": 4.968, "step": 491100 }, { "epoch": 2.961553768044283, "grad_norm": 2.469815969467163, "learning_rate": 0.00018658026316386734, "loss": 5.2623, "step": 491150 }, { "epoch": 2.961855259825617, "grad_norm": 2.2247567176818848, "learning_rate": 0.00018656052823124583, "loss": 5.6141, "step": 491200 }, { "epoch": 2.962156751606951, "grad_norm": 1.4975228309631348, "learning_rate": 0.00018654079262578005, "loss": 4.888, "step": 491250 }, { "epoch": 2.9624582433882853, "grad_norm": 1.5557520389556885, "learning_rate": 0.0001865210563478331, "loss": 4.8782, "step": 491300 }, { "epoch": 2.9627597351696195, "grad_norm": 1.4202994108200073, "learning_rate": 0.00018650131939776816, "loss": 4.9158, "step": 491350 }, { "epoch": 2.9630612269509533, "grad_norm": 1.8808863162994385, "learning_rate": 0.00018648158177594858, "loss": 4.9442, "step": 491400 }, { "epoch": 2.9633627187322875, "grad_norm": 1.6651962995529175, "learning_rate": 0.00018646184348273756, "loss": 4.9522, "step": 491450 }, { "epoch": 2.9636642105136213, "grad_norm": 0.901952862739563, "learning_rate": 0.00018644210451849834, "loss": 4.473, "step": 491500 }, { "epoch": 2.9639657022949555, "grad_norm": 1.936919927597046, "learning_rate": 0.00018642236488359418, "loss": 4.9451, "step": 491550 }, { "epoch": 2.9642671940762897, "grad_norm": 1.7133064270019531, "learning_rate": 0.0001864026245783884, "loss": 5.1342, "step": 491600 }, { "epoch": 2.9645686858576235, "grad_norm": 1.8377978801727295, "learning_rate": 0.00018638288360324422, "loss": 5.119, "step": 491650 }, { "epoch": 2.9648701776389577, "grad_norm": 1.6773120164871216, "learning_rate": 0.00018636314195852503, "loss": 4.9036, "step": 491700 }, { "epoch": 2.9651716694202914, "grad_norm": 0.9803944230079651, "learning_rate": 0.00018634339964459408, "loss": 4.5319, "step": 491750 }, { "epoch": 2.9654731612016256, "grad_norm": 1.8291908502578735, "learning_rate": 0.00018632365666181473, "loss": 5.1401, "step": 491800 }, { "epoch": 2.96577465298296, "grad_norm": 2.517038583755493, "learning_rate": 0.0001863039130105503, "loss": 5.0549, "step": 491850 }, { "epoch": 2.9660761447642936, "grad_norm": 2.140916347503662, "learning_rate": 0.0001862841686911642, "loss": 4.359, "step": 491900 }, { "epoch": 2.966377636545628, "grad_norm": 1.910070776939392, "learning_rate": 0.0001862644237040197, "loss": 5.106, "step": 491950 }, { "epoch": 2.9666791283269616, "grad_norm": 1.1323330402374268, "learning_rate": 0.00018624467804948027, "loss": 4.716, "step": 492000 }, { "epoch": 2.9666791283269616, "eval_loss": 5.382606506347656, "eval_runtime": 38.8088, "eval_samples_per_second": 13.193, "eval_steps_per_second": 6.596, "eval_tts_loss": 7.462707099899911, "step": 492000 }, { "epoch": 2.9669806201082958, "grad_norm": 2.24629282951355, "learning_rate": 0.0001862249317279093, "loss": 5.4572, "step": 492050 }, { "epoch": 2.96728211188963, "grad_norm": 1.2180079221725464, "learning_rate": 0.0001862051847396701, "loss": 4.7379, "step": 492100 }, { "epoch": 2.967583603670964, "grad_norm": 2.328089475631714, "learning_rate": 0.00018618543708512616, "loss": 4.9558, "step": 492150 }, { "epoch": 2.967885095452298, "grad_norm": 1.3277615308761597, "learning_rate": 0.00018616568876464085, "loss": 4.8213, "step": 492200 }, { "epoch": 2.968186587233632, "grad_norm": 1.8075981140136719, "learning_rate": 0.00018614593977857769, "loss": 4.9608, "step": 492250 }, { "epoch": 2.968488079014966, "grad_norm": 2.2147376537323, "learning_rate": 0.00018612619012730013, "loss": 4.4825, "step": 492300 }, { "epoch": 2.9687895707963, "grad_norm": 1.9716358184814453, "learning_rate": 0.00018610643981117151, "loss": 4.906, "step": 492350 }, { "epoch": 2.9690910625776343, "grad_norm": 1.1882251501083374, "learning_rate": 0.00018608668883055544, "loss": 4.8749, "step": 492400 }, { "epoch": 2.969392554358968, "grad_norm": 1.6744645833969116, "learning_rate": 0.00018606693718581537, "loss": 4.7964, "step": 492450 }, { "epoch": 2.9696940461403023, "grad_norm": 2.463465452194214, "learning_rate": 0.0001860471848773148, "loss": 4.9367, "step": 492500 }, { "epoch": 2.969995537921636, "grad_norm": 1.4924885034561157, "learning_rate": 0.00018602743190541718, "loss": 5.3869, "step": 492550 }, { "epoch": 2.9702970297029703, "grad_norm": 2.61120867729187, "learning_rate": 0.00018600767827048613, "loss": 5.1408, "step": 492600 }, { "epoch": 2.9705985214843045, "grad_norm": 2.2865233421325684, "learning_rate": 0.00018598792397288516, "loss": 4.8152, "step": 492650 }, { "epoch": 2.9709000132656382, "grad_norm": 2.645756244659424, "learning_rate": 0.00018596816901297777, "loss": 4.9372, "step": 492700 }, { "epoch": 2.9712015050469724, "grad_norm": 0.61232590675354, "learning_rate": 0.0001859484133911276, "loss": 4.6852, "step": 492750 }, { "epoch": 2.971502996828306, "grad_norm": 2.325610399246216, "learning_rate": 0.00018592865710769814, "loss": 5.0595, "step": 492800 }, { "epoch": 2.9718044886096404, "grad_norm": 2.224430799484253, "learning_rate": 0.00018590890016305305, "loss": 4.5957, "step": 492850 }, { "epoch": 2.9721059803909746, "grad_norm": 2.116002321243286, "learning_rate": 0.0001858891425575559, "loss": 5.0738, "step": 492900 }, { "epoch": 2.972407472172309, "grad_norm": 1.4503391981124878, "learning_rate": 0.00018586938429157026, "loss": 4.7114, "step": 492950 }, { "epoch": 2.9727089639536426, "grad_norm": 1.6504504680633545, "learning_rate": 0.0001858496253654598, "loss": 4.7944, "step": 493000 }, { "epoch": 2.973010455734977, "grad_norm": 2.047123670578003, "learning_rate": 0.00018582986577958817, "loss": 4.9709, "step": 493050 }, { "epoch": 2.9733119475163106, "grad_norm": 1.821801781654358, "learning_rate": 0.00018581010553431898, "loss": 5.2718, "step": 493100 }, { "epoch": 2.9736134392976448, "grad_norm": 1.0614919662475586, "learning_rate": 0.0001857903446300159, "loss": 4.9971, "step": 493150 }, { "epoch": 2.973914931078979, "grad_norm": 0.8596398234367371, "learning_rate": 0.00018577058306704262, "loss": 5.0028, "step": 493200 }, { "epoch": 2.9742164228603127, "grad_norm": 2.09301495552063, "learning_rate": 0.0001857508208457628, "loss": 4.834, "step": 493250 }, { "epoch": 2.974517914641647, "grad_norm": 0.4898781478404999, "learning_rate": 0.0001857310579665401, "loss": 4.821, "step": 493300 }, { "epoch": 2.9748194064229807, "grad_norm": 3.0883588790893555, "learning_rate": 0.00018571129442973832, "loss": 4.52, "step": 493350 }, { "epoch": 2.975120898204315, "grad_norm": 1.451945185661316, "learning_rate": 0.0001856915302357211, "loss": 5.1138, "step": 493400 }, { "epoch": 2.975422389985649, "grad_norm": 1.4707807302474976, "learning_rate": 0.0001856717653848522, "loss": 4.9332, "step": 493450 }, { "epoch": 2.975723881766983, "grad_norm": 2.1867361068725586, "learning_rate": 0.0001856519998774954, "loss": 5.2446, "step": 493500 }, { "epoch": 2.976025373548317, "grad_norm": 2.056126356124878, "learning_rate": 0.00018563223371401437, "loss": 5.0163, "step": 493550 }, { "epoch": 2.976326865329651, "grad_norm": 1.4984785318374634, "learning_rate": 0.00018561246689477294, "loss": 5.102, "step": 493600 }, { "epoch": 2.976628357110985, "grad_norm": 1.5041289329528809, "learning_rate": 0.00018559269942013492, "loss": 4.7726, "step": 493650 }, { "epoch": 2.9769298488923193, "grad_norm": 4.21092414855957, "learning_rate": 0.00018557293129046402, "loss": 5.0708, "step": 493700 }, { "epoch": 2.9772313406736535, "grad_norm": 2.1806344985961914, "learning_rate": 0.00018555316250612412, "loss": 5.085, "step": 493750 }, { "epoch": 2.9775328324549872, "grad_norm": 1.6425265073776245, "learning_rate": 0.00018553339306747894, "loss": 5.281, "step": 493800 }, { "epoch": 2.9778343242363214, "grad_norm": 2.2072155475616455, "learning_rate": 0.00018551362297489243, "loss": 4.9741, "step": 493850 }, { "epoch": 2.978135816017655, "grad_norm": 1.9444726705551147, "learning_rate": 0.00018549385222872838, "loss": 5.0227, "step": 493900 }, { "epoch": 2.9784373077989894, "grad_norm": 1.428072452545166, "learning_rate": 0.00018547408082935054, "loss": 4.81, "step": 493950 }, { "epoch": 2.9787387995803236, "grad_norm": 0.30138590931892395, "learning_rate": 0.00018545430877712295, "loss": 4.8467, "step": 494000 }, { "epoch": 2.9790402913616574, "grad_norm": 1.77151620388031, "learning_rate": 0.00018543453607240936, "loss": 5.0363, "step": 494050 }, { "epoch": 2.9793417831429916, "grad_norm": 2.692133903503418, "learning_rate": 0.00018541476271557373, "loss": 4.7396, "step": 494100 }, { "epoch": 2.9796432749243253, "grad_norm": 0.6094012260437012, "learning_rate": 0.00018539498870697994, "loss": 5.2036, "step": 494150 }, { "epoch": 2.9799447667056596, "grad_norm": 1.5933867692947388, "learning_rate": 0.00018537521404699188, "loss": 4.706, "step": 494200 }, { "epoch": 2.9802462584869938, "grad_norm": 2.1332924365997314, "learning_rate": 0.00018535543873597347, "loss": 4.8232, "step": 494250 }, { "epoch": 2.9805477502683275, "grad_norm": 1.3278063535690308, "learning_rate": 0.0001853356627742887, "loss": 4.9223, "step": 494300 }, { "epoch": 2.9808492420496617, "grad_norm": 1.7767270803451538, "learning_rate": 0.00018531588616230147, "loss": 4.682, "step": 494350 }, { "epoch": 2.9811507338309955, "grad_norm": 2.2455649375915527, "learning_rate": 0.00018529610890037572, "loss": 4.7148, "step": 494400 }, { "epoch": 2.9814522256123297, "grad_norm": 1.5640569925308228, "learning_rate": 0.0001852763309888755, "loss": 4.9834, "step": 494450 }, { "epoch": 2.981753717393664, "grad_norm": 1.6799533367156982, "learning_rate": 0.00018525655242816475, "loss": 5.2028, "step": 494500 }, { "epoch": 2.982055209174998, "grad_norm": 1.3156894445419312, "learning_rate": 0.00018523677321860741, "loss": 4.4282, "step": 494550 }, { "epoch": 2.982356700956332, "grad_norm": 2.4435157775878906, "learning_rate": 0.00018521699336056761, "loss": 4.8052, "step": 494600 }, { "epoch": 2.982658192737666, "grad_norm": 3.38358736038208, "learning_rate": 0.0001851972128544093, "loss": 5.3224, "step": 494650 }, { "epoch": 2.982959684519, "grad_norm": 0.7056095600128174, "learning_rate": 0.00018517743170049647, "loss": 5.3631, "step": 494700 }, { "epoch": 2.983261176300334, "grad_norm": 2.3139398097991943, "learning_rate": 0.00018515764989919325, "loss": 5.076, "step": 494750 }, { "epoch": 2.9835626680816683, "grad_norm": 2.5954525470733643, "learning_rate": 0.00018513786745086362, "loss": 5.2788, "step": 494800 }, { "epoch": 2.983864159863002, "grad_norm": 1.2993394136428833, "learning_rate": 0.00018511808435587172, "loss": 4.7186, "step": 494850 }, { "epoch": 2.9841656516443362, "grad_norm": 2.87802791595459, "learning_rate": 0.0001850983006145816, "loss": 4.7867, "step": 494900 }, { "epoch": 2.98446714342567, "grad_norm": 1.6869983673095703, "learning_rate": 0.00018507851622735733, "loss": 5.2277, "step": 494950 }, { "epoch": 2.984768635207004, "grad_norm": 2.5125980377197266, "learning_rate": 0.00018505873119456298, "loss": 4.9325, "step": 495000 }, { "epoch": 2.984768635207004, "eval_loss": 5.376990795135498, "eval_runtime": 38.8766, "eval_samples_per_second": 13.17, "eval_steps_per_second": 6.585, "eval_tts_loss": 7.459579689951467, "step": 495000 }, { "epoch": 2.9850701269883384, "grad_norm": 1.8387424945831299, "learning_rate": 0.00018503894551656275, "loss": 4.8912, "step": 495050 }, { "epoch": 2.9853716187696726, "grad_norm": 2.3910181522369385, "learning_rate": 0.00018501915919372079, "loss": 5.0408, "step": 495100 }, { "epoch": 2.9856731105510064, "grad_norm": 1.0605469942092896, "learning_rate": 0.00018499937222640108, "loss": 5.0022, "step": 495150 }, { "epoch": 2.9859746023323406, "grad_norm": 1.2042174339294434, "learning_rate": 0.00018497958461496795, "loss": 4.3634, "step": 495200 }, { "epoch": 2.9862760941136743, "grad_norm": 1.4237060546875, "learning_rate": 0.00018495979635978544, "loss": 4.7827, "step": 495250 }, { "epoch": 2.9865775858950085, "grad_norm": 1.8054640293121338, "learning_rate": 0.0001849400074612178, "loss": 4.8814, "step": 495300 }, { "epoch": 2.9868790776763428, "grad_norm": 2.522623300552368, "learning_rate": 0.00018492021791962919, "loss": 5.3551, "step": 495350 }, { "epoch": 2.9871805694576765, "grad_norm": 3.6063501834869385, "learning_rate": 0.00018490042773538373, "loss": 4.6158, "step": 495400 }, { "epoch": 2.9874820612390107, "grad_norm": 2.1944828033447266, "learning_rate": 0.00018488063690884576, "loss": 4.9174, "step": 495450 }, { "epoch": 2.9877835530203445, "grad_norm": 2.5135385990142822, "learning_rate": 0.00018486084544037945, "loss": 4.8272, "step": 495500 }, { "epoch": 2.9880850448016787, "grad_norm": 0.7620628476142883, "learning_rate": 0.000184841053330349, "loss": 4.906, "step": 495550 }, { "epoch": 2.988386536583013, "grad_norm": 1.6455497741699219, "learning_rate": 0.0001848212605791187, "loss": 4.6615, "step": 495600 }, { "epoch": 2.9886880283643467, "grad_norm": 1.6383532285690308, "learning_rate": 0.0001848014671870528, "loss": 5.6451, "step": 495650 }, { "epoch": 2.988989520145681, "grad_norm": 3.1217148303985596, "learning_rate": 0.00018478167315451555, "loss": 4.9321, "step": 495700 }, { "epoch": 2.9892910119270146, "grad_norm": 1.963880181312561, "learning_rate": 0.00018476187848187123, "loss": 4.4236, "step": 495750 }, { "epoch": 2.989592503708349, "grad_norm": 1.7373898029327393, "learning_rate": 0.0001847420831694842, "loss": 5.0611, "step": 495800 }, { "epoch": 2.989893995489683, "grad_norm": 1.9891407489776611, "learning_rate": 0.00018472228721771864, "loss": 5.4134, "step": 495850 }, { "epoch": 2.9901954872710172, "grad_norm": 2.872598648071289, "learning_rate": 0.00018470249062693897, "loss": 4.973, "step": 495900 }, { "epoch": 2.990496979052351, "grad_norm": 1.7280243635177612, "learning_rate": 0.00018468269339750947, "loss": 4.9203, "step": 495950 }, { "epoch": 2.990798470833685, "grad_norm": 2.3423805236816406, "learning_rate": 0.0001846628955297945, "loss": 5.2035, "step": 496000 }, { "epoch": 2.991099962615019, "grad_norm": 1.6107436418533325, "learning_rate": 0.00018464309702415842, "loss": 4.9897, "step": 496050 }, { "epoch": 2.991401454396353, "grad_norm": 1.824464201927185, "learning_rate": 0.00018462329788096558, "loss": 5.072, "step": 496100 }, { "epoch": 2.9917029461776874, "grad_norm": 1.6170806884765625, "learning_rate": 0.0001846034981005803, "loss": 4.7141, "step": 496150 }, { "epoch": 2.992004437959021, "grad_norm": 2.6109230518341064, "learning_rate": 0.00018458369768336707, "loss": 4.8377, "step": 496200 }, { "epoch": 2.9923059297403554, "grad_norm": 2.4420006275177, "learning_rate": 0.00018456389662969023, "loss": 5.301, "step": 496250 }, { "epoch": 2.992607421521689, "grad_norm": 1.8601607084274292, "learning_rate": 0.00018454409493991417, "loss": 4.8187, "step": 496300 }, { "epoch": 2.9929089133030233, "grad_norm": 1.2773469686508179, "learning_rate": 0.00018452429261440337, "loss": 5.0328, "step": 496350 }, { "epoch": 2.9932104050843575, "grad_norm": 1.2201679944992065, "learning_rate": 0.0001845044896535222, "loss": 4.8254, "step": 496400 }, { "epoch": 2.9935118968656913, "grad_norm": 1.3658162355422974, "learning_rate": 0.00018448468605763519, "loss": 5.312, "step": 496450 }, { "epoch": 2.9938133886470255, "grad_norm": 2.8359251022338867, "learning_rate": 0.0001844648818271067, "loss": 4.9624, "step": 496500 }, { "epoch": 2.9941148804283593, "grad_norm": 0.4894082546234131, "learning_rate": 0.0001844450769623013, "loss": 4.6497, "step": 496550 }, { "epoch": 2.9944163722096935, "grad_norm": 0.9307365417480469, "learning_rate": 0.0001844252714635833, "loss": 4.7669, "step": 496600 }, { "epoch": 2.9947178639910277, "grad_norm": 2.3539907932281494, "learning_rate": 0.00018440546533131743, "loss": 4.9415, "step": 496650 }, { "epoch": 2.995019355772362, "grad_norm": 2.179802894592285, "learning_rate": 0.00018438565856586804, "loss": 5.5236, "step": 496700 }, { "epoch": 2.9953208475536957, "grad_norm": 1.6066192388534546, "learning_rate": 0.0001843658511675996, "loss": 4.7623, "step": 496750 }, { "epoch": 2.99562233933503, "grad_norm": 2.5145695209503174, "learning_rate": 0.00018434604313687677, "loss": 4.8073, "step": 496800 }, { "epoch": 2.9959238311163636, "grad_norm": 4.29248571395874, "learning_rate": 0.00018432623447406404, "loss": 4.9915, "step": 496850 }, { "epoch": 2.996225322897698, "grad_norm": 2.097285032272339, "learning_rate": 0.00018430642517952594, "loss": 5.0591, "step": 496900 }, { "epoch": 2.996526814679032, "grad_norm": 4.816714286804199, "learning_rate": 0.00018428661525362705, "loss": 5.1404, "step": 496950 }, { "epoch": 2.996828306460366, "grad_norm": 1.496118187904358, "learning_rate": 0.0001842668046967319, "loss": 4.1761, "step": 497000 }, { "epoch": 2.9971297982417, "grad_norm": 1.5595402717590332, "learning_rate": 0.00018424699350920516, "loss": 5.0584, "step": 497050 }, { "epoch": 2.9974312900230338, "grad_norm": 0.748493492603302, "learning_rate": 0.00018422718169141137, "loss": 4.9975, "step": 497100 }, { "epoch": 2.997732781804368, "grad_norm": 2.3826181888580322, "learning_rate": 0.0001842073692437151, "loss": 5.2728, "step": 497150 }, { "epoch": 2.998034273585702, "grad_norm": 2.438774347305298, "learning_rate": 0.00018418755616648102, "loss": 5.242, "step": 497200 }, { "epoch": 2.998335765367036, "grad_norm": 2.1082332134246826, "learning_rate": 0.00018416774246007386, "loss": 5.0542, "step": 497250 }, { "epoch": 2.99863725714837, "grad_norm": 1.840404987335205, "learning_rate": 0.00018414792812485804, "loss": 5.4649, "step": 497300 }, { "epoch": 2.998938748929704, "grad_norm": 3.5620803833007812, "learning_rate": 0.00018412811316119838, "loss": 4.6603, "step": 497350 }, { "epoch": 2.999240240711038, "grad_norm": 1.1913255453109741, "learning_rate": 0.0001841082975694595, "loss": 5.0768, "step": 497400 }, { "epoch": 2.9995417324923723, "grad_norm": 1.479533076286316, "learning_rate": 0.00018408848135000605, "loss": 4.9465, "step": 497450 }, { "epoch": 2.9998432242737065, "grad_norm": 1.4596161842346191, "learning_rate": 0.0001840686645032028, "loss": 4.6417, "step": 497500 }, { "epoch": 3.0001447160550403, "grad_norm": 1.4864403009414673, "learning_rate": 0.0001840488470294144, "loss": 4.7512, "step": 497550 }, { "epoch": 3.0004462078363745, "grad_norm": 1.9388543367385864, "learning_rate": 0.00018402902892900548, "loss": 4.6872, "step": 497600 }, { "epoch": 3.0007476996177083, "grad_norm": 2.8901727199554443, "learning_rate": 0.00018400921020234092, "loss": 4.8983, "step": 497650 }, { "epoch": 3.0010491913990425, "grad_norm": 1.5081247091293335, "learning_rate": 0.00018398939084978536, "loss": 4.9624, "step": 497700 }, { "epoch": 3.0013506831803767, "grad_norm": 2.659144878387451, "learning_rate": 0.00018396957087170353, "loss": 5.0659, "step": 497750 }, { "epoch": 3.0016521749617104, "grad_norm": 1.7320916652679443, "learning_rate": 0.0001839497502684603, "loss": 4.9179, "step": 497800 }, { "epoch": 3.0019536667430446, "grad_norm": 2.668128728866577, "learning_rate": 0.0001839299290404203, "loss": 5.0474, "step": 497850 }, { "epoch": 3.0022551585243784, "grad_norm": 2.2403290271759033, "learning_rate": 0.00018391010718794842, "loss": 4.716, "step": 497900 }, { "epoch": 3.0025566503057126, "grad_norm": 1.3741767406463623, "learning_rate": 0.0001838902847114094, "loss": 4.9625, "step": 497950 }, { "epoch": 3.002858142087047, "grad_norm": 3.0742931365966797, "learning_rate": 0.00018387046161116803, "loss": 5.3137, "step": 498000 }, { "epoch": 3.002858142087047, "eval_loss": 5.381691932678223, "eval_runtime": 39.2225, "eval_samples_per_second": 13.054, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.4215618377510175, "step": 498000 }, { "epoch": 3.0031596338683806, "grad_norm": 1.573826551437378, "learning_rate": 0.0001838506378875892, "loss": 5.1397, "step": 498050 }, { "epoch": 3.003461125649715, "grad_norm": 3.299466133117676, "learning_rate": 0.00018383081354103766, "loss": 4.7521, "step": 498100 }, { "epoch": 3.003762617431049, "grad_norm": 0.8618778586387634, "learning_rate": 0.00018381098857187824, "loss": 4.9603, "step": 498150 }, { "epoch": 3.0040641092123828, "grad_norm": 2.811439037322998, "learning_rate": 0.00018379116298047584, "loss": 5.1484, "step": 498200 }, { "epoch": 3.004365600993717, "grad_norm": 2.699979066848755, "learning_rate": 0.00018377133676719537, "loss": 5.0627, "step": 498250 }, { "epoch": 3.004667092775051, "grad_norm": 1.7916966676712036, "learning_rate": 0.00018375150993240157, "loss": 4.9496, "step": 498300 }, { "epoch": 3.004968584556385, "grad_norm": 1.9163732528686523, "learning_rate": 0.00018373168247645943, "loss": 5.1034, "step": 498350 }, { "epoch": 3.005270076337719, "grad_norm": 2.351226568222046, "learning_rate": 0.00018371185439973377, "loss": 5.2124, "step": 498400 }, { "epoch": 3.005571568119053, "grad_norm": 1.8270293474197388, "learning_rate": 0.00018369202570258958, "loss": 4.9221, "step": 498450 }, { "epoch": 3.005873059900387, "grad_norm": 3.8952295780181885, "learning_rate": 0.00018367219638539174, "loss": 4.7009, "step": 498500 }, { "epoch": 3.0061745516817213, "grad_norm": 1.6766382455825806, "learning_rate": 0.00018365236644850516, "loss": 5.0049, "step": 498550 }, { "epoch": 3.006476043463055, "grad_norm": 1.768906831741333, "learning_rate": 0.0001836325358922948, "loss": 5.3932, "step": 498600 }, { "epoch": 3.0067775352443893, "grad_norm": 1.9220067262649536, "learning_rate": 0.00018361270471712565, "loss": 5.0119, "step": 498650 }, { "epoch": 3.0070790270257235, "grad_norm": 1.5263807773590088, "learning_rate": 0.00018359287292336258, "loss": 4.7764, "step": 498700 }, { "epoch": 3.0073805188070573, "grad_norm": 2.59165620803833, "learning_rate": 0.00018357304051137062, "loss": 4.8415, "step": 498750 }, { "epoch": 3.0076820105883915, "grad_norm": 2.1846845149993896, "learning_rate": 0.0001835532074815148, "loss": 5.0954, "step": 498800 }, { "epoch": 3.0079835023697252, "grad_norm": 2.5068957805633545, "learning_rate": 0.0001835333738341601, "loss": 5.0878, "step": 498850 }, { "epoch": 3.0082849941510594, "grad_norm": 2.5998706817626953, "learning_rate": 0.00018351353956967144, "loss": 4.925, "step": 498900 }, { "epoch": 3.0085864859323936, "grad_norm": 1.7896102666854858, "learning_rate": 0.00018349370468841397, "loss": 4.9881, "step": 498950 }, { "epoch": 3.0088879777137274, "grad_norm": 3.031702995300293, "learning_rate": 0.00018347386919075265, "loss": 5.0396, "step": 499000 }, { "epoch": 3.0091894694950616, "grad_norm": 1.7490640878677368, "learning_rate": 0.00018345403307705254, "loss": 5.0991, "step": 499050 }, { "epoch": 3.009490961276396, "grad_norm": 2.468308448791504, "learning_rate": 0.0001834341963476787, "loss": 4.9211, "step": 499100 }, { "epoch": 3.0097924530577296, "grad_norm": 2.3041183948516846, "learning_rate": 0.00018341435900299616, "loss": 4.7871, "step": 499150 }, { "epoch": 3.010093944839064, "grad_norm": 2.529087781906128, "learning_rate": 0.00018339452104337002, "loss": 4.8905, "step": 499200 }, { "epoch": 3.0103954366203975, "grad_norm": 0.4863596558570862, "learning_rate": 0.00018337468246916545, "loss": 5.1159, "step": 499250 }, { "epoch": 3.0106969284017318, "grad_norm": 1.5758169889450073, "learning_rate": 0.0001833548432807474, "loss": 4.824, "step": 499300 }, { "epoch": 3.010998420183066, "grad_norm": 1.441619634628296, "learning_rate": 0.00018333500347848106, "loss": 5.1131, "step": 499350 }, { "epoch": 3.0112999119643997, "grad_norm": 2.135317325592041, "learning_rate": 0.0001833151630627316, "loss": 4.7794, "step": 499400 }, { "epoch": 3.011601403745734, "grad_norm": 2.1248908042907715, "learning_rate": 0.00018329532203386408, "loss": 4.9712, "step": 499450 }, { "epoch": 3.011902895527068, "grad_norm": 2.023766040802002, "learning_rate": 0.00018327548039224367, "loss": 5.2207, "step": 499500 }, { "epoch": 3.012204387308402, "grad_norm": 1.0352814197540283, "learning_rate": 0.00018325563813823554, "loss": 4.5547, "step": 499550 }, { "epoch": 3.012505879089736, "grad_norm": 1.825132966041565, "learning_rate": 0.00018323579527220485, "loss": 4.9347, "step": 499600 }, { "epoch": 3.01280737087107, "grad_norm": 1.686265230178833, "learning_rate": 0.00018321595179451676, "loss": 4.6924, "step": 499650 }, { "epoch": 3.013108862652404, "grad_norm": 1.0884888172149658, "learning_rate": 0.00018319610770553652, "loss": 4.8138, "step": 499700 }, { "epoch": 3.0134103544337383, "grad_norm": 1.8047538995742798, "learning_rate": 0.00018317626300562925, "loss": 5.15, "step": 499750 }, { "epoch": 3.013711846215072, "grad_norm": 4.56895637512207, "learning_rate": 0.0001831564176951602, "loss": 5.0401, "step": 499800 }, { "epoch": 3.0140133379964062, "grad_norm": 2.041957378387451, "learning_rate": 0.00018313657177449464, "loss": 5.2045, "step": 499850 }, { "epoch": 3.0143148297777405, "grad_norm": 1.9469541311264038, "learning_rate": 0.00018311672524399776, "loss": 5.1465, "step": 499900 }, { "epoch": 3.014616321559074, "grad_norm": 2.0983736515045166, "learning_rate": 0.00018309687810403477, "loss": 5.0346, "step": 499950 }, { "epoch": 3.0149178133404084, "grad_norm": 1.8186582326889038, "learning_rate": 0.000183077030354971, "loss": 4.6953, "step": 500000 }, { "epoch": 3.015219305121742, "grad_norm": 1.0208760499954224, "learning_rate": 0.00018305718199717165, "loss": 4.9286, "step": 500050 }, { "epoch": 3.0155207969030764, "grad_norm": 1.5905505418777466, "learning_rate": 0.0001830373330310021, "loss": 4.6464, "step": 500100 }, { "epoch": 3.0158222886844106, "grad_norm": 1.9073454141616821, "learning_rate": 0.00018301748345682755, "loss": 4.703, "step": 500150 }, { "epoch": 3.0161237804657444, "grad_norm": 0.5704882144927979, "learning_rate": 0.00018299763327501336, "loss": 4.9089, "step": 500200 }, { "epoch": 3.0164252722470786, "grad_norm": 2.1431684494018555, "learning_rate": 0.0001829777824859248, "loss": 4.8649, "step": 500250 }, { "epoch": 3.0167267640284128, "grad_norm": 2.0339577198028564, "learning_rate": 0.00018295793108992722, "loss": 5.0511, "step": 500300 }, { "epoch": 3.0170282558097465, "grad_norm": 1.4360448122024536, "learning_rate": 0.00018293807908738592, "loss": 4.8857, "step": 500350 }, { "epoch": 3.0173297475910807, "grad_norm": 1.6108825206756592, "learning_rate": 0.0001829182264786663, "loss": 4.6435, "step": 500400 }, { "epoch": 3.0176312393724145, "grad_norm": 1.9373730421066284, "learning_rate": 0.00018289837326413374, "loss": 5.1485, "step": 500450 }, { "epoch": 3.0179327311537487, "grad_norm": 2.1841485500335693, "learning_rate": 0.0001828785194441535, "loss": 5.0388, "step": 500500 }, { "epoch": 3.018234222935083, "grad_norm": 1.1415400505065918, "learning_rate": 0.00018285866501909108, "loss": 4.7551, "step": 500550 }, { "epoch": 3.0185357147164167, "grad_norm": 1.9238280057907104, "learning_rate": 0.00018283880998931182, "loss": 4.8786, "step": 500600 }, { "epoch": 3.018837206497751, "grad_norm": 0.8078948259353638, "learning_rate": 0.0001828189543551811, "loss": 4.6714, "step": 500650 }, { "epoch": 3.019138698279085, "grad_norm": 1.8228732347488403, "learning_rate": 0.0001827990981170644, "loss": 5.0951, "step": 500700 }, { "epoch": 3.019440190060419, "grad_norm": 2.0170884132385254, "learning_rate": 0.00018277924127532708, "loss": 4.9641, "step": 500750 }, { "epoch": 3.019741681841753, "grad_norm": 2.477402925491333, "learning_rate": 0.00018275938383033454, "loss": 4.9767, "step": 500800 }, { "epoch": 3.020043173623087, "grad_norm": 1.5050145387649536, "learning_rate": 0.00018273952578245242, "loss": 4.8179, "step": 500850 }, { "epoch": 3.020344665404421, "grad_norm": 2.4933927059173584, "learning_rate": 0.00018271966713204597, "loss": 5.0369, "step": 500900 }, { "epoch": 3.0206461571857552, "grad_norm": 2.5156753063201904, "learning_rate": 0.00018269980787948071, "loss": 4.8164, "step": 500950 }, { "epoch": 3.020947648967089, "grad_norm": 0.8975661396980286, "learning_rate": 0.0001826799480251222, "loss": 4.7272, "step": 501000 }, { "epoch": 3.020947648967089, "eval_loss": 5.376887798309326, "eval_runtime": 39.2389, "eval_samples_per_second": 13.048, "eval_steps_per_second": 6.524, "eval_tts_loss": 7.448367356133929, "step": 501000 }, { "epoch": 3.021249140748423, "grad_norm": 1.5700774192810059, "learning_rate": 0.00018266008756933594, "loss": 5.1297, "step": 501050 }, { "epoch": 3.0215506325297574, "grad_norm": 1.2151265144348145, "learning_rate": 0.0001826402265124873, "loss": 4.8211, "step": 501100 }, { "epoch": 3.021852124311091, "grad_norm": 1.991579294204712, "learning_rate": 0.00018262036485494184, "loss": 4.6953, "step": 501150 }, { "epoch": 3.0221536160924254, "grad_norm": 1.7107423543930054, "learning_rate": 0.0001826005025970652, "loss": 4.7616, "step": 501200 }, { "epoch": 3.022455107873759, "grad_norm": 2.390552282333374, "learning_rate": 0.00018258063973922278, "loss": 5.2184, "step": 501250 }, { "epoch": 3.0227565996550934, "grad_norm": 2.811225414276123, "learning_rate": 0.0001825607762817802, "loss": 4.5556, "step": 501300 }, { "epoch": 3.0230580914364276, "grad_norm": 1.8208047151565552, "learning_rate": 0.000182540912225103, "loss": 4.8439, "step": 501350 }, { "epoch": 3.0233595832177613, "grad_norm": 1.655744194984436, "learning_rate": 0.0001825210475695567, "loss": 5.0473, "step": 501400 }, { "epoch": 3.0236610749990955, "grad_norm": 0.8965066075325012, "learning_rate": 0.00018250118231550704, "loss": 4.712, "step": 501450 }, { "epoch": 3.0239625667804297, "grad_norm": 2.507246255874634, "learning_rate": 0.00018248131646331943, "loss": 4.3747, "step": 501500 }, { "epoch": 3.0242640585617635, "grad_norm": 1.8145111799240112, "learning_rate": 0.00018246145001335953, "loss": 5.1531, "step": 501550 }, { "epoch": 3.0245655503430977, "grad_norm": 1.6065095663070679, "learning_rate": 0.00018244158296599295, "loss": 4.5853, "step": 501600 }, { "epoch": 3.0248670421244315, "grad_norm": 2.005120038986206, "learning_rate": 0.0001824217153215854, "loss": 4.8065, "step": 501650 }, { "epoch": 3.0251685339057657, "grad_norm": 1.827449083328247, "learning_rate": 0.0001824018470805024, "loss": 4.9596, "step": 501700 }, { "epoch": 3.0254700256871, "grad_norm": 1.7361226081848145, "learning_rate": 0.00018238197824310966, "loss": 4.9103, "step": 501750 }, { "epoch": 3.0257715174684336, "grad_norm": 1.5353577136993408, "learning_rate": 0.0001823621088097728, "loss": 4.6944, "step": 501800 }, { "epoch": 3.026073009249768, "grad_norm": 1.3334472179412842, "learning_rate": 0.00018234223878085758, "loss": 4.8357, "step": 501850 }, { "epoch": 3.026374501031102, "grad_norm": 2.2917609214782715, "learning_rate": 0.00018232236815672958, "loss": 5.0324, "step": 501900 }, { "epoch": 3.026675992812436, "grad_norm": 1.9038244485855103, "learning_rate": 0.00018230249693775447, "loss": 4.6687, "step": 501950 }, { "epoch": 3.02697748459377, "grad_norm": 0.5080597996711731, "learning_rate": 0.00018228262512429803, "loss": 4.8967, "step": 502000 }, { "epoch": 3.0272789763751042, "grad_norm": 1.889756441116333, "learning_rate": 0.00018226275271672598, "loss": 4.9778, "step": 502050 }, { "epoch": 3.027580468156438, "grad_norm": 2.034696340560913, "learning_rate": 0.00018224287971540392, "loss": 4.492, "step": 502100 }, { "epoch": 3.027881959937772, "grad_norm": 1.767118215560913, "learning_rate": 0.00018222300612069775, "loss": 4.7791, "step": 502150 }, { "epoch": 3.028183451719106, "grad_norm": 1.4375756978988647, "learning_rate": 0.00018220313193297312, "loss": 4.9083, "step": 502200 }, { "epoch": 3.02848494350044, "grad_norm": 2.269655227661133, "learning_rate": 0.0001821832571525958, "loss": 4.6772, "step": 502250 }, { "epoch": 3.0287864352817744, "grad_norm": 1.7222604751586914, "learning_rate": 0.00018216338177993156, "loss": 4.8771, "step": 502300 }, { "epoch": 3.029087927063108, "grad_norm": 1.291981816291809, "learning_rate": 0.0001821435058153462, "loss": 4.6701, "step": 502350 }, { "epoch": 3.0293894188444423, "grad_norm": 1.3032926321029663, "learning_rate": 0.00018212362925920542, "loss": 5.0484, "step": 502400 }, { "epoch": 3.0296909106257766, "grad_norm": 1.6158177852630615, "learning_rate": 0.00018210375211187517, "loss": 5.2317, "step": 502450 }, { "epoch": 3.0299924024071103, "grad_norm": 1.6385760307312012, "learning_rate": 0.00018208387437372112, "loss": 4.8176, "step": 502500 }, { "epoch": 3.0302938941884445, "grad_norm": 1.7103458642959595, "learning_rate": 0.00018206399604510915, "loss": 5.1799, "step": 502550 }, { "epoch": 3.0305953859697783, "grad_norm": 2.5777103900909424, "learning_rate": 0.0001820441171264051, "loss": 4.6863, "step": 502600 }, { "epoch": 3.0308968777511125, "grad_norm": 1.6946381330490112, "learning_rate": 0.00018202423761797483, "loss": 4.7993, "step": 502650 }, { "epoch": 3.0311983695324467, "grad_norm": 2.490407705307007, "learning_rate": 0.00018200435752018415, "loss": 4.8094, "step": 502700 }, { "epoch": 3.0314998613137805, "grad_norm": 1.8549646139144897, "learning_rate": 0.0001819844768333989, "loss": 4.219, "step": 502750 }, { "epoch": 3.0318013530951147, "grad_norm": 2.1018435955047607, "learning_rate": 0.00018196459555798507, "loss": 5.0192, "step": 502800 }, { "epoch": 3.032102844876449, "grad_norm": 1.5146703720092773, "learning_rate": 0.00018194471369430845, "loss": 5.3858, "step": 502850 }, { "epoch": 3.0324043366577826, "grad_norm": 2.5559823513031006, "learning_rate": 0.00018192483124273497, "loss": 4.9652, "step": 502900 }, { "epoch": 3.032705828439117, "grad_norm": 1.1558301448822021, "learning_rate": 0.0001819049482036305, "loss": 4.8214, "step": 502950 }, { "epoch": 3.0330073202204506, "grad_norm": 1.5501089096069336, "learning_rate": 0.000181885064577361, "loss": 4.9807, "step": 503000 }, { "epoch": 3.033308812001785, "grad_norm": 3.331268310546875, "learning_rate": 0.0001818651803642924, "loss": 4.7307, "step": 503050 }, { "epoch": 3.033610303783119, "grad_norm": 1.7236475944519043, "learning_rate": 0.00018184529556479064, "loss": 4.8355, "step": 503100 }, { "epoch": 3.033911795564453, "grad_norm": 2.0906119346618652, "learning_rate": 0.0001818254101792216, "loss": 4.8694, "step": 503150 }, { "epoch": 3.034213287345787, "grad_norm": 2.0708370208740234, "learning_rate": 0.00018180552420795135, "loss": 5.1379, "step": 503200 }, { "epoch": 3.034514779127121, "grad_norm": 1.2787425518035889, "learning_rate": 0.00018178563765134585, "loss": 4.9066, "step": 503250 }, { "epoch": 3.034816270908455, "grad_norm": 0.7775710821151733, "learning_rate": 0.00018176575050977097, "loss": 5.1383, "step": 503300 }, { "epoch": 3.035117762689789, "grad_norm": 1.5356943607330322, "learning_rate": 0.00018174586278359286, "loss": 4.5342, "step": 503350 }, { "epoch": 3.035419254471123, "grad_norm": 3.6990387439727783, "learning_rate": 0.0001817259744731774, "loss": 5.0288, "step": 503400 }, { "epoch": 3.035720746252457, "grad_norm": 2.1417107582092285, "learning_rate": 0.00018170608557889067, "loss": 4.4789, "step": 503450 }, { "epoch": 3.0360222380337913, "grad_norm": 2.0781326293945312, "learning_rate": 0.00018168619610109868, "loss": 4.9058, "step": 503500 }, { "epoch": 3.036323729815125, "grad_norm": 1.6676075458526611, "learning_rate": 0.00018166630604016747, "loss": 4.6271, "step": 503550 }, { "epoch": 3.0366252215964593, "grad_norm": 1.9187462329864502, "learning_rate": 0.00018164641539646307, "loss": 5.2877, "step": 503600 }, { "epoch": 3.0369267133777935, "grad_norm": 3.15153431892395, "learning_rate": 0.00018162652417035163, "loss": 5.3054, "step": 503650 }, { "epoch": 3.0372282051591273, "grad_norm": 1.6808408498764038, "learning_rate": 0.00018160663236219904, "loss": 4.7162, "step": 503700 }, { "epoch": 3.0375296969404615, "grad_norm": 1.5923329591751099, "learning_rate": 0.00018158673997237156, "loss": 4.5152, "step": 503750 }, { "epoch": 3.0378311887217953, "grad_norm": 0.626876950263977, "learning_rate": 0.00018156684700123518, "loss": 4.9884, "step": 503800 }, { "epoch": 3.0381326805031295, "grad_norm": 1.1745069026947021, "learning_rate": 0.00018154695344915604, "loss": 4.8205, "step": 503850 }, { "epoch": 3.0384341722844637, "grad_norm": 1.006983757019043, "learning_rate": 0.00018152705931650024, "loss": 5.0681, "step": 503900 }, { "epoch": 3.0387356640657974, "grad_norm": 1.0837091207504272, "learning_rate": 0.00018150716460363391, "loss": 4.6059, "step": 503950 }, { "epoch": 3.0390371558471316, "grad_norm": 1.6878939867019653, "learning_rate": 0.00018148726931092312, "loss": 4.7997, "step": 504000 }, { "epoch": 3.0390371558471316, "eval_loss": 5.373140335083008, "eval_runtime": 39.0375, "eval_samples_per_second": 13.116, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.528159624775351, "step": 504000 }, { "epoch": 3.039338647628466, "grad_norm": 2.5848653316497803, "learning_rate": 0.00018146737343873416, "loss": 4.833, "step": 504050 }, { "epoch": 3.0396401394097996, "grad_norm": 2.777703285217285, "learning_rate": 0.00018144747698743308, "loss": 5.0643, "step": 504100 }, { "epoch": 3.039941631191134, "grad_norm": 2.6847121715545654, "learning_rate": 0.00018142757995738603, "loss": 4.9776, "step": 504150 }, { "epoch": 3.0402431229724676, "grad_norm": 2.1072375774383545, "learning_rate": 0.0001814076823489592, "loss": 4.9113, "step": 504200 }, { "epoch": 3.0405446147538018, "grad_norm": 2.1224451065063477, "learning_rate": 0.00018138778416251888, "loss": 5.094, "step": 504250 }, { "epoch": 3.040846106535136, "grad_norm": 1.7856453657150269, "learning_rate": 0.0001813678853984311, "loss": 5.119, "step": 504300 }, { "epoch": 3.0411475983164697, "grad_norm": 1.5345622301101685, "learning_rate": 0.0001813479860570622, "loss": 4.9254, "step": 504350 }, { "epoch": 3.041449090097804, "grad_norm": 3.0522661209106445, "learning_rate": 0.00018132808613877837, "loss": 4.7907, "step": 504400 }, { "epoch": 3.041750581879138, "grad_norm": 2.2625670433044434, "learning_rate": 0.0001813081856439458, "loss": 4.7904, "step": 504450 }, { "epoch": 3.042052073660472, "grad_norm": 1.5893597602844238, "learning_rate": 0.00018128828457293075, "loss": 4.7735, "step": 504500 }, { "epoch": 3.042353565441806, "grad_norm": 1.831558108329773, "learning_rate": 0.00018126838292609947, "loss": 4.5872, "step": 504550 }, { "epoch": 3.04265505722314, "grad_norm": 2.0293803215026855, "learning_rate": 0.0001812484807038182, "loss": 5.1745, "step": 504600 }, { "epoch": 3.042956549004474, "grad_norm": 2.224057674407959, "learning_rate": 0.0001812285779064533, "loss": 4.5617, "step": 504650 }, { "epoch": 3.0432580407858083, "grad_norm": 1.7800801992416382, "learning_rate": 0.00018120867453437096, "loss": 5.0772, "step": 504700 }, { "epoch": 3.043559532567142, "grad_norm": 2.2397148609161377, "learning_rate": 0.00018118877058793747, "loss": 4.729, "step": 504750 }, { "epoch": 3.0438610243484763, "grad_norm": 2.0634939670562744, "learning_rate": 0.00018116886606751916, "loss": 5.3184, "step": 504800 }, { "epoch": 3.0441625161298105, "grad_norm": 0.7248588800430298, "learning_rate": 0.00018114896097348244, "loss": 4.9228, "step": 504850 }, { "epoch": 3.0444640079111442, "grad_norm": 2.3500306606292725, "learning_rate": 0.00018112905530619344, "loss": 4.9355, "step": 504900 }, { "epoch": 3.0447654996924784, "grad_norm": 1.9773428440093994, "learning_rate": 0.0001811091490660187, "loss": 5.0213, "step": 504950 }, { "epoch": 3.045066991473812, "grad_norm": 1.7514702081680298, "learning_rate": 0.0001810892422533244, "loss": 4.6836, "step": 505000 }, { "epoch": 3.0453684832551464, "grad_norm": 2.3517961502075195, "learning_rate": 0.00018106933486847698, "loss": 5.0145, "step": 505050 }, { "epoch": 3.0456699750364806, "grad_norm": 1.6812589168548584, "learning_rate": 0.00018104942691184276, "loss": 4.8572, "step": 505100 }, { "epoch": 3.0459714668178144, "grad_norm": 2.2325222492218018, "learning_rate": 0.00018102951838378818, "loss": 4.6042, "step": 505150 }, { "epoch": 3.0462729585991486, "grad_norm": 1.5388919115066528, "learning_rate": 0.00018100960928467958, "loss": 4.703, "step": 505200 }, { "epoch": 3.046574450380483, "grad_norm": 2.3851988315582275, "learning_rate": 0.0001809896996148834, "loss": 4.8803, "step": 505250 }, { "epoch": 3.0468759421618166, "grad_norm": 1.865667700767517, "learning_rate": 0.00018096978937476603, "loss": 4.7541, "step": 505300 }, { "epoch": 3.0471774339431508, "grad_norm": 3.1380186080932617, "learning_rate": 0.00018094987856469383, "loss": 4.5507, "step": 505350 }, { "epoch": 3.0474789257244845, "grad_norm": 2.2950427532196045, "learning_rate": 0.0001809299671850333, "loss": 4.9529, "step": 505400 }, { "epoch": 3.0477804175058187, "grad_norm": 2.383563756942749, "learning_rate": 0.00018091005523615089, "loss": 4.8944, "step": 505450 }, { "epoch": 3.048081909287153, "grad_norm": 2.0879080295562744, "learning_rate": 0.000180890142718413, "loss": 4.9796, "step": 505500 }, { "epoch": 3.0483834010684867, "grad_norm": 2.4346086978912354, "learning_rate": 0.00018087022963218614, "loss": 4.6274, "step": 505550 }, { "epoch": 3.048684892849821, "grad_norm": 1.8588441610336304, "learning_rate": 0.00018085031597783676, "loss": 4.9684, "step": 505600 }, { "epoch": 3.048986384631155, "grad_norm": 1.5167957544326782, "learning_rate": 0.00018083040175573135, "loss": 4.9787, "step": 505650 }, { "epoch": 3.049287876412489, "grad_norm": 1.5903289318084717, "learning_rate": 0.00018081048696623635, "loss": 4.7447, "step": 505700 }, { "epoch": 3.049589368193823, "grad_norm": 1.7948825359344482, "learning_rate": 0.00018079057160971828, "loss": 4.8692, "step": 505750 }, { "epoch": 3.0498908599751573, "grad_norm": 1.9242061376571655, "learning_rate": 0.0001807706556865437, "loss": 5.178, "step": 505800 }, { "epoch": 3.050192351756491, "grad_norm": 1.5775941610336304, "learning_rate": 0.0001807507391970792, "loss": 4.9976, "step": 505850 }, { "epoch": 3.0504938435378253, "grad_norm": 1.9957929849624634, "learning_rate": 0.00018073082214169116, "loss": 4.8527, "step": 505900 }, { "epoch": 3.050795335319159, "grad_norm": 2.659717082977295, "learning_rate": 0.0001807109045207462, "loss": 5.1421, "step": 505950 }, { "epoch": 3.0510968271004932, "grad_norm": 2.0671780109405518, "learning_rate": 0.00018069098633461086, "loss": 4.7882, "step": 506000 }, { "epoch": 3.0513983188818274, "grad_norm": 1.6373318433761597, "learning_rate": 0.00018067106758365176, "loss": 5.1399, "step": 506050 }, { "epoch": 3.051699810663161, "grad_norm": 1.5277400016784668, "learning_rate": 0.00018065114826823538, "loss": 5.2516, "step": 506100 }, { "epoch": 3.0520013024444954, "grad_norm": 0.8223810791969299, "learning_rate": 0.00018063122838872837, "loss": 5.0115, "step": 506150 }, { "epoch": 3.0523027942258296, "grad_norm": 3.490838050842285, "learning_rate": 0.00018061130794549732, "loss": 5.001, "step": 506200 }, { "epoch": 3.0526042860071634, "grad_norm": 1.5923932790756226, "learning_rate": 0.00018059138693890883, "loss": 4.7621, "step": 506250 }, { "epoch": 3.0529057777884976, "grad_norm": 1.0493500232696533, "learning_rate": 0.00018057146536932953, "loss": 4.624, "step": 506300 }, { "epoch": 3.0532072695698314, "grad_norm": 2.856044054031372, "learning_rate": 0.000180551543237126, "loss": 4.947, "step": 506350 }, { "epoch": 3.0535087613511656, "grad_norm": 2.068650722503662, "learning_rate": 0.00018053162054266494, "loss": 5.0436, "step": 506400 }, { "epoch": 3.0538102531324998, "grad_norm": 2.624952554702759, "learning_rate": 0.000180511697286313, "loss": 5.0425, "step": 506450 }, { "epoch": 3.0541117449138335, "grad_norm": 1.9317359924316406, "learning_rate": 0.0001804917734684368, "loss": 4.5854, "step": 506500 }, { "epoch": 3.0544132366951677, "grad_norm": 2.1850743293762207, "learning_rate": 0.000180471849089403, "loss": 4.7485, "step": 506550 }, { "epoch": 3.054714728476502, "grad_norm": 2.544551372528076, "learning_rate": 0.00018045192414957833, "loss": 4.9898, "step": 506600 }, { "epoch": 3.0550162202578357, "grad_norm": 2.7277963161468506, "learning_rate": 0.00018043199864932943, "loss": 5.0515, "step": 506650 }, { "epoch": 3.05531771203917, "grad_norm": 1.957205057144165, "learning_rate": 0.00018041207258902303, "loss": 5.0038, "step": 506700 }, { "epoch": 3.0556192038205037, "grad_norm": 1.099328875541687, "learning_rate": 0.00018039214596902585, "loss": 5.3026, "step": 506750 }, { "epoch": 3.055920695601838, "grad_norm": 0.475757360458374, "learning_rate": 0.00018037221878970457, "loss": 4.5809, "step": 506800 }, { "epoch": 3.056222187383172, "grad_norm": 0.4178772270679474, "learning_rate": 0.000180352291051426, "loss": 5.0165, "step": 506850 }, { "epoch": 3.056523679164506, "grad_norm": 0.9470672607421875, "learning_rate": 0.0001803323627545568, "loss": 4.8866, "step": 506900 }, { "epoch": 3.05682517094584, "grad_norm": 2.496933937072754, "learning_rate": 0.0001803124338994637, "loss": 5.0234, "step": 506950 }, { "epoch": 3.0571266627271743, "grad_norm": 0.622715175151825, "learning_rate": 0.00018029250448651355, "loss": 4.6171, "step": 507000 }, { "epoch": 3.0571266627271743, "eval_loss": 5.361151695251465, "eval_runtime": 39.0861, "eval_samples_per_second": 13.099, "eval_steps_per_second": 6.55, "eval_tts_loss": 7.513152417024582, "step": 507000 }, { "epoch": 3.057428154508508, "grad_norm": 1.6904280185699463, "learning_rate": 0.00018027257451607315, "loss": 4.3148, "step": 507050 }, { "epoch": 3.0577296462898422, "grad_norm": 2.5270919799804688, "learning_rate": 0.0001802526439885091, "loss": 5.184, "step": 507100 }, { "epoch": 3.058031138071176, "grad_norm": 1.7054483890533447, "learning_rate": 0.00018023271290418838, "loss": 4.9257, "step": 507150 }, { "epoch": 3.05833262985251, "grad_norm": 3.066349744796753, "learning_rate": 0.0001802127812634777, "loss": 4.8631, "step": 507200 }, { "epoch": 3.0586341216338444, "grad_norm": 2.1741251945495605, "learning_rate": 0.00018019284906674391, "loss": 5.1621, "step": 507250 }, { "epoch": 3.058935613415178, "grad_norm": 1.8478280305862427, "learning_rate": 0.00018017291631435383, "loss": 4.8235, "step": 507300 }, { "epoch": 3.0592371051965124, "grad_norm": 1.761544108390808, "learning_rate": 0.00018015298300667423, "loss": 4.9085, "step": 507350 }, { "epoch": 3.0595385969778466, "grad_norm": 1.830783724784851, "learning_rate": 0.00018013304914407205, "loss": 5.0102, "step": 507400 }, { "epoch": 3.0598400887591803, "grad_norm": 2.858365297317505, "learning_rate": 0.00018011311472691414, "loss": 4.7361, "step": 507450 }, { "epoch": 3.0601415805405145, "grad_norm": 2.4380555152893066, "learning_rate": 0.00018009317975556724, "loss": 4.9298, "step": 507500 }, { "epoch": 3.0604430723218483, "grad_norm": 0.8070403337478638, "learning_rate": 0.0001800732442303984, "loss": 4.6202, "step": 507550 }, { "epoch": 3.0607445641031825, "grad_norm": 3.084885835647583, "learning_rate": 0.0001800533081517744, "loss": 4.8847, "step": 507600 }, { "epoch": 3.0610460558845167, "grad_norm": 1.61770498752594, "learning_rate": 0.0001800333715200621, "loss": 5.2132, "step": 507650 }, { "epoch": 3.0613475476658505, "grad_norm": 2.241644859313965, "learning_rate": 0.0001800134343356285, "loss": 4.499, "step": 507700 }, { "epoch": 3.0616490394471847, "grad_norm": 1.3864457607269287, "learning_rate": 0.00017999349659884046, "loss": 4.8771, "step": 507750 }, { "epoch": 3.061950531228519, "grad_norm": 2.082240104675293, "learning_rate": 0.00017997355831006492, "loss": 5.0237, "step": 507800 }, { "epoch": 3.0622520230098527, "grad_norm": 2.8078794479370117, "learning_rate": 0.0001799536194696688, "loss": 4.9001, "step": 507850 }, { "epoch": 3.062553514791187, "grad_norm": 0.6412045955657959, "learning_rate": 0.00017993368007801906, "loss": 4.7499, "step": 507900 }, { "epoch": 3.0628550065725206, "grad_norm": 3.427361011505127, "learning_rate": 0.00017991374013548268, "loss": 4.6734, "step": 507950 }, { "epoch": 3.063156498353855, "grad_norm": 1.7506808042526245, "learning_rate": 0.00017989379964242658, "loss": 4.4322, "step": 508000 }, { "epoch": 3.063457990135189, "grad_norm": 1.441367268562317, "learning_rate": 0.0001798738585992178, "loss": 5.2253, "step": 508050 }, { "epoch": 3.063759481916523, "grad_norm": 2.217108726501465, "learning_rate": 0.0001798539170062232, "loss": 4.7725, "step": 508100 }, { "epoch": 3.064060973697857, "grad_norm": 2.0856666564941406, "learning_rate": 0.00017983397486380991, "loss": 5.1836, "step": 508150 }, { "epoch": 3.0643624654791912, "grad_norm": 1.6223446130752563, "learning_rate": 0.0001798140321723449, "loss": 4.8512, "step": 508200 }, { "epoch": 3.064663957260525, "grad_norm": 1.4756176471710205, "learning_rate": 0.0001797940889321952, "loss": 4.9062, "step": 508250 }, { "epoch": 3.064965449041859, "grad_norm": 3.321669101715088, "learning_rate": 0.00017977414514372777, "loss": 5.0052, "step": 508300 }, { "epoch": 3.065266940823193, "grad_norm": 1.6960268020629883, "learning_rate": 0.0001797542008073097, "loss": 4.697, "step": 508350 }, { "epoch": 3.065568432604527, "grad_norm": 2.7523341178894043, "learning_rate": 0.00017973425592330802, "loss": 5.1713, "step": 508400 }, { "epoch": 3.0658699243858614, "grad_norm": 0.5727407336235046, "learning_rate": 0.0001797143104920898, "loss": 5.2629, "step": 508450 }, { "epoch": 3.066171416167195, "grad_norm": 0.8578125834465027, "learning_rate": 0.0001796943645140221, "loss": 5.1692, "step": 508500 }, { "epoch": 3.0664729079485293, "grad_norm": 1.7845057249069214, "learning_rate": 0.00017967441798947197, "loss": 5.2826, "step": 508550 }, { "epoch": 3.0667743997298635, "grad_norm": 1.695704698562622, "learning_rate": 0.00017965447091880654, "loss": 4.8797, "step": 508600 }, { "epoch": 3.0670758915111973, "grad_norm": 2.082819938659668, "learning_rate": 0.00017963452330239292, "loss": 4.7677, "step": 508650 }, { "epoch": 3.0673773832925315, "grad_norm": 1.8578284978866577, "learning_rate": 0.00017961457514059816, "loss": 5.0597, "step": 508700 }, { "epoch": 3.0676788750738653, "grad_norm": 2.4457359313964844, "learning_rate": 0.00017959462643378938, "loss": 4.9321, "step": 508750 }, { "epoch": 3.0679803668551995, "grad_norm": 1.8822880983352661, "learning_rate": 0.00017957467718233378, "loss": 5.2308, "step": 508800 }, { "epoch": 3.0682818586365337, "grad_norm": 1.5519551038742065, "learning_rate": 0.00017955472738659847, "loss": 5.176, "step": 508850 }, { "epoch": 3.0685833504178675, "grad_norm": 1.7564600706100464, "learning_rate": 0.00017953477704695054, "loss": 5.0235, "step": 508900 }, { "epoch": 3.0688848421992017, "grad_norm": 1.3695615530014038, "learning_rate": 0.00017951482616375717, "loss": 5.1223, "step": 508950 }, { "epoch": 3.069186333980536, "grad_norm": 1.2584507465362549, "learning_rate": 0.00017949487473738562, "loss": 4.8739, "step": 509000 }, { "epoch": 3.0694878257618696, "grad_norm": 2.5168488025665283, "learning_rate": 0.00017947492276820294, "loss": 4.8249, "step": 509050 }, { "epoch": 3.069789317543204, "grad_norm": 3.6840097904205322, "learning_rate": 0.00017945497025657634, "loss": 4.7329, "step": 509100 }, { "epoch": 3.0700908093245376, "grad_norm": 1.7257299423217773, "learning_rate": 0.00017943501720287306, "loss": 4.7453, "step": 509150 }, { "epoch": 3.070392301105872, "grad_norm": 1.6011005640029907, "learning_rate": 0.00017941506360746035, "loss": 5.0214, "step": 509200 }, { "epoch": 3.070693792887206, "grad_norm": 2.267728090286255, "learning_rate": 0.00017939510947070531, "loss": 5.0457, "step": 509250 }, { "epoch": 3.0709952846685398, "grad_norm": 2.17744779586792, "learning_rate": 0.0001793751547929753, "loss": 4.9498, "step": 509300 }, { "epoch": 3.071296776449874, "grad_norm": 1.6431134939193726, "learning_rate": 0.0001793551995746374, "loss": 4.665, "step": 509350 }, { "epoch": 3.071598268231208, "grad_norm": 1.5070064067840576, "learning_rate": 0.00017933524381605898, "loss": 4.7301, "step": 509400 }, { "epoch": 3.071899760012542, "grad_norm": 1.8689579963684082, "learning_rate": 0.0001793152875176073, "loss": 5.0314, "step": 509450 }, { "epoch": 3.072201251793876, "grad_norm": 2.1895644664764404, "learning_rate": 0.00017929533067964954, "loss": 5.2754, "step": 509500 }, { "epoch": 3.0725027435752104, "grad_norm": 1.7395596504211426, "learning_rate": 0.00017927537330255302, "loss": 5.068, "step": 509550 }, { "epoch": 3.072804235356544, "grad_norm": 1.125361680984497, "learning_rate": 0.00017925541538668505, "loss": 4.7004, "step": 509600 }, { "epoch": 3.0731057271378783, "grad_norm": 1.6038033962249756, "learning_rate": 0.00017923545693241297, "loss": 5.1704, "step": 509650 }, { "epoch": 3.073407218919212, "grad_norm": 0.8972242474555969, "learning_rate": 0.0001792154979401039, "loss": 4.8475, "step": 509700 }, { "epoch": 3.0737087107005463, "grad_norm": 1.956491470336914, "learning_rate": 0.00017919553841012537, "loss": 5.0539, "step": 509750 }, { "epoch": 3.0740102024818805, "grad_norm": 2.528501272201538, "learning_rate": 0.00017917557834284462, "loss": 4.6503, "step": 509800 }, { "epoch": 3.0743116942632143, "grad_norm": 1.5828711986541748, "learning_rate": 0.00017915561773862896, "loss": 4.632, "step": 509850 }, { "epoch": 3.0746131860445485, "grad_norm": 2.8793416023254395, "learning_rate": 0.00017913565659784575, "loss": 5.2366, "step": 509900 }, { "epoch": 3.0749146778258827, "grad_norm": 2.045189380645752, "learning_rate": 0.0001791156949208624, "loss": 4.8478, "step": 509950 }, { "epoch": 3.0752161696072164, "grad_norm": 2.4757306575775146, "learning_rate": 0.00017909573270804625, "loss": 4.9662, "step": 510000 }, { "epoch": 3.0752161696072164, "eval_loss": 5.3518171310424805, "eval_runtime": 39.1167, "eval_samples_per_second": 13.089, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.484570886542199, "step": 510000 }, { "epoch": 3.0755176613885506, "grad_norm": 2.6267898082733154, "learning_rate": 0.00017907576995976462, "loss": 5.0153, "step": 510050 }, { "epoch": 3.0758191531698844, "grad_norm": 2.440662145614624, "learning_rate": 0.00017905580667638494, "loss": 4.47, "step": 510100 }, { "epoch": 3.0761206449512186, "grad_norm": 2.1118106842041016, "learning_rate": 0.0001790358428582746, "loss": 5.3016, "step": 510150 }, { "epoch": 3.076422136732553, "grad_norm": 2.3198277950286865, "learning_rate": 0.000179015878505801, "loss": 4.9516, "step": 510200 }, { "epoch": 3.0767236285138866, "grad_norm": 1.5075132846832275, "learning_rate": 0.0001789959136193316, "loss": 5.0141, "step": 510250 }, { "epoch": 3.077025120295221, "grad_norm": 1.824207067489624, "learning_rate": 0.00017897594819923376, "loss": 4.3366, "step": 510300 }, { "epoch": 3.077326612076555, "grad_norm": 0.9874511361122131, "learning_rate": 0.00017895598224587492, "loss": 4.9459, "step": 510350 }, { "epoch": 3.0776281038578888, "grad_norm": 1.31158447265625, "learning_rate": 0.00017893601575962262, "loss": 5.0644, "step": 510400 }, { "epoch": 3.077929595639223, "grad_norm": 1.9430878162384033, "learning_rate": 0.0001789160487408442, "loss": 4.2198, "step": 510450 }, { "epoch": 3.0782310874205567, "grad_norm": 1.4978742599487305, "learning_rate": 0.00017889608118990718, "loss": 4.7151, "step": 510500 }, { "epoch": 3.078532579201891, "grad_norm": 1.7617754936218262, "learning_rate": 0.000178876113107179, "loss": 4.9477, "step": 510550 }, { "epoch": 3.078834070983225, "grad_norm": 1.7549968957901, "learning_rate": 0.00017885614449302719, "loss": 5.0374, "step": 510600 }, { "epoch": 3.079135562764559, "grad_norm": 0.5843079686164856, "learning_rate": 0.00017883617534781925, "loss": 5.0219, "step": 510650 }, { "epoch": 3.079437054545893, "grad_norm": 1.873913288116455, "learning_rate": 0.00017881620567192257, "loss": 4.9807, "step": 510700 }, { "epoch": 3.0797385463272273, "grad_norm": 1.4491641521453857, "learning_rate": 0.0001787962354657048, "loss": 5.1749, "step": 510750 }, { "epoch": 3.080040038108561, "grad_norm": 1.6106756925582886, "learning_rate": 0.00017877626472953342, "loss": 4.883, "step": 510800 }, { "epoch": 3.0803415298898953, "grad_norm": 2.037381887435913, "learning_rate": 0.00017875629346377592, "loss": 5.0356, "step": 510850 }, { "epoch": 3.080643021671229, "grad_norm": 2.0825600624084473, "learning_rate": 0.00017873632166879992, "loss": 5.025, "step": 510900 }, { "epoch": 3.0809445134525633, "grad_norm": 1.5699892044067383, "learning_rate": 0.00017871634934497291, "loss": 4.9915, "step": 510950 }, { "epoch": 3.0812460052338975, "grad_norm": 0.4653213620185852, "learning_rate": 0.00017869637649266243, "loss": 5.0496, "step": 511000 }, { "epoch": 3.0815474970152312, "grad_norm": 2.0331625938415527, "learning_rate": 0.00017867640311223615, "loss": 4.6954, "step": 511050 }, { "epoch": 3.0818489887965654, "grad_norm": 1.656111478805542, "learning_rate": 0.00017865642920406158, "loss": 5.1338, "step": 511100 }, { "epoch": 3.0821504805778996, "grad_norm": 2.95934796333313, "learning_rate": 0.00017863645476850627, "loss": 5.2312, "step": 511150 }, { "epoch": 3.0824519723592334, "grad_norm": 2.2225866317749023, "learning_rate": 0.000178616479805938, "loss": 4.6713, "step": 511200 }, { "epoch": 3.0827534641405676, "grad_norm": 1.0578502416610718, "learning_rate": 0.00017859650431672415, "loss": 4.8955, "step": 511250 }, { "epoch": 3.0830549559219014, "grad_norm": 2.030707597732544, "learning_rate": 0.00017857652830123247, "loss": 4.7844, "step": 511300 }, { "epoch": 3.0833564477032356, "grad_norm": 1.242822527885437, "learning_rate": 0.00017855655175983058, "loss": 4.7685, "step": 511350 }, { "epoch": 3.08365793948457, "grad_norm": 0.7153142690658569, "learning_rate": 0.00017853657469288608, "loss": 5.1193, "step": 511400 }, { "epoch": 3.0839594312659036, "grad_norm": 1.2739323377609253, "learning_rate": 0.0001785165971007667, "loss": 4.9605, "step": 511450 }, { "epoch": 3.0842609230472378, "grad_norm": 1.825859546661377, "learning_rate": 0.00017849661898384, "loss": 4.5678, "step": 511500 }, { "epoch": 3.084562414828572, "grad_norm": 2.970977306365967, "learning_rate": 0.00017847664034247374, "loss": 5.0232, "step": 511550 }, { "epoch": 3.0848639066099057, "grad_norm": 1.281714916229248, "learning_rate": 0.0001784566611770355, "loss": 4.65, "step": 511600 }, { "epoch": 3.08516539839124, "grad_norm": 2.290447235107422, "learning_rate": 0.00017843668148789308, "loss": 4.7982, "step": 511650 }, { "epoch": 3.0854668901725737, "grad_norm": 2.700230360031128, "learning_rate": 0.00017841670127541408, "loss": 4.5855, "step": 511700 }, { "epoch": 3.085768381953908, "grad_norm": 2.697862148284912, "learning_rate": 0.00017839672053996622, "loss": 5.5271, "step": 511750 }, { "epoch": 3.086069873735242, "grad_norm": 2.7240264415740967, "learning_rate": 0.00017837673928191726, "loss": 4.9299, "step": 511800 }, { "epoch": 3.086371365516576, "grad_norm": 2.0752100944519043, "learning_rate": 0.00017835675750163494, "loss": 5.1215, "step": 511850 }, { "epoch": 3.08667285729791, "grad_norm": 1.0073298215866089, "learning_rate": 0.00017833677519948693, "loss": 4.5872, "step": 511900 }, { "epoch": 3.0869743490792443, "grad_norm": 2.240542411804199, "learning_rate": 0.00017831679237584105, "loss": 4.628, "step": 511950 }, { "epoch": 3.087275840860578, "grad_norm": 2.14672589302063, "learning_rate": 0.00017829680903106498, "loss": 4.9064, "step": 512000 }, { "epoch": 3.0875773326419123, "grad_norm": 1.5857717990875244, "learning_rate": 0.00017827682516552654, "loss": 5.1993, "step": 512050 }, { "epoch": 3.087878824423246, "grad_norm": 1.482094168663025, "learning_rate": 0.0001782568407795935, "loss": 4.7126, "step": 512100 }, { "epoch": 3.0881803162045802, "grad_norm": 1.8449056148529053, "learning_rate": 0.00017823685587363357, "loss": 4.9373, "step": 512150 }, { "epoch": 3.0884818079859144, "grad_norm": 1.9504741430282593, "learning_rate": 0.00017821687044801467, "loss": 4.8285, "step": 512200 }, { "epoch": 3.088783299767248, "grad_norm": 2.168938636779785, "learning_rate": 0.00017819688450310453, "loss": 4.6139, "step": 512250 }, { "epoch": 3.0890847915485824, "grad_norm": 2.519611120223999, "learning_rate": 0.00017817689803927093, "loss": 4.8661, "step": 512300 }, { "epoch": 3.0893862833299166, "grad_norm": 2.4905498027801514, "learning_rate": 0.00017815691105688174, "loss": 5.253, "step": 512350 }, { "epoch": 3.0896877751112504, "grad_norm": 1.9849135875701904, "learning_rate": 0.00017813692355630483, "loss": 4.8539, "step": 512400 }, { "epoch": 3.0899892668925846, "grad_norm": 2.9898440837860107, "learning_rate": 0.00017811693553790794, "loss": 4.4714, "step": 512450 }, { "epoch": 3.0902907586739183, "grad_norm": 2.268759250640869, "learning_rate": 0.000178096947002059, "loss": 4.8157, "step": 512500 }, { "epoch": 3.0905922504552525, "grad_norm": 1.240700602531433, "learning_rate": 0.00017807695794912586, "loss": 4.6237, "step": 512550 }, { "epoch": 3.0908937422365867, "grad_norm": 2.201103448867798, "learning_rate": 0.00017805696837947638, "loss": 5.4827, "step": 512600 }, { "epoch": 3.0911952340179205, "grad_norm": 1.8964613676071167, "learning_rate": 0.00017803697829347844, "loss": 4.9268, "step": 512650 }, { "epoch": 3.0914967257992547, "grad_norm": 2.3410937786102295, "learning_rate": 0.0001780169876914999, "loss": 5.0202, "step": 512700 }, { "epoch": 3.091798217580589, "grad_norm": 1.3655465841293335, "learning_rate": 0.00017799699657390864, "loss": 4.8971, "step": 512750 }, { "epoch": 3.0920997093619227, "grad_norm": 1.6708288192749023, "learning_rate": 0.0001779770049410727, "loss": 4.7742, "step": 512800 }, { "epoch": 3.092401201143257, "grad_norm": 2.1134631633758545, "learning_rate": 0.00017795701279335987, "loss": 5.1197, "step": 512850 }, { "epoch": 3.0927026929245907, "grad_norm": 1.8876984119415283, "learning_rate": 0.00017793702013113808, "loss": 4.8201, "step": 512900 }, { "epoch": 3.093004184705925, "grad_norm": 2.2885947227478027, "learning_rate": 0.00017791702695477536, "loss": 4.9548, "step": 512950 }, { "epoch": 3.093305676487259, "grad_norm": 0.7989384531974792, "learning_rate": 0.00017789703326463957, "loss": 5.1056, "step": 513000 }, { "epoch": 3.093305676487259, "eval_loss": 5.353050231933594, "eval_runtime": 39.1127, "eval_samples_per_second": 13.09, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.42932115035911, "step": 513000 }, { "epoch": 3.093607168268593, "grad_norm": 1.580108642578125, "learning_rate": 0.0001778770390610987, "loss": 4.8619, "step": 513050 }, { "epoch": 3.093908660049927, "grad_norm": 2.1686394214630127, "learning_rate": 0.00017785704434452072, "loss": 5.0283, "step": 513100 }, { "epoch": 3.0942101518312612, "grad_norm": 1.5353010892868042, "learning_rate": 0.00017783704911527361, "loss": 5.1504, "step": 513150 }, { "epoch": 3.094511643612595, "grad_norm": 3.2291083335876465, "learning_rate": 0.00017781705337372532, "loss": 5.4454, "step": 513200 }, { "epoch": 3.094813135393929, "grad_norm": 2.716153860092163, "learning_rate": 0.00017779705712024383, "loss": 5.1741, "step": 513250 }, { "epoch": 3.0951146271752634, "grad_norm": 0.5036844611167908, "learning_rate": 0.00017777706035519721, "loss": 4.621, "step": 513300 }, { "epoch": 3.095416118956597, "grad_norm": 2.560072422027588, "learning_rate": 0.00017775706307895337, "loss": 5.2728, "step": 513350 }, { "epoch": 3.0957176107379314, "grad_norm": 1.7243037223815918, "learning_rate": 0.00017773706529188049, "loss": 4.8279, "step": 513400 }, { "epoch": 3.096019102519265, "grad_norm": 2.164006233215332, "learning_rate": 0.00017771706699434644, "loss": 4.5873, "step": 513450 }, { "epoch": 3.0963205943005994, "grad_norm": 0.41346633434295654, "learning_rate": 0.0001776970681867193, "loss": 5.224, "step": 513500 }, { "epoch": 3.0966220860819336, "grad_norm": 2.968139886856079, "learning_rate": 0.0001776770688693672, "loss": 4.7383, "step": 513550 }, { "epoch": 3.0969235778632673, "grad_norm": 2.6450982093811035, "learning_rate": 0.00017765706904265818, "loss": 4.7973, "step": 513600 }, { "epoch": 3.0972250696446015, "grad_norm": 1.8200421333312988, "learning_rate": 0.0001776370687069602, "loss": 4.7953, "step": 513650 }, { "epoch": 3.0975265614259357, "grad_norm": 1.731520175933838, "learning_rate": 0.00017761706786264145, "loss": 4.91, "step": 513700 }, { "epoch": 3.0978280532072695, "grad_norm": 2.025284767150879, "learning_rate": 0.00017759706651006998, "loss": 5.0762, "step": 513750 }, { "epoch": 3.0981295449886037, "grad_norm": 2.510307788848877, "learning_rate": 0.0001775770646496139, "loss": 4.6048, "step": 513800 }, { "epoch": 3.0984310367699375, "grad_norm": 3.367560386657715, "learning_rate": 0.00017755706228164125, "loss": 5.0081, "step": 513850 }, { "epoch": 3.0987325285512717, "grad_norm": 1.9260954856872559, "learning_rate": 0.0001775370594065202, "loss": 4.8116, "step": 513900 }, { "epoch": 3.099034020332606, "grad_norm": 2.3440890312194824, "learning_rate": 0.00017751705602461886, "loss": 5.0965, "step": 513950 }, { "epoch": 3.0993355121139397, "grad_norm": 1.9014636278152466, "learning_rate": 0.00017749705213630543, "loss": 5.0157, "step": 514000 }, { "epoch": 3.099637003895274, "grad_norm": 3.3523998260498047, "learning_rate": 0.00017747704774194798, "loss": 4.8874, "step": 514050 }, { "epoch": 3.099938495676608, "grad_norm": 2.525036096572876, "learning_rate": 0.00017745704284191462, "loss": 5.1212, "step": 514100 }, { "epoch": 3.100239987457942, "grad_norm": 2.015076160430908, "learning_rate": 0.00017743703743657366, "loss": 4.8006, "step": 514150 }, { "epoch": 3.100541479239276, "grad_norm": 3.544321060180664, "learning_rate": 0.00017741703152629313, "loss": 4.9814, "step": 514200 }, { "epoch": 3.10084297102061, "grad_norm": 0.584723174571991, "learning_rate": 0.00017739702511144123, "loss": 4.8242, "step": 514250 }, { "epoch": 3.101144462801944, "grad_norm": 1.5874136686325073, "learning_rate": 0.00017737701819238626, "loss": 5.1499, "step": 514300 }, { "epoch": 3.101445954583278, "grad_norm": 4.87234354019165, "learning_rate": 0.00017735701076949625, "loss": 5.2792, "step": 514350 }, { "epoch": 3.101747446364612, "grad_norm": 1.7643077373504639, "learning_rate": 0.00017733700284313958, "loss": 5.1451, "step": 514400 }, { "epoch": 3.102048938145946, "grad_norm": 2.2962872982025146, "learning_rate": 0.00017731699441368435, "loss": 5.1173, "step": 514450 }, { "epoch": 3.1023504299272804, "grad_norm": 1.7051228284835815, "learning_rate": 0.00017729698548149875, "loss": 4.9397, "step": 514500 }, { "epoch": 3.102651921708614, "grad_norm": 2.059448480606079, "learning_rate": 0.0001772769760469512, "loss": 4.9856, "step": 514550 }, { "epoch": 3.1029534134899484, "grad_norm": 1.4747058153152466, "learning_rate": 0.00017725696611040976, "loss": 4.7316, "step": 514600 }, { "epoch": 3.103254905271282, "grad_norm": 2.219116449356079, "learning_rate": 0.0001772369556722428, "loss": 4.6989, "step": 514650 }, { "epoch": 3.1035563970526163, "grad_norm": 1.7543554306030273, "learning_rate": 0.0001772169447328185, "loss": 4.6571, "step": 514700 }, { "epoch": 3.1038578888339505, "grad_norm": 1.8722375631332397, "learning_rate": 0.0001771969332925052, "loss": 4.6698, "step": 514750 }, { "epoch": 3.1041593806152843, "grad_norm": 2.3345248699188232, "learning_rate": 0.00017717692135167114, "loss": 4.5514, "step": 514800 }, { "epoch": 3.1044608723966185, "grad_norm": 1.9369279146194458, "learning_rate": 0.00017715690891068468, "loss": 5.0134, "step": 514850 }, { "epoch": 3.1047623641779527, "grad_norm": 1.0052541494369507, "learning_rate": 0.000177136895969914, "loss": 4.7353, "step": 514900 }, { "epoch": 3.1050638559592865, "grad_norm": 1.4098061323165894, "learning_rate": 0.00017711688252972747, "loss": 4.6864, "step": 514950 }, { "epoch": 3.1053653477406207, "grad_norm": 2.223468542098999, "learning_rate": 0.0001770968685904935, "loss": 5.2133, "step": 515000 }, { "epoch": 3.1056668395219544, "grad_norm": 1.6963001489639282, "learning_rate": 0.00017707685415258026, "loss": 4.9666, "step": 515050 }, { "epoch": 3.1059683313032886, "grad_norm": 1.7658112049102783, "learning_rate": 0.00017705683921635614, "loss": 4.9446, "step": 515100 }, { "epoch": 3.106269823084623, "grad_norm": 2.7812917232513428, "learning_rate": 0.00017703682378218956, "loss": 4.8981, "step": 515150 }, { "epoch": 3.1065713148659566, "grad_norm": 2.670189380645752, "learning_rate": 0.0001770168078504488, "loss": 5.3424, "step": 515200 }, { "epoch": 3.106872806647291, "grad_norm": 1.732504963874817, "learning_rate": 0.00017699679142150224, "loss": 4.9461, "step": 515250 }, { "epoch": 3.107174298428625, "grad_norm": 2.4727437496185303, "learning_rate": 0.0001769767744957183, "loss": 5.2404, "step": 515300 }, { "epoch": 3.107475790209959, "grad_norm": 1.7616093158721924, "learning_rate": 0.0001769567570734653, "loss": 5.0328, "step": 515350 }, { "epoch": 3.107777281991293, "grad_norm": 1.9110759496688843, "learning_rate": 0.00017693673915511164, "loss": 5.263, "step": 515400 }, { "epoch": 3.1080787737726268, "grad_norm": 1.3822282552719116, "learning_rate": 0.00017691672074102575, "loss": 4.7996, "step": 515450 }, { "epoch": 3.108380265553961, "grad_norm": 2.065363645553589, "learning_rate": 0.00017689670183157602, "loss": 4.6645, "step": 515500 }, { "epoch": 3.108681757335295, "grad_norm": 2.5441231727600098, "learning_rate": 0.00017687668242713086, "loss": 4.5654, "step": 515550 }, { "epoch": 3.108983249116629, "grad_norm": 2.3513083457946777, "learning_rate": 0.00017685666252805878, "loss": 5.0796, "step": 515600 }, { "epoch": 3.109284740897963, "grad_norm": 1.3198659420013428, "learning_rate": 0.00017683664213472808, "loss": 5.0659, "step": 515650 }, { "epoch": 3.1095862326792973, "grad_norm": 3.119502544403076, "learning_rate": 0.00017681662124750732, "loss": 5.3897, "step": 515700 }, { "epoch": 3.109887724460631, "grad_norm": 1.010546088218689, "learning_rate": 0.0001767965998667649, "loss": 4.2412, "step": 515750 }, { "epoch": 3.1101892162419653, "grad_norm": 2.7344162464141846, "learning_rate": 0.0001767765779928693, "loss": 4.7319, "step": 515800 }, { "epoch": 3.110490708023299, "grad_norm": 0.9531354308128357, "learning_rate": 0.000176756555626189, "loss": 4.836, "step": 515850 }, { "epoch": 3.1107921998046333, "grad_norm": 2.6382834911346436, "learning_rate": 0.0001767365327670925, "loss": 5.2331, "step": 515900 }, { "epoch": 3.1110936915859675, "grad_norm": 1.3095111846923828, "learning_rate": 0.0001767165094159482, "loss": 4.5237, "step": 515950 }, { "epoch": 3.1113951833673013, "grad_norm": 0.9051591753959656, "learning_rate": 0.00017669648557312476, "loss": 5.3289, "step": 516000 }, { "epoch": 3.1113951833673013, "eval_loss": 5.353259086608887, "eval_runtime": 38.9376, "eval_samples_per_second": 13.149, "eval_steps_per_second": 6.575, "eval_tts_loss": 7.451515548932979, "step": 516000 }, { "epoch": 3.1116966751486355, "grad_norm": 2.7358527183532715, "learning_rate": 0.00017667646123899056, "loss": 4.8162, "step": 516050 }, { "epoch": 3.1119981669299697, "grad_norm": 1.9670159816741943, "learning_rate": 0.0001766564364139141, "loss": 4.4477, "step": 516100 }, { "epoch": 3.1122996587113034, "grad_norm": 2.145063877105713, "learning_rate": 0.00017663641109826405, "loss": 5.1454, "step": 516150 }, { "epoch": 3.1126011504926376, "grad_norm": 1.3417434692382812, "learning_rate": 0.00017661638529240886, "loss": 4.6568, "step": 516200 }, { "epoch": 3.1129026422739714, "grad_norm": 0.8493592739105225, "learning_rate": 0.00017659635899671704, "loss": 4.9604, "step": 516250 }, { "epoch": 3.1132041340553056, "grad_norm": 2.689548969268799, "learning_rate": 0.00017657633221155724, "loss": 4.8491, "step": 516300 }, { "epoch": 3.11350562583664, "grad_norm": 2.255568504333496, "learning_rate": 0.0001765563049372979, "loss": 4.8684, "step": 516350 }, { "epoch": 3.1138071176179736, "grad_norm": 1.7031532526016235, "learning_rate": 0.00017653627717430774, "loss": 5.0609, "step": 516400 }, { "epoch": 3.114108609399308, "grad_norm": 1.9316205978393555, "learning_rate": 0.0001765162489229552, "loss": 4.9134, "step": 516450 }, { "epoch": 3.114410101180642, "grad_norm": 1.5691113471984863, "learning_rate": 0.00017649622018360903, "loss": 5.0909, "step": 516500 }, { "epoch": 3.1147115929619758, "grad_norm": 2.0506200790405273, "learning_rate": 0.00017647619095663764, "loss": 4.7964, "step": 516550 }, { "epoch": 3.11501308474331, "grad_norm": 1.7601501941680908, "learning_rate": 0.0001764561612424098, "loss": 4.43, "step": 516600 }, { "epoch": 3.1153145765246437, "grad_norm": 1.8462072610855103, "learning_rate": 0.00017643613104129406, "loss": 5.1609, "step": 516650 }, { "epoch": 3.115616068305978, "grad_norm": 1.6794084310531616, "learning_rate": 0.00017641610035365902, "loss": 4.7994, "step": 516700 }, { "epoch": 3.115917560087312, "grad_norm": 2.086246967315674, "learning_rate": 0.0001763960691798734, "loss": 4.7923, "step": 516750 }, { "epoch": 3.116219051868646, "grad_norm": 0.49532175064086914, "learning_rate": 0.00017637603752030577, "loss": 4.9384, "step": 516800 }, { "epoch": 3.11652054364998, "grad_norm": 2.041888475418091, "learning_rate": 0.0001763560053753248, "loss": 4.9616, "step": 516850 }, { "epoch": 3.1168220354313143, "grad_norm": 1.5965421199798584, "learning_rate": 0.0001763359727452992, "loss": 4.8909, "step": 516900 }, { "epoch": 3.117123527212648, "grad_norm": 1.3434386253356934, "learning_rate": 0.00017631593963059758, "loss": 4.6655, "step": 516950 }, { "epoch": 3.1174250189939823, "grad_norm": 2.8894400596618652, "learning_rate": 0.00017629590603158865, "loss": 4.9205, "step": 517000 }, { "epoch": 3.1177265107753165, "grad_norm": 2.0347301959991455, "learning_rate": 0.00017627587194864107, "loss": 4.7486, "step": 517050 }, { "epoch": 3.1180280025566502, "grad_norm": 2.4365179538726807, "learning_rate": 0.0001762558373821236, "loss": 4.9251, "step": 517100 }, { "epoch": 3.1183294943379845, "grad_norm": 2.2156732082366943, "learning_rate": 0.00017623580233240487, "loss": 4.9523, "step": 517150 }, { "epoch": 3.118630986119318, "grad_norm": 1.3003045320510864, "learning_rate": 0.00017621576679985372, "loss": 5.1849, "step": 517200 }, { "epoch": 3.1189324779006524, "grad_norm": 2.3529250621795654, "learning_rate": 0.0001761957307848387, "loss": 5.271, "step": 517250 }, { "epoch": 3.1192339696819866, "grad_norm": 2.077167510986328, "learning_rate": 0.00017617569428772867, "loss": 4.2032, "step": 517300 }, { "epoch": 3.1195354614633204, "grad_norm": 2.0916030406951904, "learning_rate": 0.00017615565730889237, "loss": 5.0834, "step": 517350 }, { "epoch": 3.1198369532446546, "grad_norm": 1.9828808307647705, "learning_rate": 0.0001761356198486985, "loss": 4.8657, "step": 517400 }, { "epoch": 3.120138445025989, "grad_norm": 1.7573198080062866, "learning_rate": 0.00017611558190751584, "loss": 4.8112, "step": 517450 }, { "epoch": 3.1204399368073226, "grad_norm": 2.3467628955841064, "learning_rate": 0.00017609554348571322, "loss": 4.7044, "step": 517500 }, { "epoch": 3.1207414285886568, "grad_norm": 2.7176899909973145, "learning_rate": 0.00017607550458365926, "loss": 4.5478, "step": 517550 }, { "epoch": 3.1210429203699905, "grad_norm": 2.8608858585357666, "learning_rate": 0.00017605546520172295, "loss": 4.9326, "step": 517600 }, { "epoch": 3.1213444121513247, "grad_norm": 2.496476411819458, "learning_rate": 0.00017603542534027296, "loss": 4.7805, "step": 517650 }, { "epoch": 3.121645903932659, "grad_norm": 2.331559419631958, "learning_rate": 0.00017601538499967807, "loss": 5.17, "step": 517700 }, { "epoch": 3.1219473957139927, "grad_norm": 2.2113139629364014, "learning_rate": 0.00017599534418030716, "loss": 4.7436, "step": 517750 }, { "epoch": 3.122248887495327, "grad_norm": 1.8744150400161743, "learning_rate": 0.00017597530288252907, "loss": 4.9256, "step": 517800 }, { "epoch": 3.122550379276661, "grad_norm": 1.424199104309082, "learning_rate": 0.00017595526110671258, "loss": 4.6358, "step": 517850 }, { "epoch": 3.122851871057995, "grad_norm": 1.4522162675857544, "learning_rate": 0.00017593521885322656, "loss": 5.5544, "step": 517900 }, { "epoch": 3.123153362839329, "grad_norm": 3.353508472442627, "learning_rate": 0.00017591517612243986, "loss": 4.8793, "step": 517950 }, { "epoch": 3.123454854620663, "grad_norm": 1.2440894842147827, "learning_rate": 0.0001758951329147213, "loss": 4.8128, "step": 518000 }, { "epoch": 3.123756346401997, "grad_norm": 4.148002624511719, "learning_rate": 0.00017587508923043977, "loss": 4.5445, "step": 518050 }, { "epoch": 3.1240578381833313, "grad_norm": 1.839614748954773, "learning_rate": 0.0001758550450699642, "loss": 4.9846, "step": 518100 }, { "epoch": 3.124359329964665, "grad_norm": 1.7043923139572144, "learning_rate": 0.00017583500043366333, "loss": 4.7995, "step": 518150 }, { "epoch": 3.1246608217459992, "grad_norm": 1.2317055463790894, "learning_rate": 0.00017581495532190624, "loss": 4.5365, "step": 518200 }, { "epoch": 3.1249623135273334, "grad_norm": 2.3959670066833496, "learning_rate": 0.00017579490973506167, "loss": 4.9631, "step": 518250 }, { "epoch": 3.125263805308667, "grad_norm": 1.8605122566223145, "learning_rate": 0.00017577486367349862, "loss": 5.0266, "step": 518300 }, { "epoch": 3.1255652970900014, "grad_norm": 2.263585090637207, "learning_rate": 0.000175754817137586, "loss": 5.2812, "step": 518350 }, { "epoch": 3.125866788871335, "grad_norm": 0.8504515886306763, "learning_rate": 0.00017573477012769272, "loss": 4.8631, "step": 518400 }, { "epoch": 3.1261682806526694, "grad_norm": 1.8639891147613525, "learning_rate": 0.0001757147226441877, "loss": 5.2211, "step": 518450 }, { "epoch": 3.1264697724340036, "grad_norm": 1.453809142112732, "learning_rate": 0.00017569467468743992, "loss": 4.2755, "step": 518500 }, { "epoch": 3.1267712642153374, "grad_norm": 0.6907035112380981, "learning_rate": 0.00017567462625781835, "loss": 4.911, "step": 518550 }, { "epoch": 3.1270727559966716, "grad_norm": 1.6686211824417114, "learning_rate": 0.0001756545773556919, "loss": 4.822, "step": 518600 }, { "epoch": 3.1273742477780058, "grad_norm": 1.4211196899414062, "learning_rate": 0.00017563452798142955, "loss": 4.9677, "step": 518650 }, { "epoch": 3.1276757395593395, "grad_norm": 2.332660436630249, "learning_rate": 0.00017561447813540032, "loss": 5.1262, "step": 518700 }, { "epoch": 3.1279772313406737, "grad_norm": 1.8660557270050049, "learning_rate": 0.00017559442781797313, "loss": 4.8557, "step": 518750 }, { "epoch": 3.1282787231220075, "grad_norm": 1.8477340936660767, "learning_rate": 0.0001755743770295171, "loss": 4.509, "step": 518800 }, { "epoch": 3.1285802149033417, "grad_norm": 0.9444066882133484, "learning_rate": 0.00017555432577040107, "loss": 5.0211, "step": 518850 }, { "epoch": 3.128881706684676, "grad_norm": 0.5925876498222351, "learning_rate": 0.00017553427404099416, "loss": 4.9585, "step": 518900 }, { "epoch": 3.1291831984660097, "grad_norm": 1.8431309461593628, "learning_rate": 0.00017551422184166544, "loss": 4.9272, "step": 518950 }, { "epoch": 3.129484690247344, "grad_norm": 1.5632435083389282, "learning_rate": 0.00017549416917278382, "loss": 4.8704, "step": 519000 }, { "epoch": 3.129484690247344, "eval_loss": 5.358013153076172, "eval_runtime": 39.0491, "eval_samples_per_second": 13.112, "eval_steps_per_second": 6.556, "eval_tts_loss": 7.520929793529403, "step": 519000 }, { "epoch": 3.129786182028678, "grad_norm": 2.3624050617218018, "learning_rate": 0.0001754741160347184, "loss": 4.9716, "step": 519050 }, { "epoch": 3.130087673810012, "grad_norm": 2.137896776199341, "learning_rate": 0.00017545406242783828, "loss": 5.219, "step": 519100 }, { "epoch": 3.130389165591346, "grad_norm": 1.4332396984100342, "learning_rate": 0.00017543400835251237, "loss": 4.5532, "step": 519150 }, { "epoch": 3.13069065737268, "grad_norm": 2.060500383377075, "learning_rate": 0.00017541395380910992, "loss": 4.5803, "step": 519200 }, { "epoch": 3.130992149154014, "grad_norm": 2.5310356616973877, "learning_rate": 0.0001753938987979999, "loss": 4.9885, "step": 519250 }, { "epoch": 3.1312936409353482, "grad_norm": 2.49712872505188, "learning_rate": 0.00017537384331955136, "loss": 5.0802, "step": 519300 }, { "epoch": 3.131595132716682, "grad_norm": 0.4548467993736267, "learning_rate": 0.0001753537873741335, "loss": 5.127, "step": 519350 }, { "epoch": 3.131896624498016, "grad_norm": 2.7375810146331787, "learning_rate": 0.0001753337309621154, "loss": 5.1733, "step": 519400 }, { "epoch": 3.1321981162793504, "grad_norm": 1.9333226680755615, "learning_rate": 0.00017531367408386603, "loss": 4.7171, "step": 519450 }, { "epoch": 3.132499608060684, "grad_norm": 2.947747230529785, "learning_rate": 0.00017529361673975467, "loss": 4.7846, "step": 519500 }, { "epoch": 3.1328010998420184, "grad_norm": 2.228076219558716, "learning_rate": 0.0001752735589301504, "loss": 5.2368, "step": 519550 }, { "epoch": 3.133102591623352, "grad_norm": 2.0700912475585938, "learning_rate": 0.00017525350065542234, "loss": 5.433, "step": 519600 }, { "epoch": 3.1334040834046863, "grad_norm": 1.745224118232727, "learning_rate": 0.00017523344191593964, "loss": 5.0926, "step": 519650 }, { "epoch": 3.1337055751860206, "grad_norm": 1.0390346050262451, "learning_rate": 0.00017521338271207142, "loss": 4.7457, "step": 519700 }, { "epoch": 3.1340070669673543, "grad_norm": 1.4587252140045166, "learning_rate": 0.00017519332304418686, "loss": 4.8298, "step": 519750 }, { "epoch": 3.1343085587486885, "grad_norm": 1.460036277770996, "learning_rate": 0.0001751732629126552, "loss": 5.1146, "step": 519800 }, { "epoch": 3.1346100505300227, "grad_norm": 2.681349992752075, "learning_rate": 0.00017515320231784558, "loss": 5.1701, "step": 519850 }, { "epoch": 3.1349115423113565, "grad_norm": 1.3321211338043213, "learning_rate": 0.00017513314126012705, "loss": 4.5206, "step": 519900 }, { "epoch": 3.1352130340926907, "grad_norm": 1.5853856801986694, "learning_rate": 0.00017511307973986902, "loss": 4.9773, "step": 519950 }, { "epoch": 3.135514525874025, "grad_norm": 1.800779938697815, "learning_rate": 0.00017509301775744062, "loss": 4.8912, "step": 520000 }, { "epoch": 3.1358160176553587, "grad_norm": 1.9814050197601318, "learning_rate": 0.00017507295531321094, "loss": 4.9862, "step": 520050 }, { "epoch": 3.136117509436693, "grad_norm": 2.5475733280181885, "learning_rate": 0.00017505289240754936, "loss": 4.7687, "step": 520100 }, { "epoch": 3.1364190012180266, "grad_norm": 1.7152879238128662, "learning_rate": 0.00017503282904082502, "loss": 4.9926, "step": 520150 }, { "epoch": 3.136720492999361, "grad_norm": 1.1752402782440186, "learning_rate": 0.0001750127652134072, "loss": 4.5512, "step": 520200 }, { "epoch": 3.137021984780695, "grad_norm": 2.673515796661377, "learning_rate": 0.00017499270092566515, "loss": 5.1759, "step": 520250 }, { "epoch": 3.137323476562029, "grad_norm": 1.509418249130249, "learning_rate": 0.00017497263617796807, "loss": 4.8664, "step": 520300 }, { "epoch": 3.137624968343363, "grad_norm": 2.135324716567993, "learning_rate": 0.00017495257097068528, "loss": 5.0558, "step": 520350 }, { "epoch": 3.137926460124697, "grad_norm": 2.5083107948303223, "learning_rate": 0.00017493250530418607, "loss": 5.1114, "step": 520400 }, { "epoch": 3.138227951906031, "grad_norm": 1.57282292842865, "learning_rate": 0.00017491243917883964, "loss": 4.6799, "step": 520450 }, { "epoch": 3.138529443687365, "grad_norm": 2.5239386558532715, "learning_rate": 0.00017489237259501528, "loss": 4.8547, "step": 520500 }, { "epoch": 3.138830935468699, "grad_norm": 0.8203474879264832, "learning_rate": 0.00017487230555308236, "loss": 4.6685, "step": 520550 }, { "epoch": 3.139132427250033, "grad_norm": 1.1042081117630005, "learning_rate": 0.00017485223805341018, "loss": 4.4537, "step": 520600 }, { "epoch": 3.1394339190313674, "grad_norm": 1.542184829711914, "learning_rate": 0.000174832170096368, "loss": 5.1853, "step": 520650 }, { "epoch": 3.139735410812701, "grad_norm": 2.600243091583252, "learning_rate": 0.00017481210168232515, "loss": 5.0655, "step": 520700 }, { "epoch": 3.1400369025940353, "grad_norm": 1.8459097146987915, "learning_rate": 0.000174792032811651, "loss": 4.9391, "step": 520750 }, { "epoch": 3.1403383943753695, "grad_norm": 1.9928926229476929, "learning_rate": 0.00017477196348471486, "loss": 5.0862, "step": 520800 }, { "epoch": 3.1406398861567033, "grad_norm": 2.14441180229187, "learning_rate": 0.0001747518937018861, "loss": 5.1591, "step": 520850 }, { "epoch": 3.1409413779380375, "grad_norm": 1.356568455696106, "learning_rate": 0.000174731823463534, "loss": 5.0659, "step": 520900 }, { "epoch": 3.1412428697193713, "grad_norm": 2.4902899265289307, "learning_rate": 0.000174711752770028, "loss": 5.0236, "step": 520950 }, { "epoch": 3.1415443615007055, "grad_norm": 2.3672826290130615, "learning_rate": 0.00017469168162173755, "loss": 5.2205, "step": 521000 }, { "epoch": 3.1418458532820397, "grad_norm": 1.7004534006118774, "learning_rate": 0.00017467161001903184, "loss": 5.0335, "step": 521050 }, { "epoch": 3.1421473450633735, "grad_norm": 4.2235426902771, "learning_rate": 0.00017465153796228036, "loss": 4.3721, "step": 521100 }, { "epoch": 3.1424488368447077, "grad_norm": 1.5594037771224976, "learning_rate": 0.00017463146545185253, "loss": 5.2626, "step": 521150 }, { "epoch": 3.1427503286260414, "grad_norm": 3.164057493209839, "learning_rate": 0.00017461139248811774, "loss": 5.0656, "step": 521200 }, { "epoch": 3.1430518204073756, "grad_norm": 2.5546834468841553, "learning_rate": 0.00017459131907144538, "loss": 4.6797, "step": 521250 }, { "epoch": 3.14335331218871, "grad_norm": 2.3868019580841064, "learning_rate": 0.0001745712452022049, "loss": 5.1729, "step": 521300 }, { "epoch": 3.1436548039700436, "grad_norm": 1.9705480337142944, "learning_rate": 0.00017455117088076565, "loss": 4.8074, "step": 521350 }, { "epoch": 3.143956295751378, "grad_norm": 2.797306776046753, "learning_rate": 0.00017453109610749725, "loss": 5.4572, "step": 521400 }, { "epoch": 3.144257787532712, "grad_norm": 2.9343905448913574, "learning_rate": 0.00017451102088276898, "loss": 5.2896, "step": 521450 }, { "epoch": 3.1445592793140458, "grad_norm": 0.6335799694061279, "learning_rate": 0.0001744909452069503, "loss": 5.0727, "step": 521500 }, { "epoch": 3.14486077109538, "grad_norm": 2.243685722351074, "learning_rate": 0.0001744708690804108, "loss": 4.6351, "step": 521550 }, { "epoch": 3.145162262876714, "grad_norm": 3.0054123401641846, "learning_rate": 0.00017445079250351985, "loss": 5.1387, "step": 521600 }, { "epoch": 3.145463754658048, "grad_norm": 2.455436944961548, "learning_rate": 0.00017443071547664692, "loss": 4.6756, "step": 521650 }, { "epoch": 3.145765246439382, "grad_norm": 1.8243180513381958, "learning_rate": 0.00017441063800016155, "loss": 4.9634, "step": 521700 }, { "epoch": 3.146066738220716, "grad_norm": 1.6644606590270996, "learning_rate": 0.00017439056007443327, "loss": 4.9924, "step": 521750 }, { "epoch": 3.14636823000205, "grad_norm": 1.6553164720535278, "learning_rate": 0.00017437048169983152, "loss": 4.7913, "step": 521800 }, { "epoch": 3.1466697217833843, "grad_norm": 1.5057439804077148, "learning_rate": 0.0001743504028767258, "loss": 5.0252, "step": 521850 }, { "epoch": 3.146971213564718, "grad_norm": 2.199765682220459, "learning_rate": 0.00017433032360548572, "loss": 4.9816, "step": 521900 }, { "epoch": 3.1472727053460523, "grad_norm": 2.910576105117798, "learning_rate": 0.00017431024388648068, "loss": 4.8737, "step": 521950 }, { "epoch": 3.1475741971273865, "grad_norm": 2.310594081878662, "learning_rate": 0.00017429016372008037, "loss": 5.0592, "step": 522000 }, { "epoch": 3.1475741971273865, "eval_loss": 5.355422496795654, "eval_runtime": 39.217, "eval_samples_per_second": 13.056, "eval_steps_per_second": 6.528, "eval_tts_loss": 7.512918176698532, "step": 522000 }, { "epoch": 3.1478756889087203, "grad_norm": 1.536448359489441, "learning_rate": 0.00017427008310665423, "loss": 5.002, "step": 522050 }, { "epoch": 3.1481771806900545, "grad_norm": 1.5884127616882324, "learning_rate": 0.00017425000204657182, "loss": 5.0401, "step": 522100 }, { "epoch": 3.1484786724713882, "grad_norm": 1.3906800746917725, "learning_rate": 0.00017422992054020274, "loss": 4.505, "step": 522150 }, { "epoch": 3.1487801642527224, "grad_norm": 1.0774608850479126, "learning_rate": 0.00017420983858791662, "loss": 4.5359, "step": 522200 }, { "epoch": 3.1490816560340567, "grad_norm": 1.6501001119613647, "learning_rate": 0.00017418975619008285, "loss": 4.7138, "step": 522250 }, { "epoch": 3.1493831478153904, "grad_norm": 3.0316195487976074, "learning_rate": 0.00017416967334707122, "loss": 4.3485, "step": 522300 }, { "epoch": 3.1496846395967246, "grad_norm": 2.114272117614746, "learning_rate": 0.0001741495900592512, "loss": 4.775, "step": 522350 }, { "epoch": 3.149986131378059, "grad_norm": 2.076129674911499, "learning_rate": 0.00017412950632699249, "loss": 4.889, "step": 522400 }, { "epoch": 3.1502876231593926, "grad_norm": 1.5950244665145874, "learning_rate": 0.00017410942215066461, "loss": 4.5144, "step": 522450 }, { "epoch": 3.150589114940727, "grad_norm": 1.715949535369873, "learning_rate": 0.00017408933753063722, "loss": 4.7727, "step": 522500 }, { "epoch": 3.1508906067220606, "grad_norm": 2.1365299224853516, "learning_rate": 0.00017406925246727997, "loss": 5.0376, "step": 522550 }, { "epoch": 3.1511920985033948, "grad_norm": 1.2876569032669067, "learning_rate": 0.0001740491669609625, "loss": 4.8506, "step": 522600 }, { "epoch": 3.151493590284729, "grad_norm": 0.9202232956886292, "learning_rate": 0.0001740290810120544, "loss": 4.724, "step": 522650 }, { "epoch": 3.1517950820660627, "grad_norm": 1.590993881225586, "learning_rate": 0.00017400899462092535, "loss": 5.0152, "step": 522700 }, { "epoch": 3.152096573847397, "grad_norm": 1.5765889883041382, "learning_rate": 0.000173988907787945, "loss": 4.8826, "step": 522750 }, { "epoch": 3.152398065628731, "grad_norm": 1.6341218948364258, "learning_rate": 0.0001739688205134831, "loss": 5.2188, "step": 522800 }, { "epoch": 3.152699557410065, "grad_norm": 1.6424039602279663, "learning_rate": 0.0001739487327979092, "loss": 5.3336, "step": 522850 }, { "epoch": 3.153001049191399, "grad_norm": 1.69691002368927, "learning_rate": 0.00017392864464159307, "loss": 4.8631, "step": 522900 }, { "epoch": 3.153302540972733, "grad_norm": 3.27485990524292, "learning_rate": 0.0001739085560449044, "loss": 4.7879, "step": 522950 }, { "epoch": 3.153604032754067, "grad_norm": 1.4994739294052124, "learning_rate": 0.00017388846700821286, "loss": 5.4401, "step": 523000 }, { "epoch": 3.1539055245354013, "grad_norm": 2.899679183959961, "learning_rate": 0.0001738683775318882, "loss": 4.7049, "step": 523050 }, { "epoch": 3.154207016316735, "grad_norm": 1.6599615812301636, "learning_rate": 0.00017384828761630005, "loss": 4.7107, "step": 523100 }, { "epoch": 3.1545085080980693, "grad_norm": 0.7529410719871521, "learning_rate": 0.00017382819726181827, "loss": 4.7923, "step": 523150 }, { "epoch": 3.1548099998794035, "grad_norm": 1.5250452756881714, "learning_rate": 0.0001738081064688125, "loss": 4.642, "step": 523200 }, { "epoch": 3.1551114916607372, "grad_norm": 1.7437465190887451, "learning_rate": 0.0001737880152376525, "loss": 4.6269, "step": 523250 }, { "epoch": 3.1554129834420714, "grad_norm": 2.032694101333618, "learning_rate": 0.00017376792356870802, "loss": 5.2981, "step": 523300 }, { "epoch": 3.155714475223405, "grad_norm": 3.5283401012420654, "learning_rate": 0.00017374783146234884, "loss": 4.9894, "step": 523350 }, { "epoch": 3.1560159670047394, "grad_norm": 2.228360414505005, "learning_rate": 0.0001737277389189447, "loss": 4.8102, "step": 523400 }, { "epoch": 3.1563174587860736, "grad_norm": 1.5037055015563965, "learning_rate": 0.0001737076459388654, "loss": 4.8989, "step": 523450 }, { "epoch": 3.1566189505674074, "grad_norm": 2.375148296356201, "learning_rate": 0.0001736875525224807, "loss": 4.8584, "step": 523500 }, { "epoch": 3.1569204423487416, "grad_norm": 2.2732720375061035, "learning_rate": 0.0001736674586701604, "loss": 4.8457, "step": 523550 }, { "epoch": 3.157221934130076, "grad_norm": 2.138634443283081, "learning_rate": 0.0001736473643822743, "loss": 4.8389, "step": 523600 }, { "epoch": 3.1575234259114096, "grad_norm": 1.5124648809432983, "learning_rate": 0.00017362726965919223, "loss": 5.0916, "step": 523650 }, { "epoch": 3.1578249176927438, "grad_norm": 1.8223772048950195, "learning_rate": 0.0001736071745012839, "loss": 5.0772, "step": 523700 }, { "epoch": 3.158126409474078, "grad_norm": 1.9677399396896362, "learning_rate": 0.0001735870789089193, "loss": 4.852, "step": 523750 }, { "epoch": 3.1584279012554117, "grad_norm": 1.9097201824188232, "learning_rate": 0.00017356698288246818, "loss": 5.1849, "step": 523800 }, { "epoch": 3.158729393036746, "grad_norm": 1.2577407360076904, "learning_rate": 0.0001735468864223003, "loss": 4.813, "step": 523850 }, { "epoch": 3.1590308848180797, "grad_norm": 2.353637456893921, "learning_rate": 0.0001735267895287856, "loss": 5.0503, "step": 523900 }, { "epoch": 3.159332376599414, "grad_norm": 2.0634617805480957, "learning_rate": 0.00017350669220229394, "loss": 5.0011, "step": 523950 }, { "epoch": 3.159633868380748, "grad_norm": 2.296448230743408, "learning_rate": 0.00017348659444319515, "loss": 5.1341, "step": 524000 }, { "epoch": 3.159935360162082, "grad_norm": 1.441435694694519, "learning_rate": 0.0001734664962518591, "loss": 4.7906, "step": 524050 }, { "epoch": 3.160236851943416, "grad_norm": 2.339468002319336, "learning_rate": 0.00017344639762865565, "loss": 5.0597, "step": 524100 }, { "epoch": 3.16053834372475, "grad_norm": 2.2230172157287598, "learning_rate": 0.00017342629857395475, "loss": 4.9692, "step": 524150 }, { "epoch": 3.160839835506084, "grad_norm": 2.097663164138794, "learning_rate": 0.00017340619908812622, "loss": 4.7106, "step": 524200 }, { "epoch": 3.1611413272874183, "grad_norm": 1.7026686668395996, "learning_rate": 0.00017338609917154002, "loss": 4.88, "step": 524250 }, { "epoch": 3.161442819068752, "grad_norm": 2.547255039215088, "learning_rate": 0.00017336599882456603, "loss": 5.1188, "step": 524300 }, { "epoch": 3.1617443108500862, "grad_norm": 5.164364814758301, "learning_rate": 0.00017334589804757417, "loss": 5.2246, "step": 524350 }, { "epoch": 3.1620458026314204, "grad_norm": 2.691413640975952, "learning_rate": 0.0001733257968409344, "loss": 4.8137, "step": 524400 }, { "epoch": 3.162347294412754, "grad_norm": 1.0059916973114014, "learning_rate": 0.00017330569520501664, "loss": 4.9615, "step": 524450 }, { "epoch": 3.1626487861940884, "grad_norm": 1.383103370666504, "learning_rate": 0.0001732855931401908, "loss": 4.951, "step": 524500 }, { "epoch": 3.1629502779754226, "grad_norm": 2.714597225189209, "learning_rate": 0.00017326549064682683, "loss": 4.8688, "step": 524550 }, { "epoch": 3.1632517697567564, "grad_norm": 2.157595634460449, "learning_rate": 0.00017324538772529473, "loss": 5.0362, "step": 524600 }, { "epoch": 3.1635532615380906, "grad_norm": 0.8230142593383789, "learning_rate": 0.0001732252843759644, "loss": 5.0961, "step": 524650 }, { "epoch": 3.1638547533194243, "grad_norm": 1.9256168603897095, "learning_rate": 0.00017320518059920584, "loss": 4.9662, "step": 524700 }, { "epoch": 3.1641562451007585, "grad_norm": 0.646404504776001, "learning_rate": 0.0001731850763953891, "loss": 4.7298, "step": 524750 }, { "epoch": 3.1644577368820928, "grad_norm": 1.763959288597107, "learning_rate": 0.00017316497176488415, "loss": 4.516, "step": 524800 }, { "epoch": 3.1647592286634265, "grad_norm": 2.325831174850464, "learning_rate": 0.0001731448667080609, "loss": 4.9351, "step": 524850 }, { "epoch": 3.1650607204447607, "grad_norm": 1.8277740478515625, "learning_rate": 0.0001731247612252894, "loss": 4.3775, "step": 524900 }, { "epoch": 3.1653622122260945, "grad_norm": 1.20475172996521, "learning_rate": 0.0001731046553169397, "loss": 4.8426, "step": 524950 }, { "epoch": 3.1656637040074287, "grad_norm": 2.759150743484497, "learning_rate": 0.00017308454898338178, "loss": 5.1689, "step": 525000 }, { "epoch": 3.1656637040074287, "eval_loss": 5.364665508270264, "eval_runtime": 38.9848, "eval_samples_per_second": 13.133, "eval_steps_per_second": 6.567, "eval_tts_loss": 7.40700595307358, "step": 525000 }, { "epoch": 3.165965195788763, "grad_norm": 1.9095969200134277, "learning_rate": 0.0001730644422249857, "loss": 4.684, "step": 525050 }, { "epoch": 3.1662666875700967, "grad_norm": 2.4388179779052734, "learning_rate": 0.00017304433504212145, "loss": 4.9094, "step": 525100 }, { "epoch": 3.166568179351431, "grad_norm": 1.6381971836090088, "learning_rate": 0.00017302422743515911, "loss": 5.0989, "step": 525150 }, { "epoch": 3.166869671132765, "grad_norm": 1.5541378259658813, "learning_rate": 0.00017300411940446872, "loss": 4.7008, "step": 525200 }, { "epoch": 3.167171162914099, "grad_norm": 1.6383110284805298, "learning_rate": 0.0001729840109504203, "loss": 4.6145, "step": 525250 }, { "epoch": 3.167472654695433, "grad_norm": 2.659895896911621, "learning_rate": 0.000172963902073384, "loss": 4.9591, "step": 525300 }, { "epoch": 3.1677741464767672, "grad_norm": 2.1260087490081787, "learning_rate": 0.00017294379277372983, "loss": 4.7742, "step": 525350 }, { "epoch": 3.168075638258101, "grad_norm": 2.215527057647705, "learning_rate": 0.00017292368305182795, "loss": 4.8784, "step": 525400 }, { "epoch": 3.168377130039435, "grad_norm": 1.750215768814087, "learning_rate": 0.00017290357290804836, "loss": 5.2791, "step": 525450 }, { "epoch": 3.168678621820769, "grad_norm": 7.8331122398376465, "learning_rate": 0.00017288346234276117, "loss": 4.9079, "step": 525500 }, { "epoch": 3.168980113602103, "grad_norm": 2.1449742317199707, "learning_rate": 0.00017286335135633654, "loss": 4.84, "step": 525550 }, { "epoch": 3.1692816053834374, "grad_norm": 1.7260701656341553, "learning_rate": 0.00017284323994914456, "loss": 4.9885, "step": 525600 }, { "epoch": 3.169583097164771, "grad_norm": 3.198624849319458, "learning_rate": 0.00017282312812155537, "loss": 4.7966, "step": 525650 }, { "epoch": 3.1698845889461054, "grad_norm": 2.4738986492156982, "learning_rate": 0.00017280301587393898, "loss": 4.9516, "step": 525700 }, { "epoch": 3.1701860807274396, "grad_norm": 1.7566393613815308, "learning_rate": 0.00017278290320666575, "loss": 5.2465, "step": 525750 }, { "epoch": 3.1704875725087733, "grad_norm": 1.8342303037643433, "learning_rate": 0.00017276279012010562, "loss": 4.9099, "step": 525800 }, { "epoch": 3.1707890642901075, "grad_norm": 2.696669101715088, "learning_rate": 0.00017274267661462886, "loss": 4.6008, "step": 525850 }, { "epoch": 3.1710905560714413, "grad_norm": 1.1698880195617676, "learning_rate": 0.00017272256269060558, "loss": 4.7173, "step": 525900 }, { "epoch": 3.1713920478527755, "grad_norm": 2.8051207065582275, "learning_rate": 0.00017270244834840596, "loss": 5.0302, "step": 525950 }, { "epoch": 3.1716935396341097, "grad_norm": 2.283381700515747, "learning_rate": 0.00017268233358840016, "loss": 4.7011, "step": 526000 }, { "epoch": 3.1719950314154435, "grad_norm": 1.8411930799484253, "learning_rate": 0.00017266221841095842, "loss": 4.98, "step": 526050 }, { "epoch": 3.1722965231967777, "grad_norm": 1.629899501800537, "learning_rate": 0.00017264210281645086, "loss": 4.9182, "step": 526100 }, { "epoch": 3.172598014978112, "grad_norm": 0.3326248228549957, "learning_rate": 0.0001726219868052477, "loss": 4.8793, "step": 526150 }, { "epoch": 3.1728995067594457, "grad_norm": 1.6261347532272339, "learning_rate": 0.0001726018703777192, "loss": 5.1159, "step": 526200 }, { "epoch": 3.17320099854078, "grad_norm": 2.190878391265869, "learning_rate": 0.0001725817535342355, "loss": 4.8566, "step": 526250 }, { "epoch": 3.1735024903221136, "grad_norm": 1.8026677370071411, "learning_rate": 0.00017256163627516684, "loss": 4.6844, "step": 526300 }, { "epoch": 3.173803982103448, "grad_norm": 1.033980369567871, "learning_rate": 0.00017254151860088344, "loss": 4.749, "step": 526350 }, { "epoch": 3.174105473884782, "grad_norm": 3.252394199371338, "learning_rate": 0.00017252140051175565, "loss": 4.8349, "step": 526400 }, { "epoch": 3.174406965666116, "grad_norm": 1.8145920038223267, "learning_rate": 0.00017250128200815354, "loss": 4.8055, "step": 526450 }, { "epoch": 3.17470845744745, "grad_norm": 1.5671838521957397, "learning_rate": 0.00017248116309044747, "loss": 5.0018, "step": 526500 }, { "epoch": 3.175009949228784, "grad_norm": 2.0641863346099854, "learning_rate": 0.00017246104375900768, "loss": 5.1932, "step": 526550 }, { "epoch": 3.175311441010118, "grad_norm": 2.6062629222869873, "learning_rate": 0.00017244092401420442, "loss": 4.5216, "step": 526600 }, { "epoch": 3.175612932791452, "grad_norm": 2.2905189990997314, "learning_rate": 0.000172420803856408, "loss": 5.4555, "step": 526650 }, { "epoch": 3.175914424572786, "grad_norm": 2.661569118499756, "learning_rate": 0.00017240068328598868, "loss": 5.0592, "step": 526700 }, { "epoch": 3.17621591635412, "grad_norm": 1.6544580459594727, "learning_rate": 0.00017238056230331676, "loss": 4.6157, "step": 526750 }, { "epoch": 3.1765174081354544, "grad_norm": 2.00821852684021, "learning_rate": 0.0001723604409087625, "loss": 5.0433, "step": 526800 }, { "epoch": 3.176818899916788, "grad_norm": 1.6829044818878174, "learning_rate": 0.00017234031910269625, "loss": 5.4679, "step": 526850 }, { "epoch": 3.1771203916981223, "grad_norm": 2.797772169113159, "learning_rate": 0.00017232019688548828, "loss": 4.779, "step": 526900 }, { "epoch": 3.1774218834794565, "grad_norm": 1.5916993618011475, "learning_rate": 0.00017230007425750896, "loss": 4.9174, "step": 526950 }, { "epoch": 3.1777233752607903, "grad_norm": 2.243802547454834, "learning_rate": 0.00017227995121912865, "loss": 4.484, "step": 527000 }, { "epoch": 3.1780248670421245, "grad_norm": 2.0089774131774902, "learning_rate": 0.00017225982777071755, "loss": 5.0618, "step": 527050 }, { "epoch": 3.1783263588234583, "grad_norm": 2.0039591789245605, "learning_rate": 0.00017223970391264614, "loss": 4.4931, "step": 527100 }, { "epoch": 3.1786278506047925, "grad_norm": 1.615010142326355, "learning_rate": 0.0001722195796452847, "loss": 4.4363, "step": 527150 }, { "epoch": 3.1789293423861267, "grad_norm": 1.9165319204330444, "learning_rate": 0.0001721994549690036, "loss": 5.0398, "step": 527200 }, { "epoch": 3.1792308341674604, "grad_norm": 1.7991127967834473, "learning_rate": 0.0001721793298841732, "loss": 4.6994, "step": 527250 }, { "epoch": 3.1795323259487946, "grad_norm": 1.7658263444900513, "learning_rate": 0.00017215920439116392, "loss": 5.1991, "step": 527300 }, { "epoch": 3.179833817730129, "grad_norm": 1.3348642587661743, "learning_rate": 0.00017213907849034609, "loss": 4.9143, "step": 527350 }, { "epoch": 3.1801353095114626, "grad_norm": 1.9552010297775269, "learning_rate": 0.00017211895218209011, "loss": 4.7341, "step": 527400 }, { "epoch": 3.180436801292797, "grad_norm": 1.633687973022461, "learning_rate": 0.00017209882546676634, "loss": 5.0751, "step": 527450 }, { "epoch": 3.180738293074131, "grad_norm": 2.3393585681915283, "learning_rate": 0.0001720786983447453, "loss": 4.8348, "step": 527500 }, { "epoch": 3.181039784855465, "grad_norm": 1.8375122547149658, "learning_rate": 0.0001720585708163973, "loss": 4.8693, "step": 527550 }, { "epoch": 3.181341276636799, "grad_norm": 2.647280216217041, "learning_rate": 0.00017203844288209277, "loss": 5.35, "step": 527600 }, { "epoch": 3.1816427684181328, "grad_norm": 2.5761470794677734, "learning_rate": 0.00017201831454220217, "loss": 4.9448, "step": 527650 }, { "epoch": 3.181944260199467, "grad_norm": 1.4469029903411865, "learning_rate": 0.0001719981857970959, "loss": 4.487, "step": 527700 }, { "epoch": 3.182245751980801, "grad_norm": 1.5541242361068726, "learning_rate": 0.0001719780566471444, "loss": 5.0475, "step": 527750 }, { "epoch": 3.182547243762135, "grad_norm": 1.6818562746047974, "learning_rate": 0.00017195792709271814, "loss": 4.8979, "step": 527800 }, { "epoch": 3.182848735543469, "grad_norm": 1.94388747215271, "learning_rate": 0.00017193779713418756, "loss": 5.2825, "step": 527850 }, { "epoch": 3.183150227324803, "grad_norm": 1.869884729385376, "learning_rate": 0.00017191766677192313, "loss": 4.8363, "step": 527900 }, { "epoch": 3.183451719106137, "grad_norm": 1.7867002487182617, "learning_rate": 0.00017189753600629538, "loss": 5.0216, "step": 527950 }, { "epoch": 3.1837532108874713, "grad_norm": 2.3943300247192383, "learning_rate": 0.00017187740483767466, "loss": 4.3367, "step": 528000 }, { "epoch": 3.1837532108874713, "eval_loss": 5.355640411376953, "eval_runtime": 39.1824, "eval_samples_per_second": 13.067, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.402308767963908, "step": 528000 }, { "epoch": 3.184054702668805, "grad_norm": 1.5384130477905273, "learning_rate": 0.00017185727326643151, "loss": 4.6527, "step": 528050 }, { "epoch": 3.1843561944501393, "grad_norm": 2.0968387126922607, "learning_rate": 0.00017183714129293649, "loss": 4.8407, "step": 528100 }, { "epoch": 3.1846576862314735, "grad_norm": 2.666015625, "learning_rate": 0.00017181700891756002, "loss": 5.0976, "step": 528150 }, { "epoch": 3.1849591780128073, "grad_norm": 1.9787880182266235, "learning_rate": 0.00017179687614067267, "loss": 4.9696, "step": 528200 }, { "epoch": 3.1852606697941415, "grad_norm": 1.6457319259643555, "learning_rate": 0.0001717767429626449, "loss": 5.0012, "step": 528250 }, { "epoch": 3.1855621615754757, "grad_norm": 2.04343843460083, "learning_rate": 0.00017175660938384724, "loss": 4.9722, "step": 528300 }, { "epoch": 3.1858636533568094, "grad_norm": 2.0188403129577637, "learning_rate": 0.00017173647540465024, "loss": 4.8313, "step": 528350 }, { "epoch": 3.1861651451381436, "grad_norm": 1.5889134407043457, "learning_rate": 0.00017171634102542443, "loss": 4.8644, "step": 528400 }, { "epoch": 3.1864666369194774, "grad_norm": 3.555753231048584, "learning_rate": 0.00017169620624654037, "loss": 5.1112, "step": 528450 }, { "epoch": 3.1867681287008116, "grad_norm": 1.6414412260055542, "learning_rate": 0.00017167607106836856, "loss": 4.661, "step": 528500 }, { "epoch": 3.187069620482146, "grad_norm": 1.4608319997787476, "learning_rate": 0.00017165593549127965, "loss": 4.8129, "step": 528550 }, { "epoch": 3.1873711122634796, "grad_norm": 2.647141933441162, "learning_rate": 0.00017163579951564413, "loss": 4.8497, "step": 528600 }, { "epoch": 3.187672604044814, "grad_norm": 2.190570592880249, "learning_rate": 0.00017161566314183256, "loss": 5.1527, "step": 528650 }, { "epoch": 3.1879740958261475, "grad_norm": 1.327775478363037, "learning_rate": 0.0001715955263702156, "loss": 5.149, "step": 528700 }, { "epoch": 3.1882755876074818, "grad_norm": 1.3468149900436401, "learning_rate": 0.00017157538920116374, "loss": 5.2808, "step": 528750 }, { "epoch": 3.188577079388816, "grad_norm": 1.7643723487854004, "learning_rate": 0.00017155525163504767, "loss": 4.8024, "step": 528800 }, { "epoch": 3.1888785711701497, "grad_norm": 3.3771986961364746, "learning_rate": 0.00017153511367223797, "loss": 4.5126, "step": 528850 }, { "epoch": 3.189180062951484, "grad_norm": 2.5909252166748047, "learning_rate": 0.0001715149753131052, "loss": 5.0238, "step": 528900 }, { "epoch": 3.189481554732818, "grad_norm": 1.7497591972351074, "learning_rate": 0.00017149483655802008, "loss": 5.0537, "step": 528950 }, { "epoch": 3.189783046514152, "grad_norm": 1.0270529985427856, "learning_rate": 0.0001714746974073531, "loss": 4.3143, "step": 529000 }, { "epoch": 3.190084538295486, "grad_norm": 2.3953468799591064, "learning_rate": 0.00017145455786147495, "loss": 4.8248, "step": 529050 }, { "epoch": 3.1903860300768203, "grad_norm": 1.922655463218689, "learning_rate": 0.0001714344179207563, "loss": 4.807, "step": 529100 }, { "epoch": 3.190687521858154, "grad_norm": 1.688612699508667, "learning_rate": 0.0001714142775855678, "loss": 4.9107, "step": 529150 }, { "epoch": 3.1909890136394883, "grad_norm": 1.8458821773529053, "learning_rate": 0.00017139413685628009, "loss": 4.8587, "step": 529200 }, { "epoch": 3.191290505420822, "grad_norm": 0.9745505452156067, "learning_rate": 0.0001713739957332638, "loss": 4.7609, "step": 529250 }, { "epoch": 3.1915919972021563, "grad_norm": 1.864030361175537, "learning_rate": 0.00017135385421688962, "loss": 4.93, "step": 529300 }, { "epoch": 3.1918934889834905, "grad_norm": 2.2964882850646973, "learning_rate": 0.00017133371230752822, "loss": 4.793, "step": 529350 }, { "epoch": 3.192194980764824, "grad_norm": 2.5501222610473633, "learning_rate": 0.0001713135700055503, "loss": 4.7803, "step": 529400 }, { "epoch": 3.1924964725461584, "grad_norm": 1.1241086721420288, "learning_rate": 0.00017129342731132656, "loss": 4.5953, "step": 529450 }, { "epoch": 3.1927979643274926, "grad_norm": 2.1479387283325195, "learning_rate": 0.0001712732842252276, "loss": 5.0638, "step": 529500 }, { "epoch": 3.1930994561088264, "grad_norm": 0.6479436755180359, "learning_rate": 0.0001712531407476243, "loss": 5.1024, "step": 529550 }, { "epoch": 3.1934009478901606, "grad_norm": 2.4116289615631104, "learning_rate": 0.00017123299687888722, "loss": 5.2983, "step": 529600 }, { "epoch": 3.1937024396714944, "grad_norm": 1.8013395071029663, "learning_rate": 0.00017121285261938716, "loss": 5.115, "step": 529650 }, { "epoch": 3.1940039314528286, "grad_norm": 2.2521438598632812, "learning_rate": 0.0001711927079694948, "loss": 4.9365, "step": 529700 }, { "epoch": 3.1943054232341628, "grad_norm": 1.6450895071029663, "learning_rate": 0.00017117256292958089, "loss": 5.0653, "step": 529750 }, { "epoch": 3.1946069150154965, "grad_norm": 1.739038109779358, "learning_rate": 0.00017115241750001618, "loss": 4.8918, "step": 529800 }, { "epoch": 3.1949084067968307, "grad_norm": 1.6258697509765625, "learning_rate": 0.00017113227168117143, "loss": 4.8607, "step": 529850 }, { "epoch": 3.195209898578165, "grad_norm": 3.144932270050049, "learning_rate": 0.00017111212547341737, "loss": 5.0653, "step": 529900 }, { "epoch": 3.1955113903594987, "grad_norm": 1.555379033088684, "learning_rate": 0.00017109197887712476, "loss": 4.361, "step": 529950 }, { "epoch": 3.195812882140833, "grad_norm": 2.3867533206939697, "learning_rate": 0.0001710718318926644, "loss": 4.767, "step": 530000 }, { "epoch": 3.1961143739221667, "grad_norm": 1.5021907091140747, "learning_rate": 0.00017105168452040702, "loss": 4.7733, "step": 530050 }, { "epoch": 3.196415865703501, "grad_norm": 1.7825911045074463, "learning_rate": 0.00017103153676072342, "loss": 4.9896, "step": 530100 }, { "epoch": 3.196717357484835, "grad_norm": 2.0097200870513916, "learning_rate": 0.00017101138861398444, "loss": 4.8102, "step": 530150 }, { "epoch": 3.197018849266169, "grad_norm": 1.8799618482589722, "learning_rate": 0.00017099124008056082, "loss": 5.2054, "step": 530200 }, { "epoch": 3.197320341047503, "grad_norm": 4.168281555175781, "learning_rate": 0.00017097109116082334, "loss": 4.9289, "step": 530250 }, { "epoch": 3.1976218328288373, "grad_norm": 2.268188714981079, "learning_rate": 0.00017095094185514288, "loss": 4.7145, "step": 530300 }, { "epoch": 3.197923324610171, "grad_norm": 1.949014663696289, "learning_rate": 0.00017093079216389022, "loss": 4.9103, "step": 530350 }, { "epoch": 3.1982248163915052, "grad_norm": 2.471611976623535, "learning_rate": 0.00017091064208743623, "loss": 5.2107, "step": 530400 }, { "epoch": 3.198526308172839, "grad_norm": 1.659845232963562, "learning_rate": 0.00017089049162615175, "loss": 4.9919, "step": 530450 }, { "epoch": 3.198827799954173, "grad_norm": 1.0731585025787354, "learning_rate": 0.00017087034078040752, "loss": 4.3537, "step": 530500 }, { "epoch": 3.1991292917355074, "grad_norm": 1.5199915170669556, "learning_rate": 0.00017085018955057446, "loss": 4.9903, "step": 530550 }, { "epoch": 3.199430783516841, "grad_norm": 1.8354980945587158, "learning_rate": 0.00017083003793702344, "loss": 4.6405, "step": 530600 }, { "epoch": 3.1997322752981754, "grad_norm": 2.7291882038116455, "learning_rate": 0.00017080988594012528, "loss": 4.7267, "step": 530650 }, { "epoch": 3.2000337670795096, "grad_norm": 2.4642984867095947, "learning_rate": 0.00017078973356025087, "loss": 5.1062, "step": 530700 }, { "epoch": 3.2003352588608434, "grad_norm": 2.7716145515441895, "learning_rate": 0.0001707695807977711, "loss": 4.7674, "step": 530750 }, { "epoch": 3.2006367506421776, "grad_norm": 2.3391029834747314, "learning_rate": 0.00017074942765305678, "loss": 4.5716, "step": 530800 }, { "epoch": 3.2009382424235113, "grad_norm": 2.3203296661376953, "learning_rate": 0.00017072927412647892, "loss": 4.8879, "step": 530850 }, { "epoch": 3.2012397342048455, "grad_norm": 2.6272952556610107, "learning_rate": 0.00017070912021840832, "loss": 4.9329, "step": 530900 }, { "epoch": 3.2015412259861797, "grad_norm": 1.934136152267456, "learning_rate": 0.00017068896592921593, "loss": 5.1836, "step": 530950 }, { "epoch": 3.2018427177675135, "grad_norm": 2.0259759426116943, "learning_rate": 0.00017066881125927266, "loss": 4.978, "step": 531000 }, { "epoch": 3.2018427177675135, "eval_loss": 5.358515739440918, "eval_runtime": 39.2201, "eval_samples_per_second": 13.055, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.447715706172776, "step": 531000 }, { "epoch": 3.2021442095488477, "grad_norm": 2.2636373043060303, "learning_rate": 0.00017064865620894937, "loss": 4.7307, "step": 531050 }, { "epoch": 3.202445701330182, "grad_norm": 0.8504832983016968, "learning_rate": 0.00017062850077861704, "loss": 4.9058, "step": 531100 }, { "epoch": 3.2027471931115157, "grad_norm": 1.8951836824417114, "learning_rate": 0.0001706083449686466, "loss": 4.832, "step": 531150 }, { "epoch": 3.20304868489285, "grad_norm": 2.0157227516174316, "learning_rate": 0.00017058818877940902, "loss": 5.144, "step": 531200 }, { "epoch": 3.203350176674184, "grad_norm": 1.5267088413238525, "learning_rate": 0.00017056803221127516, "loss": 4.7511, "step": 531250 }, { "epoch": 3.203651668455518, "grad_norm": 2.3891165256500244, "learning_rate": 0.00017054787526461603, "loss": 4.8168, "step": 531300 }, { "epoch": 3.203953160236852, "grad_norm": 1.855264663696289, "learning_rate": 0.00017052771793980262, "loss": 4.4027, "step": 531350 }, { "epoch": 3.204254652018186, "grad_norm": 2.4100987911224365, "learning_rate": 0.00017050756023720577, "loss": 5.0109, "step": 531400 }, { "epoch": 3.20455614379952, "grad_norm": 2.634173631668091, "learning_rate": 0.00017048740215719658, "loss": 4.912, "step": 531450 }, { "epoch": 3.2048576355808542, "grad_norm": 3.566349506378174, "learning_rate": 0.000170467243700146, "loss": 4.6772, "step": 531500 }, { "epoch": 3.205159127362188, "grad_norm": 3.4282233715057373, "learning_rate": 0.000170447084866425, "loss": 4.6201, "step": 531550 }, { "epoch": 3.205460619143522, "grad_norm": 2.090195417404175, "learning_rate": 0.0001704269256564046, "loss": 4.9177, "step": 531600 }, { "epoch": 3.205762110924856, "grad_norm": 2.5365347862243652, "learning_rate": 0.00017040676607045578, "loss": 4.9887, "step": 531650 }, { "epoch": 3.20606360270619, "grad_norm": 1.745421290397644, "learning_rate": 0.0001703866061089495, "loss": 4.8573, "step": 531700 }, { "epoch": 3.2063650944875244, "grad_norm": 1.7069387435913086, "learning_rate": 0.00017036644577225692, "loss": 5.0235, "step": 531750 }, { "epoch": 3.206666586268858, "grad_norm": 2.784000873565674, "learning_rate": 0.00017034628506074892, "loss": 5.1825, "step": 531800 }, { "epoch": 3.2069680780501924, "grad_norm": 1.961119294166565, "learning_rate": 0.00017032612397479654, "loss": 5.0927, "step": 531850 }, { "epoch": 3.2072695698315266, "grad_norm": 1.3539713621139526, "learning_rate": 0.00017030596251477095, "loss": 4.7772, "step": 531900 }, { "epoch": 3.2075710616128603, "grad_norm": 1.9231935739517212, "learning_rate": 0.000170285800681043, "loss": 4.9612, "step": 531950 }, { "epoch": 3.2078725533941945, "grad_norm": 1.9461277723312378, "learning_rate": 0.0001702656384739839, "loss": 5.1139, "step": 532000 }, { "epoch": 3.2081740451755287, "grad_norm": 1.5711677074432373, "learning_rate": 0.0001702454758939646, "loss": 5.0114, "step": 532050 }, { "epoch": 3.2084755369568625, "grad_norm": 2.5279695987701416, "learning_rate": 0.00017022531294135626, "loss": 4.9126, "step": 532100 }, { "epoch": 3.2087770287381967, "grad_norm": 2.428035020828247, "learning_rate": 0.0001702051496165299, "loss": 5.0107, "step": 532150 }, { "epoch": 3.2090785205195305, "grad_norm": 2.1485581398010254, "learning_rate": 0.00017018498591985653, "loss": 5.0812, "step": 532200 }, { "epoch": 3.2093800123008647, "grad_norm": 1.628616213798523, "learning_rate": 0.00017016482185170734, "loss": 4.562, "step": 532250 }, { "epoch": 3.209681504082199, "grad_norm": 1.960590124130249, "learning_rate": 0.00017014465741245336, "loss": 5.0137, "step": 532300 }, { "epoch": 3.2099829958635326, "grad_norm": 1.2858655452728271, "learning_rate": 0.00017012449260246578, "loss": 4.7299, "step": 532350 }, { "epoch": 3.210284487644867, "grad_norm": 2.7674367427825928, "learning_rate": 0.00017010432742211554, "loss": 5.0296, "step": 532400 }, { "epoch": 3.2105859794262006, "grad_norm": 1.9165449142456055, "learning_rate": 0.00017008416187177389, "loss": 5.2497, "step": 532450 }, { "epoch": 3.210887471207535, "grad_norm": 2.388997793197632, "learning_rate": 0.00017006399595181188, "loss": 4.9847, "step": 532500 }, { "epoch": 3.211188962988869, "grad_norm": 2.2378828525543213, "learning_rate": 0.00017004382966260068, "loss": 4.1399, "step": 532550 }, { "epoch": 3.211490454770203, "grad_norm": 1.9334874153137207, "learning_rate": 0.00017002366300451142, "loss": 5.2024, "step": 532600 }, { "epoch": 3.211791946551537, "grad_norm": 1.7367193698883057, "learning_rate": 0.00017000349597791518, "loss": 4.891, "step": 532650 }, { "epoch": 3.212093438332871, "grad_norm": 1.5800871849060059, "learning_rate": 0.00016998332858318314, "loss": 4.9029, "step": 532700 }, { "epoch": 3.212394930114205, "grad_norm": 2.8712687492370605, "learning_rate": 0.00016996316082068652, "loss": 4.6758, "step": 532750 }, { "epoch": 3.212696421895539, "grad_norm": 2.7453248500823975, "learning_rate": 0.00016994299269079634, "loss": 5.1205, "step": 532800 }, { "epoch": 3.2129979136768734, "grad_norm": 2.16841197013855, "learning_rate": 0.00016992282419388385, "loss": 4.5626, "step": 532850 }, { "epoch": 3.213299405458207, "grad_norm": 2.152946710586548, "learning_rate": 0.00016990265533032022, "loss": 5.4166, "step": 532900 }, { "epoch": 3.2136008972395413, "grad_norm": 1.7815630435943604, "learning_rate": 0.00016988248610047665, "loss": 4.8522, "step": 532950 }, { "epoch": 3.213902389020875, "grad_norm": 1.416269302368164, "learning_rate": 0.00016986231650472424, "loss": 5.0282, "step": 533000 }, { "epoch": 3.2142038808022093, "grad_norm": 2.546868324279785, "learning_rate": 0.00016984214654343425, "loss": 4.7005, "step": 533050 }, { "epoch": 3.2145053725835435, "grad_norm": 2.2176520824432373, "learning_rate": 0.00016982197621697793, "loss": 4.7959, "step": 533100 }, { "epoch": 3.2148068643648773, "grad_norm": 0.4669398367404938, "learning_rate": 0.00016980180552572639, "loss": 4.8651, "step": 533150 }, { "epoch": 3.2151083561462115, "grad_norm": 0.8287612199783325, "learning_rate": 0.00016978163447005089, "loss": 4.7493, "step": 533200 }, { "epoch": 3.2154098479275457, "grad_norm": 1.877258539199829, "learning_rate": 0.0001697614630503226, "loss": 5.1046, "step": 533250 }, { "epoch": 3.2157113397088795, "grad_norm": 0.9293419718742371, "learning_rate": 0.00016974129126691277, "loss": 4.6802, "step": 533300 }, { "epoch": 3.2160128314902137, "grad_norm": 2.2635669708251953, "learning_rate": 0.00016972111912019272, "loss": 4.9401, "step": 533350 }, { "epoch": 3.2163143232715474, "grad_norm": 1.7707537412643433, "learning_rate": 0.00016970094661053362, "loss": 5.3678, "step": 533400 }, { "epoch": 3.2166158150528816, "grad_norm": 2.15598464012146, "learning_rate": 0.0001696807737383066, "loss": 4.4092, "step": 533450 }, { "epoch": 3.216917306834216, "grad_norm": 2.114504098892212, "learning_rate": 0.0001696606005038831, "loss": 4.6633, "step": 533500 }, { "epoch": 3.2172187986155496, "grad_norm": 2.6896281242370605, "learning_rate": 0.00016964042690763436, "loss": 5.1089, "step": 533550 }, { "epoch": 3.217520290396884, "grad_norm": 2.427124261856079, "learning_rate": 0.0001696202529499315, "loss": 4.9645, "step": 533600 }, { "epoch": 3.217821782178218, "grad_norm": 1.1119171380996704, "learning_rate": 0.00016960007863114592, "loss": 4.6349, "step": 533650 }, { "epoch": 3.2181232739595518, "grad_norm": 1.9169023036956787, "learning_rate": 0.00016957990395164886, "loss": 4.8998, "step": 533700 }, { "epoch": 3.218424765740886, "grad_norm": 1.6366753578186035, "learning_rate": 0.00016955972891181159, "loss": 5.0529, "step": 533750 }, { "epoch": 3.2187262575222197, "grad_norm": 3.0154457092285156, "learning_rate": 0.00016953955351200545, "loss": 4.8256, "step": 533800 }, { "epoch": 3.219027749303554, "grad_norm": 1.822326421737671, "learning_rate": 0.00016951937775260172, "loss": 5.1798, "step": 533850 }, { "epoch": 3.219329241084888, "grad_norm": 1.8432223796844482, "learning_rate": 0.00016949920163397166, "loss": 4.7811, "step": 533900 }, { "epoch": 3.219630732866222, "grad_norm": 2.2680459022521973, "learning_rate": 0.0001694790251564867, "loss": 5.2482, "step": 533950 }, { "epoch": 3.219932224647556, "grad_norm": 1.720241904258728, "learning_rate": 0.000169458848320518, "loss": 5.1614, "step": 534000 }, { "epoch": 3.219932224647556, "eval_loss": 5.348339080810547, "eval_runtime": 39.0196, "eval_samples_per_second": 13.122, "eval_steps_per_second": 6.561, "eval_tts_loss": 7.5147321494358215, "step": 534000 }, { "epoch": 3.2202337164288903, "grad_norm": 1.6179674863815308, "learning_rate": 0.00016943867112643698, "loss": 4.8779, "step": 534050 }, { "epoch": 3.220535208210224, "grad_norm": 2.0526785850524902, "learning_rate": 0.00016941849357461495, "loss": 5.1761, "step": 534100 }, { "epoch": 3.2208366999915583, "grad_norm": 1.8896493911743164, "learning_rate": 0.0001693983156654233, "loss": 4.6801, "step": 534150 }, { "epoch": 3.221138191772892, "grad_norm": 3.2997450828552246, "learning_rate": 0.0001693781373992333, "loss": 4.9376, "step": 534200 }, { "epoch": 3.2214396835542263, "grad_norm": 2.0063881874084473, "learning_rate": 0.00016935795877641636, "loss": 4.5525, "step": 534250 }, { "epoch": 3.2217411753355605, "grad_norm": 2.298438549041748, "learning_rate": 0.00016933777979734376, "loss": 4.7224, "step": 534300 }, { "epoch": 3.2220426671168942, "grad_norm": 2.0458569526672363, "learning_rate": 0.00016931760046238697, "loss": 4.618, "step": 534350 }, { "epoch": 3.2223441588982285, "grad_norm": 1.7131202220916748, "learning_rate": 0.0001692974207719173, "loss": 5.0125, "step": 534400 }, { "epoch": 3.2226456506795627, "grad_norm": 1.8443487882614136, "learning_rate": 0.00016927724072630612, "loss": 4.9104, "step": 534450 }, { "epoch": 3.2229471424608964, "grad_norm": 2.1077985763549805, "learning_rate": 0.00016925706032592486, "loss": 4.8505, "step": 534500 }, { "epoch": 3.2232486342422306, "grad_norm": 1.6866995096206665, "learning_rate": 0.0001692368795711449, "loss": 4.9062, "step": 534550 }, { "epoch": 3.2235501260235644, "grad_norm": 2.383613109588623, "learning_rate": 0.0001692166984623376, "loss": 4.8627, "step": 534600 }, { "epoch": 3.2238516178048986, "grad_norm": 1.7515428066253662, "learning_rate": 0.0001691965169998744, "loss": 5.0859, "step": 534650 }, { "epoch": 3.224153109586233, "grad_norm": 0.4247313439846039, "learning_rate": 0.00016917633518412668, "loss": 4.9026, "step": 534700 }, { "epoch": 3.2244546013675666, "grad_norm": 2.052443504333496, "learning_rate": 0.0001691561530154659, "loss": 4.6734, "step": 534750 }, { "epoch": 3.2247560931489008, "grad_norm": 1.533076286315918, "learning_rate": 0.00016913597049426343, "loss": 4.965, "step": 534800 }, { "epoch": 3.225057584930235, "grad_norm": 2.1284165382385254, "learning_rate": 0.00016911578762089074, "loss": 5.1668, "step": 534850 }, { "epoch": 3.2253590767115687, "grad_norm": 1.2728251218795776, "learning_rate": 0.00016909560439571926, "loss": 4.2004, "step": 534900 }, { "epoch": 3.225660568492903, "grad_norm": 1.8841429948806763, "learning_rate": 0.00016907542081912047, "loss": 5.0037, "step": 534950 }, { "epoch": 3.225962060274237, "grad_norm": 1.5297731161117554, "learning_rate": 0.00016905523689146577, "loss": 4.6588, "step": 535000 }, { "epoch": 3.226263552055571, "grad_norm": 2.28486967086792, "learning_rate": 0.00016903505261312658, "loss": 5.1666, "step": 535050 }, { "epoch": 3.226565043836905, "grad_norm": 1.9771019220352173, "learning_rate": 0.00016901486798447443, "loss": 5.37, "step": 535100 }, { "epoch": 3.226866535618239, "grad_norm": 1.7963173389434814, "learning_rate": 0.00016899468300588083, "loss": 4.5943, "step": 535150 }, { "epoch": 3.227168027399573, "grad_norm": 3.9917194843292236, "learning_rate": 0.00016897449767771714, "loss": 4.7219, "step": 535200 }, { "epoch": 3.2274695191809073, "grad_norm": 2.582514762878418, "learning_rate": 0.00016895431200035488, "loss": 4.6823, "step": 535250 }, { "epoch": 3.227771010962241, "grad_norm": 2.224210739135742, "learning_rate": 0.0001689341259741656, "loss": 5.2217, "step": 535300 }, { "epoch": 3.2280725027435753, "grad_norm": 1.4167190790176392, "learning_rate": 0.00016891393959952075, "loss": 5.2122, "step": 535350 }, { "epoch": 3.228373994524909, "grad_norm": 1.6148371696472168, "learning_rate": 0.0001688937528767918, "loss": 5.1431, "step": 535400 }, { "epoch": 3.2286754863062432, "grad_norm": 1.657332420349121, "learning_rate": 0.00016887356580635033, "loss": 4.7465, "step": 535450 }, { "epoch": 3.2289769780875774, "grad_norm": 1.8606579303741455, "learning_rate": 0.00016885337838856774, "loss": 4.9706, "step": 535500 }, { "epoch": 3.229278469868911, "grad_norm": 1.900138258934021, "learning_rate": 0.00016883319062381574, "loss": 5.1384, "step": 535550 }, { "epoch": 3.2295799616502454, "grad_norm": 2.9452085494995117, "learning_rate": 0.00016881300251246566, "loss": 4.8429, "step": 535600 }, { "epoch": 3.2298814534315796, "grad_norm": 1.715912103652954, "learning_rate": 0.00016879281405488914, "loss": 4.8398, "step": 535650 }, { "epoch": 3.2301829452129134, "grad_norm": 0.8240451216697693, "learning_rate": 0.0001687726252514577, "loss": 4.8456, "step": 535700 }, { "epoch": 3.2304844369942476, "grad_norm": 2.0837273597717285, "learning_rate": 0.0001687524361025429, "loss": 4.79, "step": 535750 }, { "epoch": 3.230785928775582, "grad_norm": 2.3996644020080566, "learning_rate": 0.00016873224660851624, "loss": 4.7194, "step": 535800 }, { "epoch": 3.2310874205569156, "grad_norm": 1.9179961681365967, "learning_rate": 0.0001687120567697493, "loss": 4.7663, "step": 535850 }, { "epoch": 3.2313889123382498, "grad_norm": 1.8101115226745605, "learning_rate": 0.00016869186658661367, "loss": 5.213, "step": 535900 }, { "epoch": 3.2316904041195835, "grad_norm": 1.4067811965942383, "learning_rate": 0.00016867167605948095, "loss": 4.9436, "step": 535950 }, { "epoch": 3.2319918959009177, "grad_norm": 1.7799397706985474, "learning_rate": 0.00016865148518872266, "loss": 4.9506, "step": 536000 }, { "epoch": 3.232293387682252, "grad_norm": 1.0021952390670776, "learning_rate": 0.00016863129397471038, "loss": 4.7638, "step": 536050 }, { "epoch": 3.2325948794635857, "grad_norm": 2.5456926822662354, "learning_rate": 0.00016861110241781573, "loss": 4.5538, "step": 536100 }, { "epoch": 3.23289637124492, "grad_norm": 2.7271289825439453, "learning_rate": 0.00016859091051841038, "loss": 4.7712, "step": 536150 }, { "epoch": 3.2331978630262537, "grad_norm": 1.1112289428710938, "learning_rate": 0.0001685707182768657, "loss": 4.8276, "step": 536200 }, { "epoch": 3.233499354807588, "grad_norm": 2.658254623413086, "learning_rate": 0.00016855052569355355, "loss": 4.7626, "step": 536250 }, { "epoch": 3.233800846588922, "grad_norm": 2.668686866760254, "learning_rate": 0.00016853033276884545, "loss": 4.7368, "step": 536300 }, { "epoch": 3.234102338370256, "grad_norm": 3.2954561710357666, "learning_rate": 0.000168510139503113, "loss": 4.6657, "step": 536350 }, { "epoch": 3.23440383015159, "grad_norm": 0.6454799175262451, "learning_rate": 0.00016848994589672785, "loss": 4.8071, "step": 536400 }, { "epoch": 3.2347053219329243, "grad_norm": 2.4973161220550537, "learning_rate": 0.00016846975195006163, "loss": 5.2323, "step": 536450 }, { "epoch": 3.235006813714258, "grad_norm": 2.353862762451172, "learning_rate": 0.00016844955766348596, "loss": 4.9742, "step": 536500 }, { "epoch": 3.2353083054955922, "grad_norm": 0.4493044912815094, "learning_rate": 0.00016842936303737257, "loss": 4.7712, "step": 536550 }, { "epoch": 3.2356097972769264, "grad_norm": 1.5072897672653198, "learning_rate": 0.00016840916807209303, "loss": 4.5788, "step": 536600 }, { "epoch": 3.23591128905826, "grad_norm": 1.6824394464492798, "learning_rate": 0.00016838897276801897, "loss": 4.9212, "step": 536650 }, { "epoch": 3.2362127808395944, "grad_norm": 1.7058099508285522, "learning_rate": 0.00016836877712552215, "loss": 5.0207, "step": 536700 }, { "epoch": 3.236514272620928, "grad_norm": 1.7380971908569336, "learning_rate": 0.00016834858114497423, "loss": 4.7701, "step": 536750 }, { "epoch": 3.2368157644022624, "grad_norm": 1.745123028755188, "learning_rate": 0.0001683283848267468, "loss": 4.6569, "step": 536800 }, { "epoch": 3.2371172561835966, "grad_norm": 1.569689154624939, "learning_rate": 0.00016830818817121162, "loss": 5.0534, "step": 536850 }, { "epoch": 3.2374187479649303, "grad_norm": 1.894822359085083, "learning_rate": 0.0001682879911787404, "loss": 4.7987, "step": 536900 }, { "epoch": 3.2377202397462646, "grad_norm": 2.779022693634033, "learning_rate": 0.0001682677938497048, "loss": 4.9575, "step": 536950 }, { "epoch": 3.2380217315275988, "grad_norm": 3.165158748626709, "learning_rate": 0.00016824759618447647, "loss": 4.6888, "step": 537000 }, { "epoch": 3.2380217315275988, "eval_loss": 5.333556175231934, "eval_runtime": 39.0447, "eval_samples_per_second": 13.113, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.446248149592669, "step": 537000 }, { "epoch": 3.2383232233089325, "grad_norm": 2.8470871448516846, "learning_rate": 0.00016822739818342722, "loss": 5.1735, "step": 537050 }, { "epoch": 3.2386247150902667, "grad_norm": 0.9156721234321594, "learning_rate": 0.00016820719984692866, "loss": 5.1949, "step": 537100 }, { "epoch": 3.2389262068716005, "grad_norm": 2.4968698024749756, "learning_rate": 0.00016818700117535266, "loss": 5.1065, "step": 537150 }, { "epoch": 3.2392276986529347, "grad_norm": 2.5095160007476807, "learning_rate": 0.00016816680216907084, "loss": 5.1186, "step": 537200 }, { "epoch": 3.239529190434269, "grad_norm": 4.421990394592285, "learning_rate": 0.00016814660282845484, "loss": 5.2832, "step": 537250 }, { "epoch": 3.2398306822156027, "grad_norm": 1.6927552223205566, "learning_rate": 0.0001681264031538766, "loss": 5.2528, "step": 537300 }, { "epoch": 3.240132173996937, "grad_norm": 2.577000856399536, "learning_rate": 0.0001681062031457078, "loss": 5.0152, "step": 537350 }, { "epoch": 3.240433665778271, "grad_norm": 1.3639508485794067, "learning_rate": 0.00016808600280432013, "loss": 5.0923, "step": 537400 }, { "epoch": 3.240735157559605, "grad_norm": 2.127197742462158, "learning_rate": 0.0001680658021300854, "loss": 4.8717, "step": 537450 }, { "epoch": 3.241036649340939, "grad_norm": 2.697166681289673, "learning_rate": 0.00016804560112337538, "loss": 4.8088, "step": 537500 }, { "epoch": 3.241338141122273, "grad_norm": 2.218966484069824, "learning_rate": 0.0001680253997845618, "loss": 5.0792, "step": 537550 }, { "epoch": 3.241639632903607, "grad_norm": 2.526296615600586, "learning_rate": 0.00016800519811401647, "loss": 5.2965, "step": 537600 }, { "epoch": 3.2419411246849412, "grad_norm": 2.8266818523406982, "learning_rate": 0.00016798499611211113, "loss": 4.8601, "step": 537650 }, { "epoch": 3.242242616466275, "grad_norm": 2.279810667037964, "learning_rate": 0.00016796479377921761, "loss": 4.7779, "step": 537700 }, { "epoch": 3.242544108247609, "grad_norm": 2.085824728012085, "learning_rate": 0.0001679445911157078, "loss": 4.977, "step": 537750 }, { "epoch": 3.2428456000289434, "grad_norm": 2.377958297729492, "learning_rate": 0.00016792438812195324, "loss": 4.9063, "step": 537800 }, { "epoch": 3.243147091810277, "grad_norm": 1.902059555053711, "learning_rate": 0.00016790418479832603, "loss": 5.1897, "step": 537850 }, { "epoch": 3.2434485835916114, "grad_norm": 2.39229416847229, "learning_rate": 0.00016788398114519777, "loss": 4.9265, "step": 537900 }, { "epoch": 3.243750075372945, "grad_norm": 1.7669178247451782, "learning_rate": 0.00016786377716294037, "loss": 4.7806, "step": 537950 }, { "epoch": 3.2440515671542793, "grad_norm": 2.9604651927948, "learning_rate": 0.00016784357285192566, "loss": 5.465, "step": 538000 }, { "epoch": 3.2443530589356135, "grad_norm": 2.4431774616241455, "learning_rate": 0.0001678233682125255, "loss": 4.7533, "step": 538050 }, { "epoch": 3.2446545507169473, "grad_norm": 1.9798355102539062, "learning_rate": 0.00016780316324511158, "loss": 4.9982, "step": 538100 }, { "epoch": 3.2449560424982815, "grad_norm": 1.677951455116272, "learning_rate": 0.00016778295795005592, "loss": 4.5652, "step": 538150 }, { "epoch": 3.2452575342796157, "grad_norm": 2.3644890785217285, "learning_rate": 0.0001677627523277303, "loss": 4.868, "step": 538200 }, { "epoch": 3.2455590260609495, "grad_norm": 1.6237713098526, "learning_rate": 0.0001677425463785065, "loss": 5.0079, "step": 538250 }, { "epoch": 3.2458605178422837, "grad_norm": 1.1176100969314575, "learning_rate": 0.00016772234010275654, "loss": 4.7859, "step": 538300 }, { "epoch": 3.2461620096236175, "grad_norm": 0.7529390454292297, "learning_rate": 0.00016770213350085218, "loss": 4.8171, "step": 538350 }, { "epoch": 3.2464635014049517, "grad_norm": 2.9087793827056885, "learning_rate": 0.00016768192657316526, "loss": 5.3988, "step": 538400 }, { "epoch": 3.246764993186286, "grad_norm": 2.142756223678589, "learning_rate": 0.00016766171932006778, "loss": 4.9478, "step": 538450 }, { "epoch": 3.2470664849676196, "grad_norm": 3.6095845699310303, "learning_rate": 0.00016764151174193155, "loss": 5.0298, "step": 538500 }, { "epoch": 3.247367976748954, "grad_norm": 1.8951175212860107, "learning_rate": 0.00016762130383912843, "loss": 4.9296, "step": 538550 }, { "epoch": 3.247669468530288, "grad_norm": 2.814016342163086, "learning_rate": 0.00016760109561203042, "loss": 4.5023, "step": 538600 }, { "epoch": 3.247970960311622, "grad_norm": 1.8327059745788574, "learning_rate": 0.00016758088706100933, "loss": 4.2413, "step": 538650 }, { "epoch": 3.248272452092956, "grad_norm": 3.1376800537109375, "learning_rate": 0.00016756067818643705, "loss": 5.0975, "step": 538700 }, { "epoch": 3.24857394387429, "grad_norm": 2.9536938667297363, "learning_rate": 0.00016754046898868568, "loss": 5.047, "step": 538750 }, { "epoch": 3.248875435655624, "grad_norm": 1.292479157447815, "learning_rate": 0.0001675202594681269, "loss": 5.0931, "step": 538800 }, { "epoch": 3.249176927436958, "grad_norm": 2.6448311805725098, "learning_rate": 0.00016750004962513276, "loss": 4.8681, "step": 538850 }, { "epoch": 3.249478419218292, "grad_norm": 2.6067044734954834, "learning_rate": 0.0001674798394600752, "loss": 5.1408, "step": 538900 }, { "epoch": 3.249779910999626, "grad_norm": 3.091658353805542, "learning_rate": 0.0001674596289733262, "loss": 5.0935, "step": 538950 }, { "epoch": 3.2500814027809604, "grad_norm": 1.6578903198242188, "learning_rate": 0.00016743941816525754, "loss": 4.5364, "step": 539000 }, { "epoch": 3.250382894562294, "grad_norm": 3.3058009147644043, "learning_rate": 0.00016741920703624135, "loss": 4.5704, "step": 539050 }, { "epoch": 3.2506843863436283, "grad_norm": 3.4694888591766357, "learning_rate": 0.00016739899558664947, "loss": 4.9924, "step": 539100 }, { "epoch": 3.250985878124962, "grad_norm": 1.7706509828567505, "learning_rate": 0.00016737878381685393, "loss": 5.5093, "step": 539150 }, { "epoch": 3.2512873699062963, "grad_norm": 1.0535743236541748, "learning_rate": 0.0001673585717272267, "loss": 4.6514, "step": 539200 }, { "epoch": 3.2515888616876305, "grad_norm": 2.0709714889526367, "learning_rate": 0.00016733835931813964, "loss": 4.6587, "step": 539250 }, { "epoch": 3.2518903534689643, "grad_norm": 1.7367031574249268, "learning_rate": 0.00016731814658996487, "loss": 4.8845, "step": 539300 }, { "epoch": 3.2521918452502985, "grad_norm": 1.2312508821487427, "learning_rate": 0.00016729793354307437, "loss": 4.9961, "step": 539350 }, { "epoch": 3.2524933370316327, "grad_norm": 1.613107681274414, "learning_rate": 0.00016727772017784002, "loss": 5.0417, "step": 539400 }, { "epoch": 3.2527948288129664, "grad_norm": 2.781865358352661, "learning_rate": 0.00016725750649463394, "loss": 5.0465, "step": 539450 }, { "epoch": 3.2530963205943006, "grad_norm": 3.9251739978790283, "learning_rate": 0.00016723729249382805, "loss": 4.8851, "step": 539500 }, { "epoch": 3.253397812375635, "grad_norm": 2.8138365745544434, "learning_rate": 0.00016721707817579442, "loss": 4.8651, "step": 539550 }, { "epoch": 3.2536993041569686, "grad_norm": 2.193035364151001, "learning_rate": 0.00016719686354090498, "loss": 5.4346, "step": 539600 }, { "epoch": 3.254000795938303, "grad_norm": 2.2210750579833984, "learning_rate": 0.00016717664858953185, "loss": 4.7607, "step": 539650 }, { "epoch": 3.2543022877196366, "grad_norm": 2.5493013858795166, "learning_rate": 0.000167156433322047, "loss": 4.7241, "step": 539700 }, { "epoch": 3.254603779500971, "grad_norm": 1.884093999862671, "learning_rate": 0.00016713621773882252, "loss": 4.6492, "step": 539750 }, { "epoch": 3.254905271282305, "grad_norm": 0.5557171106338501, "learning_rate": 0.00016711600184023036, "loss": 4.7176, "step": 539800 }, { "epoch": 3.2552067630636388, "grad_norm": 0.8826969265937805, "learning_rate": 0.0001670957856266426, "loss": 5.2831, "step": 539850 }, { "epoch": 3.255508254844973, "grad_norm": 1.8852087259292603, "learning_rate": 0.00016707556909843134, "loss": 4.8141, "step": 539900 }, { "epoch": 3.2558097466263067, "grad_norm": 2.738905429840088, "learning_rate": 0.00016705535225596864, "loss": 5.1236, "step": 539950 }, { "epoch": 3.256111238407641, "grad_norm": 2.106414794921875, "learning_rate": 0.00016703513509962642, "loss": 5.0274, "step": 540000 }, { "epoch": 3.256111238407641, "eval_loss": 5.3416852951049805, "eval_runtime": 39.1713, "eval_samples_per_second": 13.071, "eval_steps_per_second": 6.535, "eval_tts_loss": 7.399530182651385, "step": 540000 }, { "epoch": 3.256412730188975, "grad_norm": 3.2900874614715576, "learning_rate": 0.0001670149176297769, "loss": 5.3501, "step": 540050 }, { "epoch": 3.256714221970309, "grad_norm": 1.7658823728561401, "learning_rate": 0.0001669946998467921, "loss": 4.9436, "step": 540100 }, { "epoch": 3.257015713751643, "grad_norm": 2.049161434173584, "learning_rate": 0.0001669744817510441, "loss": 4.9353, "step": 540150 }, { "epoch": 3.2573172055329773, "grad_norm": 0.6434311270713806, "learning_rate": 0.000166954263342905, "loss": 4.636, "step": 540200 }, { "epoch": 3.257618697314311, "grad_norm": 2.3007848262786865, "learning_rate": 0.0001669340446227469, "loss": 4.8198, "step": 540250 }, { "epoch": 3.2579201890956453, "grad_norm": 1.5033890008926392, "learning_rate": 0.0001669138255909418, "loss": 4.6935, "step": 540300 }, { "epoch": 3.2582216808769795, "grad_norm": 1.8398364782333374, "learning_rate": 0.00016689360624786193, "loss": 5.0148, "step": 540350 }, { "epoch": 3.2585231726583133, "grad_norm": 2.342298746109009, "learning_rate": 0.00016687338659387933, "loss": 4.8079, "step": 540400 }, { "epoch": 3.2588246644396475, "grad_norm": 2.7364895343780518, "learning_rate": 0.00016685316662936613, "loss": 5.1209, "step": 540450 }, { "epoch": 3.2591261562209812, "grad_norm": 1.7647405862808228, "learning_rate": 0.00016683294635469446, "loss": 4.9963, "step": 540500 }, { "epoch": 3.2594276480023154, "grad_norm": 2.46317720413208, "learning_rate": 0.00016681272577023646, "loss": 4.7144, "step": 540550 }, { "epoch": 3.2597291397836496, "grad_norm": 2.1644864082336426, "learning_rate": 0.00016679250487636416, "loss": 4.5835, "step": 540600 }, { "epoch": 3.2600306315649834, "grad_norm": 3.2633216381073, "learning_rate": 0.0001667722836734498, "loss": 4.4475, "step": 540650 }, { "epoch": 3.2603321233463176, "grad_norm": 2.0946035385131836, "learning_rate": 0.00016675206216186554, "loss": 4.8645, "step": 540700 }, { "epoch": 3.2606336151276514, "grad_norm": 1.8435572385787964, "learning_rate": 0.00016673184034198346, "loss": 4.7997, "step": 540750 }, { "epoch": 3.2609351069089856, "grad_norm": 1.7724968194961548, "learning_rate": 0.0001667116182141757, "loss": 5.0619, "step": 540800 }, { "epoch": 3.26123659869032, "grad_norm": 2.5919156074523926, "learning_rate": 0.0001666913957788145, "loss": 5.1568, "step": 540850 }, { "epoch": 3.2615380904716536, "grad_norm": 0.6716054081916809, "learning_rate": 0.00016667117303627193, "loss": 4.9243, "step": 540900 }, { "epoch": 3.2618395822529878, "grad_norm": 1.8115931749343872, "learning_rate": 0.00016665094998692028, "loss": 5.1285, "step": 540950 }, { "epoch": 3.262141074034322, "grad_norm": 2.2706971168518066, "learning_rate": 0.0001666307266311316, "loss": 5.3326, "step": 541000 }, { "epoch": 3.2624425658156557, "grad_norm": 1.8759478330612183, "learning_rate": 0.00016661050296927816, "loss": 4.6979, "step": 541050 }, { "epoch": 3.26274405759699, "grad_norm": 1.9302911758422852, "learning_rate": 0.0001665902790017321, "loss": 5.1181, "step": 541100 }, { "epoch": 3.263045549378324, "grad_norm": 2.586294651031494, "learning_rate": 0.00016657005472886564, "loss": 4.9951, "step": 541150 }, { "epoch": 3.263347041159658, "grad_norm": 1.837702989578247, "learning_rate": 0.00016654983015105098, "loss": 4.7819, "step": 541200 }, { "epoch": 3.263648532940992, "grad_norm": 2.2107431888580322, "learning_rate": 0.00016652960526866028, "loss": 4.9047, "step": 541250 }, { "epoch": 3.263950024722326, "grad_norm": 2.3289027214050293, "learning_rate": 0.00016650938008206582, "loss": 4.8981, "step": 541300 }, { "epoch": 3.26425151650366, "grad_norm": 2.0479557514190674, "learning_rate": 0.00016648915459163978, "loss": 5.2227, "step": 541350 }, { "epoch": 3.2645530082849943, "grad_norm": 1.8924099206924438, "learning_rate": 0.00016646892879775436, "loss": 5.1957, "step": 541400 }, { "epoch": 3.264854500066328, "grad_norm": 1.84400475025177, "learning_rate": 0.0001664487027007818, "loss": 4.1056, "step": 541450 }, { "epoch": 3.2651559918476623, "grad_norm": 1.818199872970581, "learning_rate": 0.00016642847630109436, "loss": 4.4228, "step": 541500 }, { "epoch": 3.2654574836289965, "grad_norm": 4.697254657745361, "learning_rate": 0.0001664082495990643, "loss": 5.1626, "step": 541550 }, { "epoch": 3.2657589754103302, "grad_norm": 1.803161382675171, "learning_rate": 0.0001663880225950637, "loss": 5.2616, "step": 541600 }, { "epoch": 3.2660604671916644, "grad_norm": 3.043133020401001, "learning_rate": 0.00016636779528946504, "loss": 5.2799, "step": 541650 }, { "epoch": 3.2663619589729986, "grad_norm": 2.020179510116577, "learning_rate": 0.00016634756768264044, "loss": 5.2575, "step": 541700 }, { "epoch": 3.2666634507543324, "grad_norm": 1.476392149925232, "learning_rate": 0.0001663273397749622, "loss": 4.9123, "step": 541750 }, { "epoch": 3.2669649425356666, "grad_norm": 2.182405471801758, "learning_rate": 0.00016630711156680254, "loss": 4.566, "step": 541800 }, { "epoch": 3.2672664343170004, "grad_norm": 2.0452451705932617, "learning_rate": 0.00016628688305853377, "loss": 4.9476, "step": 541850 }, { "epoch": 3.2675679260983346, "grad_norm": 2.118725061416626, "learning_rate": 0.0001662666542505282, "loss": 4.9175, "step": 541900 }, { "epoch": 3.267869417879669, "grad_norm": 1.546452522277832, "learning_rate": 0.000166246425143158, "loss": 5.3214, "step": 541950 }, { "epoch": 3.2681709096610025, "grad_norm": 1.7741338014602661, "learning_rate": 0.00016622619573679556, "loss": 5.084, "step": 542000 }, { "epoch": 3.2684724014423367, "grad_norm": 2.1375861167907715, "learning_rate": 0.00016620596603181314, "loss": 5.251, "step": 542050 }, { "epoch": 3.2687738932236705, "grad_norm": 0.5116544961929321, "learning_rate": 0.00016618573602858306, "loss": 5.3175, "step": 542100 }, { "epoch": 3.2690753850050047, "grad_norm": 2.580289840698242, "learning_rate": 0.00016616550572747763, "loss": 5.1167, "step": 542150 }, { "epoch": 3.269376876786339, "grad_norm": 3.2356090545654297, "learning_rate": 0.00016614527512886907, "loss": 4.9707, "step": 542200 }, { "epoch": 3.2696783685676727, "grad_norm": 2.29459810256958, "learning_rate": 0.0001661250442331298, "loss": 4.7853, "step": 542250 }, { "epoch": 3.269979860349007, "grad_norm": 2.043627977371216, "learning_rate": 0.00016610481304063211, "loss": 4.6457, "step": 542300 }, { "epoch": 3.270281352130341, "grad_norm": 2.2010414600372314, "learning_rate": 0.00016608458155174832, "loss": 4.9852, "step": 542350 }, { "epoch": 3.270582843911675, "grad_norm": 2.1835434436798096, "learning_rate": 0.00016606434976685076, "loss": 4.6307, "step": 542400 }, { "epoch": 3.270884335693009, "grad_norm": 2.949920892715454, "learning_rate": 0.00016604411768631174, "loss": 4.8653, "step": 542450 }, { "epoch": 3.2711858274743433, "grad_norm": 1.5793156623840332, "learning_rate": 0.00016602388531050365, "loss": 4.6967, "step": 542500 }, { "epoch": 3.271487319255677, "grad_norm": 2.6924571990966797, "learning_rate": 0.0001660036526397988, "loss": 4.7314, "step": 542550 }, { "epoch": 3.2717888110370112, "grad_norm": 2.7061586380004883, "learning_rate": 0.00016598341967456956, "loss": 4.847, "step": 542600 }, { "epoch": 3.272090302818345, "grad_norm": 0.6589584946632385, "learning_rate": 0.00016596318641518834, "loss": 4.8331, "step": 542650 }, { "epoch": 3.272391794599679, "grad_norm": 3.2564992904663086, "learning_rate": 0.00016594295286202738, "loss": 4.9614, "step": 542700 }, { "epoch": 3.2726932863810134, "grad_norm": 3.1479811668395996, "learning_rate": 0.0001659227190154592, "loss": 4.755, "step": 542750 }, { "epoch": 3.272994778162347, "grad_norm": 2.491528034210205, "learning_rate": 0.00016590248487585607, "loss": 5.0681, "step": 542800 }, { "epoch": 3.2732962699436814, "grad_norm": 2.1194090843200684, "learning_rate": 0.00016588225044359038, "loss": 5.0688, "step": 542850 }, { "epoch": 3.273597761725015, "grad_norm": 2.688295364379883, "learning_rate": 0.00016586201571903455, "loss": 5.3464, "step": 542900 }, { "epoch": 3.2738992535063494, "grad_norm": 3.05876088142395, "learning_rate": 0.000165841780702561, "loss": 4.8774, "step": 542950 }, { "epoch": 3.2742007452876836, "grad_norm": 1.673090934753418, "learning_rate": 0.00016582154539454205, "loss": 4.8199, "step": 543000 }, { "epoch": 3.2742007452876836, "eval_loss": 5.320701599121094, "eval_runtime": 39.249, "eval_samples_per_second": 13.045, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.521426215106857, "step": 543000 }, { "epoch": 3.2745022370690173, "grad_norm": 0.472184956073761, "learning_rate": 0.00016580130979535008, "loss": 4.7175, "step": 543050 }, { "epoch": 3.2748037288503515, "grad_norm": 2.624027729034424, "learning_rate": 0.00016578107390535763, "loss": 5.5418, "step": 543100 }, { "epoch": 3.2751052206316857, "grad_norm": 1.6338427066802979, "learning_rate": 0.000165760837724937, "loss": 4.8148, "step": 543150 }, { "epoch": 3.2754067124130195, "grad_norm": 2.090013265609741, "learning_rate": 0.00016574060125446063, "loss": 4.5896, "step": 543200 }, { "epoch": 3.2757082041943537, "grad_norm": 1.4990521669387817, "learning_rate": 0.00016572036449430099, "loss": 4.9266, "step": 543250 }, { "epoch": 3.276009695975688, "grad_norm": 1.2589480876922607, "learning_rate": 0.00016570012744483045, "loss": 4.5399, "step": 543300 }, { "epoch": 3.2763111877570217, "grad_norm": 1.9501891136169434, "learning_rate": 0.0001656798901064215, "loss": 5.0319, "step": 543350 }, { "epoch": 3.276612679538356, "grad_norm": 2.977355718612671, "learning_rate": 0.00016565965247944652, "loss": 5.1643, "step": 543400 }, { "epoch": 3.2769141713196897, "grad_norm": 2.0448315143585205, "learning_rate": 0.00016563941456427798, "loss": 5.3151, "step": 543450 }, { "epoch": 3.277215663101024, "grad_norm": 1.272970199584961, "learning_rate": 0.00016561917636128835, "loss": 5.421, "step": 543500 }, { "epoch": 3.277517154882358, "grad_norm": 6.210190773010254, "learning_rate": 0.0001655989378708501, "loss": 5.0955, "step": 543550 }, { "epoch": 3.277818646663692, "grad_norm": 2.297229051589966, "learning_rate": 0.00016557869909333562, "loss": 5.0974, "step": 543600 }, { "epoch": 3.278120138445026, "grad_norm": 3.917351722717285, "learning_rate": 0.0001655584600291174, "loss": 5.0167, "step": 543650 }, { "epoch": 3.27842163022636, "grad_norm": 1.6764241456985474, "learning_rate": 0.00016553822067856793, "loss": 4.5412, "step": 543700 }, { "epoch": 3.278723122007694, "grad_norm": 2.012441873550415, "learning_rate": 0.00016551798104205974, "loss": 5.122, "step": 543750 }, { "epoch": 3.279024613789028, "grad_norm": 2.2443666458129883, "learning_rate": 0.0001654977411199652, "loss": 4.7873, "step": 543800 }, { "epoch": 3.279326105570362, "grad_norm": 2.314424514770508, "learning_rate": 0.00016547750091265687, "loss": 4.5439, "step": 543850 }, { "epoch": 3.279627597351696, "grad_norm": 2.1054186820983887, "learning_rate": 0.00016545726042050722, "loss": 4.9668, "step": 543900 }, { "epoch": 3.2799290891330304, "grad_norm": 1.8356786966323853, "learning_rate": 0.00016543701964388875, "loss": 5.0078, "step": 543950 }, { "epoch": 3.280230580914364, "grad_norm": 2.572514057159424, "learning_rate": 0.00016541677858317397, "loss": 5.1807, "step": 544000 }, { "epoch": 3.2805320726956984, "grad_norm": 2.052032232284546, "learning_rate": 0.00016539653723873535, "loss": 4.9905, "step": 544050 }, { "epoch": 3.2808335644770326, "grad_norm": 4.768472194671631, "learning_rate": 0.00016537629561094548, "loss": 4.9635, "step": 544100 }, { "epoch": 3.2811350562583663, "grad_norm": 2.4626293182373047, "learning_rate": 0.00016535605370017684, "loss": 5.0573, "step": 544150 }, { "epoch": 3.2814365480397005, "grad_norm": 4.485067367553711, "learning_rate": 0.00016533581150680193, "loss": 4.7781, "step": 544200 }, { "epoch": 3.2817380398210343, "grad_norm": 4.418830871582031, "learning_rate": 0.00016531556903119323, "loss": 5.2743, "step": 544250 }, { "epoch": 3.2820395316023685, "grad_norm": 1.9066729545593262, "learning_rate": 0.00016529532627372338, "loss": 5.3015, "step": 544300 }, { "epoch": 3.2823410233837027, "grad_norm": 1.854164719581604, "learning_rate": 0.00016527508323476488, "loss": 4.8988, "step": 544350 }, { "epoch": 3.2826425151650365, "grad_norm": 3.284372091293335, "learning_rate": 0.0001652548399146903, "loss": 4.7624, "step": 544400 }, { "epoch": 3.2829440069463707, "grad_norm": 1.1683566570281982, "learning_rate": 0.00016523459631387214, "loss": 4.938, "step": 544450 }, { "epoch": 3.2832454987277044, "grad_norm": 3.6458961963653564, "learning_rate": 0.00016521435243268298, "loss": 5.4023, "step": 544500 }, { "epoch": 3.2835469905090386, "grad_norm": 2.402506113052368, "learning_rate": 0.00016519410827149538, "loss": 5.1311, "step": 544550 }, { "epoch": 3.283848482290373, "grad_norm": 3.1956284046173096, "learning_rate": 0.00016517386383068188, "loss": 4.9244, "step": 544600 }, { "epoch": 3.2841499740717066, "grad_norm": 1.2664631605148315, "learning_rate": 0.00016515361911061507, "loss": 4.8888, "step": 544650 }, { "epoch": 3.284451465853041, "grad_norm": 0.6065784692764282, "learning_rate": 0.0001651333741116676, "loss": 4.8095, "step": 544700 }, { "epoch": 3.284752957634375, "grad_norm": 2.0545101165771484, "learning_rate": 0.0001651131288342119, "loss": 4.7664, "step": 544750 }, { "epoch": 3.285054449415709, "grad_norm": 2.0021634101867676, "learning_rate": 0.0001650928832786206, "loss": 4.6728, "step": 544800 }, { "epoch": 3.285355941197043, "grad_norm": 2.178678035736084, "learning_rate": 0.00016507263744526638, "loss": 4.9193, "step": 544850 }, { "epoch": 3.285657432978377, "grad_norm": 2.9277455806732178, "learning_rate": 0.00016505239133452176, "loss": 4.6191, "step": 544900 }, { "epoch": 3.285958924759711, "grad_norm": 2.7876009941101074, "learning_rate": 0.00016503214494675936, "loss": 4.9719, "step": 544950 }, { "epoch": 3.286260416541045, "grad_norm": 2.2976157665252686, "learning_rate": 0.0001650118982823518, "loss": 5.0368, "step": 545000 }, { "epoch": 3.286561908322379, "grad_norm": 1.5947798490524292, "learning_rate": 0.00016499165134167167, "loss": 4.6339, "step": 545050 }, { "epoch": 3.286863400103713, "grad_norm": 1.868100643157959, "learning_rate": 0.0001649714041250916, "loss": 5.5439, "step": 545100 }, { "epoch": 3.2871648918850473, "grad_norm": 0.4185992181301117, "learning_rate": 0.00016495115663298416, "loss": 4.9522, "step": 545150 }, { "epoch": 3.287466383666381, "grad_norm": 2.8853442668914795, "learning_rate": 0.00016493090886572206, "loss": 4.7455, "step": 545200 }, { "epoch": 3.2877678754477153, "grad_norm": 2.087554693222046, "learning_rate": 0.00016491066082367786, "loss": 4.7298, "step": 545250 }, { "epoch": 3.2880693672290495, "grad_norm": 1.7970178127288818, "learning_rate": 0.00016489041250722426, "loss": 4.9617, "step": 545300 }, { "epoch": 3.2883708590103833, "grad_norm": 2.5991759300231934, "learning_rate": 0.00016487016391673388, "loss": 5.0768, "step": 545350 }, { "epoch": 3.2886723507917175, "grad_norm": 1.8642101287841797, "learning_rate": 0.0001648499150525793, "loss": 4.9407, "step": 545400 }, { "epoch": 3.2889738425730517, "grad_norm": 3.0736284255981445, "learning_rate": 0.00016482966591513322, "loss": 4.9901, "step": 545450 }, { "epoch": 3.2892753343543855, "grad_norm": 1.8795859813690186, "learning_rate": 0.0001648094165047683, "loss": 4.8878, "step": 545500 }, { "epoch": 3.2895768261357197, "grad_norm": 2.517043352127075, "learning_rate": 0.00016478916682185728, "loss": 5.4654, "step": 545550 }, { "epoch": 3.2898783179170534, "grad_norm": 2.6627817153930664, "learning_rate": 0.00016476891686677272, "loss": 4.8895, "step": 545600 }, { "epoch": 3.2901798096983876, "grad_norm": 2.335132122039795, "learning_rate": 0.0001647486666398873, "loss": 5.052, "step": 545650 }, { "epoch": 3.290481301479722, "grad_norm": 2.7636659145355225, "learning_rate": 0.00016472841614157373, "loss": 4.806, "step": 545700 }, { "epoch": 3.2907827932610556, "grad_norm": 2.307792901992798, "learning_rate": 0.00016470816537220465, "loss": 4.9474, "step": 545750 }, { "epoch": 3.29108428504239, "grad_norm": 2.750884532928467, "learning_rate": 0.0001646879143321528, "loss": 4.6953, "step": 545800 }, { "epoch": 3.2913857768237236, "grad_norm": 0.9069152474403381, "learning_rate": 0.00016466766302179083, "loss": 4.918, "step": 545850 }, { "epoch": 3.291687268605058, "grad_norm": 2.228304386138916, "learning_rate": 0.00016464741144149147, "loss": 4.8435, "step": 545900 }, { "epoch": 3.291988760386392, "grad_norm": 2.2116010189056396, "learning_rate": 0.0001646271595916274, "loss": 5.1089, "step": 545950 }, { "epoch": 3.2922902521677258, "grad_norm": 3.085869789123535, "learning_rate": 0.00016460690747257135, "loss": 4.8219, "step": 546000 }, { "epoch": 3.2922902521677258, "eval_loss": 5.345025062561035, "eval_runtime": 38.8347, "eval_samples_per_second": 13.184, "eval_steps_per_second": 6.592, "eval_tts_loss": 7.506353658156971, "step": 546000 }, { "epoch": 3.29259174394906, "grad_norm": 1.5056822299957275, "learning_rate": 0.000164586655084696, "loss": 4.6086, "step": 546050 }, { "epoch": 3.292893235730394, "grad_norm": 1.6984015703201294, "learning_rate": 0.00016456640242837407, "loss": 4.7747, "step": 546100 }, { "epoch": 3.293194727511728, "grad_norm": 2.093250274658203, "learning_rate": 0.00016454614950397832, "loss": 5.3089, "step": 546150 }, { "epoch": 3.293496219293062, "grad_norm": 2.939835548400879, "learning_rate": 0.00016452589631188143, "loss": 4.7069, "step": 546200 }, { "epoch": 3.2937977110743963, "grad_norm": 2.667644500732422, "learning_rate": 0.00016450564285245612, "loss": 4.4588, "step": 546250 }, { "epoch": 3.29409920285573, "grad_norm": 2.336256980895996, "learning_rate": 0.00016448538912607522, "loss": 4.8698, "step": 546300 }, { "epoch": 3.2944006946370643, "grad_norm": 1.7711706161499023, "learning_rate": 0.00016446513513311136, "loss": 5.1341, "step": 546350 }, { "epoch": 3.294702186418398, "grad_norm": 0.9404348731040955, "learning_rate": 0.00016444488087393732, "loss": 4.6868, "step": 546400 }, { "epoch": 3.2950036781997323, "grad_norm": 2.37646746635437, "learning_rate": 0.0001644246263489259, "loss": 4.6559, "step": 546450 }, { "epoch": 3.2953051699810665, "grad_norm": 0.6552010774612427, "learning_rate": 0.00016440437155844977, "loss": 4.9401, "step": 546500 }, { "epoch": 3.2956066617624002, "grad_norm": 0.9836570024490356, "learning_rate": 0.00016438411650288178, "loss": 4.8885, "step": 546550 }, { "epoch": 3.2959081535437345, "grad_norm": 2.0159854888916016, "learning_rate": 0.00016436386118259464, "loss": 4.3683, "step": 546600 }, { "epoch": 3.296209645325068, "grad_norm": 2.1032674312591553, "learning_rate": 0.0001643436055979611, "loss": 4.6613, "step": 546650 }, { "epoch": 3.2965111371064024, "grad_norm": 2.321263074874878, "learning_rate": 0.000164323349749354, "loss": 4.9762, "step": 546700 }, { "epoch": 3.2968126288877366, "grad_norm": 1.2649192810058594, "learning_rate": 0.0001643030936371461, "loss": 4.8887, "step": 546750 }, { "epoch": 3.2971141206690704, "grad_norm": 2.597104549407959, "learning_rate": 0.00016428283726171015, "loss": 4.4592, "step": 546800 }, { "epoch": 3.2974156124504046, "grad_norm": 2.3912785053253174, "learning_rate": 0.00016426258062341893, "loss": 5.0608, "step": 546850 }, { "epoch": 3.297717104231739, "grad_norm": 1.9443074464797974, "learning_rate": 0.00016424232372264536, "loss": 4.7685, "step": 546900 }, { "epoch": 3.2980185960130726, "grad_norm": 2.2047464847564697, "learning_rate": 0.0001642220665597621, "loss": 4.8153, "step": 546950 }, { "epoch": 3.2983200877944068, "grad_norm": 2.0270631313323975, "learning_rate": 0.00016420180913514194, "loss": 4.7874, "step": 547000 }, { "epoch": 3.298621579575741, "grad_norm": 1.8400410413742065, "learning_rate": 0.00016418155144915783, "loss": 4.6732, "step": 547050 }, { "epoch": 3.2989230713570747, "grad_norm": 1.8798480033874512, "learning_rate": 0.00016416129350218248, "loss": 4.4172, "step": 547100 }, { "epoch": 3.299224563138409, "grad_norm": 1.9195160865783691, "learning_rate": 0.00016414103529458873, "loss": 4.9256, "step": 547150 }, { "epoch": 3.2995260549197427, "grad_norm": 0.9018272757530212, "learning_rate": 0.00016412077682674943, "loss": 4.8072, "step": 547200 }, { "epoch": 3.299827546701077, "grad_norm": 2.2731549739837646, "learning_rate": 0.00016410051809903734, "loss": 5.0883, "step": 547250 }, { "epoch": 3.300129038482411, "grad_norm": 1.9804891347885132, "learning_rate": 0.00016408025911182535, "loss": 4.8306, "step": 547300 }, { "epoch": 3.300430530263745, "grad_norm": 3.7231476306915283, "learning_rate": 0.00016405999986548631, "loss": 4.9394, "step": 547350 }, { "epoch": 3.300732022045079, "grad_norm": 2.719857931137085, "learning_rate": 0.00016403974036039302, "loss": 5.2276, "step": 547400 }, { "epoch": 3.301033513826413, "grad_norm": 1.7323230504989624, "learning_rate": 0.00016401948059691833, "loss": 5.2103, "step": 547450 }, { "epoch": 3.301335005607747, "grad_norm": 1.612640380859375, "learning_rate": 0.00016399922057543515, "loss": 4.8095, "step": 547500 }, { "epoch": 3.3016364973890813, "grad_norm": 1.9045532941818237, "learning_rate": 0.00016397896029631627, "loss": 4.2952, "step": 547550 }, { "epoch": 3.301937989170415, "grad_norm": 1.9304953813552856, "learning_rate": 0.00016395869975993453, "loss": 4.8929, "step": 547600 }, { "epoch": 3.3022394809517492, "grad_norm": 1.8006596565246582, "learning_rate": 0.00016393843896666288, "loss": 4.5564, "step": 547650 }, { "epoch": 3.3025409727330834, "grad_norm": 3.0256330966949463, "learning_rate": 0.00016391817791687417, "loss": 5.2676, "step": 547700 }, { "epoch": 3.302842464514417, "grad_norm": 2.512284278869629, "learning_rate": 0.00016389791661094126, "loss": 5.1346, "step": 547750 }, { "epoch": 3.3031439562957514, "grad_norm": 2.8854899406433105, "learning_rate": 0.00016387765504923698, "loss": 4.9976, "step": 547800 }, { "epoch": 3.3034454480770856, "grad_norm": 1.5999494791030884, "learning_rate": 0.00016385739323213425, "loss": 4.6079, "step": 547850 }, { "epoch": 3.3037469398584194, "grad_norm": 2.2315688133239746, "learning_rate": 0.00016383713116000607, "loss": 4.885, "step": 547900 }, { "epoch": 3.3040484316397536, "grad_norm": 2.0601861476898193, "learning_rate": 0.00016381686883322518, "loss": 5.0451, "step": 547950 }, { "epoch": 3.3043499234210874, "grad_norm": 2.053042411804199, "learning_rate": 0.0001637966062521645, "loss": 5.3898, "step": 548000 }, { "epoch": 3.3046514152024216, "grad_norm": 2.2375853061676025, "learning_rate": 0.00016377634341719702, "loss": 4.8809, "step": 548050 }, { "epoch": 3.3049529069837558, "grad_norm": 2.790562391281128, "learning_rate": 0.00016375608032869564, "loss": 5.0258, "step": 548100 }, { "epoch": 3.3052543987650895, "grad_norm": 0.8763882517814636, "learning_rate": 0.00016373581698703312, "loss": 4.4172, "step": 548150 }, { "epoch": 3.3055558905464237, "grad_norm": 2.765321731567383, "learning_rate": 0.00016371555339258256, "loss": 5.044, "step": 548200 }, { "epoch": 3.3058573823277575, "grad_norm": 2.7500805854797363, "learning_rate": 0.0001636952895457168, "loss": 4.7791, "step": 548250 }, { "epoch": 3.3061588741090917, "grad_norm": 1.9051913022994995, "learning_rate": 0.0001636750254468088, "loss": 4.9706, "step": 548300 }, { "epoch": 3.306460365890426, "grad_norm": 1.599769115447998, "learning_rate": 0.00016365476109623148, "loss": 5.0295, "step": 548350 }, { "epoch": 3.3067618576717597, "grad_norm": 2.7242558002471924, "learning_rate": 0.00016363449649435774, "loss": 4.7585, "step": 548400 }, { "epoch": 3.307063349453094, "grad_norm": 1.9400691986083984, "learning_rate": 0.0001636142316415605, "loss": 4.9, "step": 548450 }, { "epoch": 3.307364841234428, "grad_norm": 2.0129895210266113, "learning_rate": 0.00016359396653821287, "loss": 5.2801, "step": 548500 }, { "epoch": 3.307666333015762, "grad_norm": 2.4215238094329834, "learning_rate": 0.00016357370118468762, "loss": 5.1778, "step": 548550 }, { "epoch": 3.307967824797096, "grad_norm": 1.0825918912887573, "learning_rate": 0.00016355343558135775, "loss": 4.9456, "step": 548600 }, { "epoch": 3.3082693165784303, "grad_norm": 2.1310436725616455, "learning_rate": 0.00016353316972859627, "loss": 4.5729, "step": 548650 }, { "epoch": 3.308570808359764, "grad_norm": 2.455695867538452, "learning_rate": 0.00016351290362677612, "loss": 4.9478, "step": 548700 }, { "epoch": 3.3088723001410982, "grad_norm": 2.267847776412964, "learning_rate": 0.00016349263727627027, "loss": 4.6945, "step": 548750 }, { "epoch": 3.309173791922432, "grad_norm": 1.7083241939544678, "learning_rate": 0.00016347237067745166, "loss": 4.9878, "step": 548800 }, { "epoch": 3.309475283703766, "grad_norm": 2.370812177658081, "learning_rate": 0.0001634521038306933, "loss": 4.6313, "step": 548850 }, { "epoch": 3.3097767754851004, "grad_norm": 2.499403238296509, "learning_rate": 0.00016343183673636817, "loss": 4.7466, "step": 548900 }, { "epoch": 3.310078267266434, "grad_norm": 1.6776736974716187, "learning_rate": 0.00016341156939484925, "loss": 5.1707, "step": 548950 }, { "epoch": 3.3103797590477684, "grad_norm": 0.7727367281913757, "learning_rate": 0.00016339130180650953, "loss": 4.7182, "step": 549000 }, { "epoch": 3.3103797590477684, "eval_loss": 5.3300886154174805, "eval_runtime": 39.3109, "eval_samples_per_second": 13.024, "eval_steps_per_second": 6.512, "eval_tts_loss": 7.48436141809053, "step": 549000 }, { "epoch": 3.3106812508291026, "grad_norm": 2.7061188220977783, "learning_rate": 0.000163371033971722, "loss": 4.801, "step": 549050 }, { "epoch": 3.3109827426104363, "grad_norm": 1.9628316164016724, "learning_rate": 0.00016335076589085973, "loss": 4.9471, "step": 549100 }, { "epoch": 3.3112842343917706, "grad_norm": 1.1195547580718994, "learning_rate": 0.00016333049756429562, "loss": 5.0719, "step": 549150 }, { "epoch": 3.3115857261731048, "grad_norm": 2.38511061668396, "learning_rate": 0.00016331022899240268, "loss": 5.013, "step": 549200 }, { "epoch": 3.3118872179544385, "grad_norm": 1.8135350942611694, "learning_rate": 0.00016328996017555406, "loss": 5.1945, "step": 549250 }, { "epoch": 3.3121887097357727, "grad_norm": 2.1838603019714355, "learning_rate": 0.00016326969111412265, "loss": 5.2797, "step": 549300 }, { "epoch": 3.3124902015171065, "grad_norm": 1.8764147758483887, "learning_rate": 0.0001632494218084815, "loss": 4.7234, "step": 549350 }, { "epoch": 3.3127916932984407, "grad_norm": 1.795884370803833, "learning_rate": 0.00016322915225900367, "loss": 4.9311, "step": 549400 }, { "epoch": 3.313093185079775, "grad_norm": 1.4587525129318237, "learning_rate": 0.00016320888246606214, "loss": 5.27, "step": 549450 }, { "epoch": 3.3133946768611087, "grad_norm": 1.2680888175964355, "learning_rate": 0.00016318861243003, "loss": 4.6568, "step": 549500 }, { "epoch": 3.313696168642443, "grad_norm": 1.7700369358062744, "learning_rate": 0.00016316834215128028, "loss": 4.7926, "step": 549550 }, { "epoch": 3.3139976604237766, "grad_norm": 2.5174407958984375, "learning_rate": 0.000163148071630186, "loss": 4.7135, "step": 549600 }, { "epoch": 3.314299152205111, "grad_norm": 2.4669365882873535, "learning_rate": 0.00016312780086712023, "loss": 4.7122, "step": 549650 }, { "epoch": 3.314600643986445, "grad_norm": 1.9581918716430664, "learning_rate": 0.00016310752986245606, "loss": 4.7587, "step": 549700 }, { "epoch": 3.314902135767779, "grad_norm": 2.3542401790618896, "learning_rate": 0.00016308725861656641, "loss": 4.5363, "step": 549750 }, { "epoch": 3.315203627549113, "grad_norm": 2.6770687103271484, "learning_rate": 0.00016306698712982454, "loss": 4.6315, "step": 549800 }, { "epoch": 3.3155051193304472, "grad_norm": 2.520735025405884, "learning_rate": 0.00016304671540260337, "loss": 4.8484, "step": 549850 }, { "epoch": 3.315806611111781, "grad_norm": 2.34332275390625, "learning_rate": 0.00016302644343527603, "loss": 4.8877, "step": 549900 }, { "epoch": 3.316108102893115, "grad_norm": 1.7432928085327148, "learning_rate": 0.0001630061712282156, "loss": 4.7093, "step": 549950 }, { "epoch": 3.3164095946744494, "grad_norm": 1.9740701913833618, "learning_rate": 0.00016298589878179515, "loss": 5.2742, "step": 550000 }, { "epoch": 3.316711086455783, "grad_norm": 2.3298768997192383, "learning_rate": 0.0001629656260963877, "loss": 4.7533, "step": 550050 }, { "epoch": 3.3170125782371174, "grad_norm": 0.7455865144729614, "learning_rate": 0.00016294535317236652, "loss": 4.5399, "step": 550100 }, { "epoch": 3.317314070018451, "grad_norm": 1.9199435710906982, "learning_rate": 0.00016292508001010452, "loss": 4.8493, "step": 550150 }, { "epoch": 3.3176155617997853, "grad_norm": 0.5814194679260254, "learning_rate": 0.00016290480660997484, "loss": 4.8391, "step": 550200 }, { "epoch": 3.3179170535811195, "grad_norm": 2.010749340057373, "learning_rate": 0.00016288453297235065, "loss": 4.7347, "step": 550250 }, { "epoch": 3.3182185453624533, "grad_norm": 1.8715983629226685, "learning_rate": 0.00016286425909760503, "loss": 4.8411, "step": 550300 }, { "epoch": 3.3185200371437875, "grad_norm": 2.624751091003418, "learning_rate": 0.00016284398498611104, "loss": 5.0528, "step": 550350 }, { "epoch": 3.3188215289251213, "grad_norm": 1.0125161409378052, "learning_rate": 0.00016282371063824184, "loss": 4.7913, "step": 550400 }, { "epoch": 3.3191230207064555, "grad_norm": 1.9589942693710327, "learning_rate": 0.00016280343605437058, "loss": 4.797, "step": 550450 }, { "epoch": 3.3194245124877897, "grad_norm": 1.9302983283996582, "learning_rate": 0.00016278316123487035, "loss": 4.9149, "step": 550500 }, { "epoch": 3.3197260042691235, "grad_norm": 2.226038694381714, "learning_rate": 0.00016276288618011426, "loss": 4.4803, "step": 550550 }, { "epoch": 3.3200274960504577, "grad_norm": 2.0109848976135254, "learning_rate": 0.00016274261089047547, "loss": 5.278, "step": 550600 }, { "epoch": 3.320328987831792, "grad_norm": 2.5905776023864746, "learning_rate": 0.00016272233536632707, "loss": 4.4729, "step": 550650 }, { "epoch": 3.3206304796131256, "grad_norm": 2.563751220703125, "learning_rate": 0.00016270205960804233, "loss": 4.983, "step": 550700 }, { "epoch": 3.32093197139446, "grad_norm": 1.6473568677902222, "learning_rate": 0.00016268178361599425, "loss": 4.8638, "step": 550750 }, { "epoch": 3.321233463175794, "grad_norm": 1.8051552772521973, "learning_rate": 0.00016266150739055604, "loss": 5.093, "step": 550800 }, { "epoch": 3.321534954957128, "grad_norm": 1.961236596107483, "learning_rate": 0.00016264123093210088, "loss": 4.8136, "step": 550850 }, { "epoch": 3.321836446738462, "grad_norm": 1.7531038522720337, "learning_rate": 0.0001626209542410019, "loss": 5.0009, "step": 550900 }, { "epoch": 3.3221379385197958, "grad_norm": 1.8432966470718384, "learning_rate": 0.00016260067731763223, "loss": 5.0014, "step": 550950 }, { "epoch": 3.32243943030113, "grad_norm": 2.496779680252075, "learning_rate": 0.00016258040016236514, "loss": 4.6733, "step": 551000 }, { "epoch": 3.322740922082464, "grad_norm": 1.7811046838760376, "learning_rate": 0.00016256012277557369, "loss": 5.1401, "step": 551050 }, { "epoch": 3.323042413863798, "grad_norm": 2.950418710708618, "learning_rate": 0.00016253984515763115, "loss": 4.3965, "step": 551100 }, { "epoch": 3.323343905645132, "grad_norm": 2.582180976867676, "learning_rate": 0.00016251956730891062, "loss": 5.012, "step": 551150 }, { "epoch": 3.323645397426466, "grad_norm": 2.542335271835327, "learning_rate": 0.00016249928922978528, "loss": 4.2072, "step": 551200 }, { "epoch": 3.3239468892078, "grad_norm": 2.033724069595337, "learning_rate": 0.0001624790109206284, "loss": 4.7295, "step": 551250 }, { "epoch": 3.3242483809891343, "grad_norm": 1.590067982673645, "learning_rate": 0.0001624587323818132, "loss": 5.7263, "step": 551300 }, { "epoch": 3.324549872770468, "grad_norm": 2.3883798122406006, "learning_rate": 0.0001624384536137127, "loss": 4.8863, "step": 551350 }, { "epoch": 3.3248513645518023, "grad_norm": 2.0470616817474365, "learning_rate": 0.00016241817461670028, "loss": 4.7172, "step": 551400 }, { "epoch": 3.3251528563331365, "grad_norm": 1.7781587839126587, "learning_rate": 0.00016239789539114905, "loss": 5.1238, "step": 551450 }, { "epoch": 3.3254543481144703, "grad_norm": 1.6806235313415527, "learning_rate": 0.00016237761593743225, "loss": 4.9528, "step": 551500 }, { "epoch": 3.3257558398958045, "grad_norm": 1.7068690061569214, "learning_rate": 0.0001623573362559231, "loss": 4.6277, "step": 551550 }, { "epoch": 3.3260573316771387, "grad_norm": 2.4204840660095215, "learning_rate": 0.0001623370563469948, "loss": 5.0199, "step": 551600 }, { "epoch": 3.3263588234584724, "grad_norm": 1.9188405275344849, "learning_rate": 0.00016231677621102058, "loss": 4.1444, "step": 551650 }, { "epoch": 3.3266603152398067, "grad_norm": 2.4394803047180176, "learning_rate": 0.00016229649584837373, "loss": 4.9971, "step": 551700 }, { "epoch": 3.3269618070211404, "grad_norm": 1.4314109086990356, "learning_rate": 0.00016227621525942735, "loss": 5.0497, "step": 551750 }, { "epoch": 3.3272632988024746, "grad_norm": 1.5686229467391968, "learning_rate": 0.00016225593444455476, "loss": 4.8393, "step": 551800 }, { "epoch": 3.327564790583809, "grad_norm": 1.9970641136169434, "learning_rate": 0.00016223565340412922, "loss": 4.9772, "step": 551850 }, { "epoch": 3.3278662823651426, "grad_norm": 1.9809480905532837, "learning_rate": 0.00016221537213852394, "loss": 4.63, "step": 551900 }, { "epoch": 3.328167774146477, "grad_norm": 1.768239974975586, "learning_rate": 0.00016219509064811213, "loss": 5.132, "step": 551950 }, { "epoch": 3.3284692659278106, "grad_norm": 1.737378478050232, "learning_rate": 0.0001621748089332671, "loss": 4.84, "step": 552000 }, { "epoch": 3.3284692659278106, "eval_loss": 5.331931114196777, "eval_runtime": 39.2495, "eval_samples_per_second": 13.045, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.458529386850328, "step": 552000 }, { "epoch": 3.3287707577091448, "grad_norm": 1.1726561784744263, "learning_rate": 0.0001621545269943621, "loss": 4.9616, "step": 552050 }, { "epoch": 3.329072249490479, "grad_norm": 3.0351734161376953, "learning_rate": 0.00016213424483177034, "loss": 5.0366, "step": 552100 }, { "epoch": 3.3293737412718127, "grad_norm": 2.338571548461914, "learning_rate": 0.00016211396244586514, "loss": 5.339, "step": 552150 }, { "epoch": 3.329675233053147, "grad_norm": 1.7399556636810303, "learning_rate": 0.00016209367983701975, "loss": 4.9052, "step": 552200 }, { "epoch": 3.329976724834481, "grad_norm": 1.7647799253463745, "learning_rate": 0.0001620733970056074, "loss": 5.0786, "step": 552250 }, { "epoch": 3.330278216615815, "grad_norm": 1.6933900117874146, "learning_rate": 0.00016205311395200148, "loss": 4.8988, "step": 552300 }, { "epoch": 3.330579708397149, "grad_norm": 1.735834002494812, "learning_rate": 0.0001620328306765752, "loss": 5.0007, "step": 552350 }, { "epoch": 3.3308812001784833, "grad_norm": 1.6770925521850586, "learning_rate": 0.00016201254717970176, "loss": 4.6127, "step": 552400 }, { "epoch": 3.331182691959817, "grad_norm": 2.5333516597747803, "learning_rate": 0.0001619922634617546, "loss": 4.8231, "step": 552450 }, { "epoch": 3.3314841837411513, "grad_norm": 3.1098623275756836, "learning_rate": 0.00016197197952310692, "loss": 4.5463, "step": 552500 }, { "epoch": 3.331785675522485, "grad_norm": 3.44054913520813, "learning_rate": 0.00016195169536413205, "loss": 5.0514, "step": 552550 }, { "epoch": 3.3320871673038193, "grad_norm": 2.5035107135772705, "learning_rate": 0.00016193141098520329, "loss": 5.3379, "step": 552600 }, { "epoch": 3.3323886590851535, "grad_norm": 2.3601202964782715, "learning_rate": 0.00016191112638669393, "loss": 4.6578, "step": 552650 }, { "epoch": 3.3326901508664872, "grad_norm": 1.8101303577423096, "learning_rate": 0.00016189084156897728, "loss": 5.208, "step": 552700 }, { "epoch": 3.3329916426478214, "grad_norm": 1.333424687385559, "learning_rate": 0.00016187055653242666, "loss": 4.9828, "step": 552750 }, { "epoch": 3.3332931344291556, "grad_norm": 1.979864239692688, "learning_rate": 0.00016185027127741538, "loss": 4.8716, "step": 552800 }, { "epoch": 3.3335946262104894, "grad_norm": 1.0823038816452026, "learning_rate": 0.00016182998580431678, "loss": 5.0734, "step": 552850 }, { "epoch": 3.3338961179918236, "grad_norm": 1.9698185920715332, "learning_rate": 0.0001618097001135042, "loss": 5.1587, "step": 552900 }, { "epoch": 3.334197609773158, "grad_norm": 1.5822430849075317, "learning_rate": 0.0001617894142053509, "loss": 4.7623, "step": 552950 }, { "epoch": 3.3344991015544916, "grad_norm": 1.2475533485412598, "learning_rate": 0.00016176912808023026, "loss": 5.4366, "step": 553000 }, { "epoch": 3.334800593335826, "grad_norm": 2.634014368057251, "learning_rate": 0.00016174884173851564, "loss": 4.858, "step": 553050 }, { "epoch": 3.3351020851171596, "grad_norm": 1.7146563529968262, "learning_rate": 0.00016172855518058036, "loss": 4.7396, "step": 553100 }, { "epoch": 3.3354035768984938, "grad_norm": 1.732162356376648, "learning_rate": 0.00016170826840679773, "loss": 5.312, "step": 553150 }, { "epoch": 3.335705068679828, "grad_norm": 2.163447856903076, "learning_rate": 0.00016168798141754113, "loss": 5.1181, "step": 553200 }, { "epoch": 3.3360065604611617, "grad_norm": 1.648613452911377, "learning_rate": 0.0001616676942131839, "loss": 4.5781, "step": 553250 }, { "epoch": 3.336308052242496, "grad_norm": 2.279747486114502, "learning_rate": 0.00016164740679409944, "loss": 5.0233, "step": 553300 }, { "epoch": 3.3366095440238297, "grad_norm": 0.9926100969314575, "learning_rate": 0.0001616271191606611, "loss": 4.8448, "step": 553350 }, { "epoch": 3.336911035805164, "grad_norm": 2.5392560958862305, "learning_rate": 0.00016160683131324216, "loss": 5.0484, "step": 553400 }, { "epoch": 3.337212527586498, "grad_norm": 1.5356245040893555, "learning_rate": 0.0001615865432522161, "loss": 4.6993, "step": 553450 }, { "epoch": 3.337514019367832, "grad_norm": 1.8331726789474487, "learning_rate": 0.00016156625497795626, "loss": 5.126, "step": 553500 }, { "epoch": 3.337815511149166, "grad_norm": 3.1616806983947754, "learning_rate": 0.00016154596649083593, "loss": 4.8653, "step": 553550 }, { "epoch": 3.3381170029305003, "grad_norm": 2.867495536804199, "learning_rate": 0.00016152567779122862, "loss": 4.7848, "step": 553600 }, { "epoch": 3.338418494711834, "grad_norm": 1.8162647485733032, "learning_rate": 0.00016150538887950764, "loss": 4.7698, "step": 553650 }, { "epoch": 3.3387199864931683, "grad_norm": 1.7680675983428955, "learning_rate": 0.00016148509975604644, "loss": 4.7102, "step": 553700 }, { "epoch": 3.3390214782745025, "grad_norm": 1.3448355197906494, "learning_rate": 0.0001614648104212183, "loss": 5.0232, "step": 553750 }, { "epoch": 3.3393229700558362, "grad_norm": 1.783835768699646, "learning_rate": 0.00016144452087539672, "loss": 5.0579, "step": 553800 }, { "epoch": 3.3396244618371704, "grad_norm": 2.0145621299743652, "learning_rate": 0.00016142423111895505, "loss": 4.8063, "step": 553850 }, { "epoch": 3.339925953618504, "grad_norm": 1.9891655445098877, "learning_rate": 0.00016140394115226676, "loss": 4.9678, "step": 553900 }, { "epoch": 3.3402274453998384, "grad_norm": 1.8931093215942383, "learning_rate": 0.0001613836509757052, "loss": 5.0011, "step": 553950 }, { "epoch": 3.3405289371811726, "grad_norm": 1.0807828903198242, "learning_rate": 0.00016136336058964374, "loss": 4.9824, "step": 554000 }, { "epoch": 3.3408304289625064, "grad_norm": 2.1967971324920654, "learning_rate": 0.00016134306999445586, "loss": 5.0766, "step": 554050 }, { "epoch": 3.3411319207438406, "grad_norm": 2.3634963035583496, "learning_rate": 0.000161322779190515, "loss": 4.6472, "step": 554100 }, { "epoch": 3.3414334125251743, "grad_norm": 1.8693263530731201, "learning_rate": 0.00016130248817819448, "loss": 4.8557, "step": 554150 }, { "epoch": 3.3417349043065085, "grad_norm": 4.332431316375732, "learning_rate": 0.00016128219695786783, "loss": 4.7566, "step": 554200 }, { "epoch": 3.3420363960878428, "grad_norm": 2.264735221862793, "learning_rate": 0.00016126190552990846, "loss": 4.7883, "step": 554250 }, { "epoch": 3.3423378878691765, "grad_norm": 2.2770144939422607, "learning_rate": 0.00016124161389468976, "loss": 5.2509, "step": 554300 }, { "epoch": 3.3426393796505107, "grad_norm": 2.559948682785034, "learning_rate": 0.00016122132205258522, "loss": 5.0209, "step": 554350 }, { "epoch": 3.342940871431845, "grad_norm": 2.4644432067871094, "learning_rate": 0.00016120103000396823, "loss": 5.0629, "step": 554400 }, { "epoch": 3.3432423632131787, "grad_norm": 3.105107545852661, "learning_rate": 0.0001611807377492123, "loss": 4.8365, "step": 554450 }, { "epoch": 3.343543854994513, "grad_norm": 1.3574250936508179, "learning_rate": 0.0001611604452886909, "loss": 4.9533, "step": 554500 }, { "epoch": 3.343845346775847, "grad_norm": 2.617723226547241, "learning_rate": 0.00016114015262277734, "loss": 5.0726, "step": 554550 }, { "epoch": 3.344146838557181, "grad_norm": 2.7944319248199463, "learning_rate": 0.0001611198597518452, "loss": 4.6854, "step": 554600 }, { "epoch": 3.344448330338515, "grad_norm": 2.2876791954040527, "learning_rate": 0.00016109956667626795, "loss": 4.5658, "step": 554650 }, { "epoch": 3.344749822119849, "grad_norm": 1.317716360092163, "learning_rate": 0.00016107927339641896, "loss": 4.9948, "step": 554700 }, { "epoch": 3.345051313901183, "grad_norm": 3.050445795059204, "learning_rate": 0.00016105897991267176, "loss": 4.8278, "step": 554750 }, { "epoch": 3.3453528056825172, "grad_norm": 1.7613506317138672, "learning_rate": 0.00016103868622539987, "loss": 5.0298, "step": 554800 }, { "epoch": 3.345654297463851, "grad_norm": 1.963987112045288, "learning_rate": 0.00016101839233497668, "loss": 4.7022, "step": 554850 }, { "epoch": 3.345955789245185, "grad_norm": 3.1227164268493652, "learning_rate": 0.00016099809824177567, "loss": 5.2532, "step": 554900 }, { "epoch": 3.346257281026519, "grad_norm": 2.584827423095703, "learning_rate": 0.00016097780394617038, "loss": 4.5457, "step": 554950 }, { "epoch": 3.346558772807853, "grad_norm": 1.6864434480667114, "learning_rate": 0.00016095750944853426, "loss": 5.1038, "step": 555000 }, { "epoch": 3.346558772807853, "eval_loss": 5.330315589904785, "eval_runtime": 39.0929, "eval_samples_per_second": 13.097, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.518017591229868, "step": 555000 }, { "epoch": 3.3468602645891874, "grad_norm": 2.8268961906433105, "learning_rate": 0.00016093721474924082, "loss": 5.0379, "step": 555050 }, { "epoch": 3.347161756370521, "grad_norm": 2.6090242862701416, "learning_rate": 0.0001609169198486636, "loss": 5.4096, "step": 555100 }, { "epoch": 3.3474632481518554, "grad_norm": 1.9676148891448975, "learning_rate": 0.00016089662474717597, "loss": 5.001, "step": 555150 }, { "epoch": 3.3477647399331896, "grad_norm": 1.708247423171997, "learning_rate": 0.0001608763294451516, "loss": 4.8225, "step": 555200 }, { "epoch": 3.3480662317145233, "grad_norm": 2.525209665298462, "learning_rate": 0.00016085603394296383, "loss": 4.8335, "step": 555250 }, { "epoch": 3.3483677234958575, "grad_norm": 1.252648115158081, "learning_rate": 0.0001608357382409863, "loss": 4.8229, "step": 555300 }, { "epoch": 3.3486692152771917, "grad_norm": 2.2298097610473633, "learning_rate": 0.00016081544233959244, "loss": 5.1595, "step": 555350 }, { "epoch": 3.3489707070585255, "grad_norm": 1.0709912776947021, "learning_rate": 0.0001607951462391558, "loss": 4.781, "step": 555400 }, { "epoch": 3.3492721988398597, "grad_norm": 2.911271095275879, "learning_rate": 0.0001607748499400499, "loss": 4.9157, "step": 555450 }, { "epoch": 3.3495736906211935, "grad_norm": 2.504171848297119, "learning_rate": 0.0001607545534426483, "loss": 4.9332, "step": 555500 }, { "epoch": 3.3498751824025277, "grad_norm": 1.2624982595443726, "learning_rate": 0.0001607342567473245, "loss": 4.6868, "step": 555550 }, { "epoch": 3.350176674183862, "grad_norm": 2.078730583190918, "learning_rate": 0.00016071395985445193, "loss": 4.8667, "step": 555600 }, { "epoch": 3.3504781659651957, "grad_norm": 2.637934446334839, "learning_rate": 0.00016069366276440429, "loss": 4.8758, "step": 555650 }, { "epoch": 3.35077965774653, "grad_norm": 0.700568675994873, "learning_rate": 0.00016067336547755508, "loss": 4.799, "step": 555700 }, { "epoch": 3.3510811495278636, "grad_norm": 1.6720647811889648, "learning_rate": 0.0001606530679942777, "loss": 4.5734, "step": 555750 }, { "epoch": 3.351382641309198, "grad_norm": 1.822680115699768, "learning_rate": 0.00016063277031494588, "loss": 4.899, "step": 555800 }, { "epoch": 3.351684133090532, "grad_norm": 0.9686749577522278, "learning_rate": 0.0001606124724399331, "loss": 4.5927, "step": 555850 }, { "epoch": 3.351985624871866, "grad_norm": 2.4779598712921143, "learning_rate": 0.00016059217436961292, "loss": 4.9169, "step": 555900 }, { "epoch": 3.3522871166532, "grad_norm": 1.7107378244400024, "learning_rate": 0.00016057187610435885, "loss": 4.6047, "step": 555950 }, { "epoch": 3.352588608434534, "grad_norm": 1.7810672521591187, "learning_rate": 0.0001605515776445445, "loss": 4.8844, "step": 556000 }, { "epoch": 3.352890100215868, "grad_norm": 1.704222559928894, "learning_rate": 0.0001605312789905434, "loss": 4.4608, "step": 556050 }, { "epoch": 3.353191591997202, "grad_norm": 3.6705832481384277, "learning_rate": 0.00016051098014272918, "loss": 5.1028, "step": 556100 }, { "epoch": 3.3534930837785364, "grad_norm": 1.6268937587738037, "learning_rate": 0.00016049068110147533, "loss": 5.1832, "step": 556150 }, { "epoch": 3.35379457555987, "grad_norm": 3.236055850982666, "learning_rate": 0.0001604703818671555, "loss": 4.8389, "step": 556200 }, { "epoch": 3.3540960673412044, "grad_norm": 2.581437826156616, "learning_rate": 0.00016045008244014322, "loss": 4.9451, "step": 556250 }, { "epoch": 3.354397559122538, "grad_norm": 1.7273577451705933, "learning_rate": 0.00016042978282081207, "loss": 5.1605, "step": 556300 }, { "epoch": 3.3546990509038723, "grad_norm": 2.024618625640869, "learning_rate": 0.00016040948300953566, "loss": 4.8234, "step": 556350 }, { "epoch": 3.3550005426852065, "grad_norm": 3.0048210620880127, "learning_rate": 0.00016038918300668762, "loss": 5.0475, "step": 556400 }, { "epoch": 3.3553020344665403, "grad_norm": 1.8046252727508545, "learning_rate": 0.0001603688828126414, "loss": 4.6644, "step": 556450 }, { "epoch": 3.3556035262478745, "grad_norm": 2.982358455657959, "learning_rate": 0.00016034858242777077, "loss": 4.6873, "step": 556500 }, { "epoch": 3.3559050180292087, "grad_norm": 1.9930771589279175, "learning_rate": 0.0001603282818524492, "loss": 4.9628, "step": 556550 }, { "epoch": 3.3562065098105425, "grad_norm": 2.8080546855926514, "learning_rate": 0.00016030798108705033, "loss": 5.0806, "step": 556600 }, { "epoch": 3.3565080015918767, "grad_norm": 2.1250040531158447, "learning_rate": 0.00016028768013194784, "loss": 4.9395, "step": 556650 }, { "epoch": 3.356809493373211, "grad_norm": 2.426421880722046, "learning_rate": 0.00016026737898751527, "loss": 4.4108, "step": 556700 }, { "epoch": 3.3571109851545446, "grad_norm": 2.520437717437744, "learning_rate": 0.00016024707765412616, "loss": 5.0666, "step": 556750 }, { "epoch": 3.357412476935879, "grad_norm": 1.852838397026062, "learning_rate": 0.00016022677613215428, "loss": 4.8856, "step": 556800 }, { "epoch": 3.3577139687172126, "grad_norm": 1.2385201454162598, "learning_rate": 0.00016020647442197317, "loss": 4.6942, "step": 556850 }, { "epoch": 3.358015460498547, "grad_norm": 0.9090608954429626, "learning_rate": 0.00016018617252395648, "loss": 4.2671, "step": 556900 }, { "epoch": 3.358316952279881, "grad_norm": 1.8367096185684204, "learning_rate": 0.0001601658704384778, "loss": 4.9487, "step": 556950 }, { "epoch": 3.358618444061215, "grad_norm": 1.754015564918518, "learning_rate": 0.0001601455681659108, "loss": 4.8593, "step": 557000 }, { "epoch": 3.358919935842549, "grad_norm": 4.752604007720947, "learning_rate": 0.00016012526570662906, "loss": 4.7002, "step": 557050 }, { "epoch": 3.3592214276238828, "grad_norm": 0.6948525905609131, "learning_rate": 0.00016010496306100627, "loss": 4.8597, "step": 557100 }, { "epoch": 3.359522919405217, "grad_norm": 1.5646406412124634, "learning_rate": 0.00016008466022941607, "loss": 4.9705, "step": 557150 }, { "epoch": 3.359824411186551, "grad_norm": 2.8325889110565186, "learning_rate": 0.00016006435721223206, "loss": 4.7547, "step": 557200 }, { "epoch": 3.360125902967885, "grad_norm": 2.722816228866577, "learning_rate": 0.00016004405400982794, "loss": 4.7081, "step": 557250 }, { "epoch": 3.360427394749219, "grad_norm": 2.6607508659362793, "learning_rate": 0.00016002375062257736, "loss": 5.161, "step": 557300 }, { "epoch": 3.3607288865305533, "grad_norm": 1.8543909788131714, "learning_rate": 0.00016000344705085392, "loss": 4.9695, "step": 557350 }, { "epoch": 3.361030378311887, "grad_norm": 2.0888166427612305, "learning_rate": 0.00015998314329503132, "loss": 4.6555, "step": 557400 }, { "epoch": 3.3613318700932213, "grad_norm": 0.6860942840576172, "learning_rate": 0.0001599628393554832, "loss": 4.7191, "step": 557450 }, { "epoch": 3.3616333618745555, "grad_norm": 2.3764607906341553, "learning_rate": 0.0001599425352325833, "loss": 4.5051, "step": 557500 }, { "epoch": 3.3619348536558893, "grad_norm": 1.6912153959274292, "learning_rate": 0.00015992223092670517, "loss": 4.7457, "step": 557550 }, { "epoch": 3.3622363454372235, "grad_norm": 2.2447352409362793, "learning_rate": 0.00015990192643822258, "loss": 4.7537, "step": 557600 }, { "epoch": 3.3625378372185573, "grad_norm": 2.303774356842041, "learning_rate": 0.00015988162176750913, "loss": 5.1981, "step": 557650 }, { "epoch": 3.3628393289998915, "grad_norm": 1.5439693927764893, "learning_rate": 0.00015986131691493854, "loss": 4.5356, "step": 557700 }, { "epoch": 3.3631408207812257, "grad_norm": 0.7737404108047485, "learning_rate": 0.00015984101188088446, "loss": 5.1746, "step": 557750 }, { "epoch": 3.3634423125625594, "grad_norm": 1.9297325611114502, "learning_rate": 0.00015982070666572065, "loss": 4.2735, "step": 557800 }, { "epoch": 3.3637438043438936, "grad_norm": 1.2251967191696167, "learning_rate": 0.0001598004012698207, "loss": 4.821, "step": 557850 }, { "epoch": 3.3640452961252274, "grad_norm": 1.5314691066741943, "learning_rate": 0.00015978009569355837, "loss": 4.9537, "step": 557900 }, { "epoch": 3.3643467879065616, "grad_norm": 1.85564386844635, "learning_rate": 0.00015975978993730737, "loss": 4.7404, "step": 557950 }, { "epoch": 3.364648279687896, "grad_norm": 1.4724818468093872, "learning_rate": 0.00015973948400144133, "loss": 4.7858, "step": 558000 }, { "epoch": 3.364648279687896, "eval_loss": 5.311558723449707, "eval_runtime": 39.0242, "eval_samples_per_second": 13.12, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.438448213731566, "step": 558000 }, { "epoch": 3.3649497714692296, "grad_norm": 2.7303173542022705, "learning_rate": 0.000159719177886334, "loss": 4.9357, "step": 558050 }, { "epoch": 3.365251263250564, "grad_norm": 4.7040114402771, "learning_rate": 0.00015969887159235905, "loss": 5.0392, "step": 558100 }, { "epoch": 3.365552755031898, "grad_norm": 0.9803646206855774, "learning_rate": 0.00015967856511989027, "loss": 4.9515, "step": 558150 }, { "epoch": 3.3658542468132318, "grad_norm": 2.014322280883789, "learning_rate": 0.00015965825846930121, "loss": 4.9196, "step": 558200 }, { "epoch": 3.366155738594566, "grad_norm": 1.873963475227356, "learning_rate": 0.00015963795164096573, "loss": 5.2495, "step": 558250 }, { "epoch": 3.3664572303759, "grad_norm": 0.6353139281272888, "learning_rate": 0.00015961764463525758, "loss": 5.0796, "step": 558300 }, { "epoch": 3.366758722157234, "grad_norm": 0.5185523629188538, "learning_rate": 0.00015959733745255032, "loss": 4.9873, "step": 558350 }, { "epoch": 3.367060213938568, "grad_norm": 2.5161051750183105, "learning_rate": 0.00015957703009321776, "loss": 5.0989, "step": 558400 }, { "epoch": 3.367361705719902, "grad_norm": 2.350782632827759, "learning_rate": 0.00015955672255763367, "loss": 5.082, "step": 558450 }, { "epoch": 3.367663197501236, "grad_norm": 1.7669780254364014, "learning_rate": 0.00015953641484617174, "loss": 5.0338, "step": 558500 }, { "epoch": 3.3679646892825703, "grad_norm": 1.6372342109680176, "learning_rate": 0.00015951610695920566, "loss": 4.7363, "step": 558550 }, { "epoch": 3.368266181063904, "grad_norm": 0.6776971220970154, "learning_rate": 0.00015949579889710925, "loss": 4.7604, "step": 558600 }, { "epoch": 3.3685676728452383, "grad_norm": 2.6202497482299805, "learning_rate": 0.0001594754906602562, "loss": 4.7686, "step": 558650 }, { "epoch": 3.368869164626572, "grad_norm": 1.938941478729248, "learning_rate": 0.00015945518224902026, "loss": 4.4853, "step": 558700 }, { "epoch": 3.3691706564079063, "grad_norm": 1.97002112865448, "learning_rate": 0.00015943487366377516, "loss": 4.7671, "step": 558750 }, { "epoch": 3.3694721481892405, "grad_norm": 2.8635008335113525, "learning_rate": 0.0001594145649048947, "loss": 4.4241, "step": 558800 }, { "epoch": 3.369773639970574, "grad_norm": 2.8230783939361572, "learning_rate": 0.0001593942559727526, "loss": 4.5661, "step": 558850 }, { "epoch": 3.3700751317519084, "grad_norm": 2.1662726402282715, "learning_rate": 0.00015937394686772266, "loss": 4.8812, "step": 558900 }, { "epoch": 3.3703766235332426, "grad_norm": 2.5959296226501465, "learning_rate": 0.00015935363759017856, "loss": 5.0902, "step": 558950 }, { "epoch": 3.3706781153145764, "grad_norm": 2.6035642623901367, "learning_rate": 0.00015933332814049416, "loss": 4.7123, "step": 559000 }, { "epoch": 3.3709796070959106, "grad_norm": 2.5535037517547607, "learning_rate": 0.00015931301851904314, "loss": 4.9427, "step": 559050 }, { "epoch": 3.371281098877245, "grad_norm": 1.564947247505188, "learning_rate": 0.00015929270872619932, "loss": 4.6997, "step": 559100 }, { "epoch": 3.3715825906585786, "grad_norm": 3.7059826850891113, "learning_rate": 0.00015927239876233643, "loss": 4.5665, "step": 559150 }, { "epoch": 3.3718840824399128, "grad_norm": 1.3811452388763428, "learning_rate": 0.0001592520886278283, "loss": 4.8185, "step": 559200 }, { "epoch": 3.3721855742212465, "grad_norm": 1.6490145921707153, "learning_rate": 0.00015923177832304866, "loss": 5.0829, "step": 559250 }, { "epoch": 3.3724870660025807, "grad_norm": 3.1205315589904785, "learning_rate": 0.00015921146784837132, "loss": 4.8666, "step": 559300 }, { "epoch": 3.372788557783915, "grad_norm": 3.3538293838500977, "learning_rate": 0.0001591911572041701, "loss": 5.2096, "step": 559350 }, { "epoch": 3.3730900495652487, "grad_norm": 2.0732691287994385, "learning_rate": 0.00015917084639081866, "loss": 4.6855, "step": 559400 }, { "epoch": 3.373391541346583, "grad_norm": 2.5640830993652344, "learning_rate": 0.00015915053540869095, "loss": 4.7771, "step": 559450 }, { "epoch": 3.3736930331279167, "grad_norm": 1.5709184408187866, "learning_rate": 0.00015913022425816068, "loss": 4.6198, "step": 559500 }, { "epoch": 3.373994524909251, "grad_norm": 1.4086692333221436, "learning_rate": 0.00015910991293960165, "loss": 4.5306, "step": 559550 }, { "epoch": 3.374296016690585, "grad_norm": 0.6508447527885437, "learning_rate": 0.0001590896014533877, "loss": 5.0239, "step": 559600 }, { "epoch": 3.3745975084719193, "grad_norm": 2.2110788822174072, "learning_rate": 0.0001590692897998926, "loss": 4.651, "step": 559650 }, { "epoch": 3.374899000253253, "grad_norm": 2.2103796005249023, "learning_rate": 0.00015904897797949013, "loss": 5.0766, "step": 559700 }, { "epoch": 3.3752004920345873, "grad_norm": 1.8577522039413452, "learning_rate": 0.00015902866599255418, "loss": 4.7014, "step": 559750 }, { "epoch": 3.375501983815921, "grad_norm": 3.752274751663208, "learning_rate": 0.00015900835383945845, "loss": 4.8741, "step": 559800 }, { "epoch": 3.3758034755972552, "grad_norm": 2.3358771800994873, "learning_rate": 0.00015898804152057692, "loss": 5.0594, "step": 559850 }, { "epoch": 3.3761049673785894, "grad_norm": 2.6981658935546875, "learning_rate": 0.00015896772903628326, "loss": 5.0771, "step": 559900 }, { "epoch": 3.376406459159923, "grad_norm": 2.3411929607391357, "learning_rate": 0.0001589474163869513, "loss": 4.8113, "step": 559950 }, { "epoch": 3.3767079509412574, "grad_norm": 1.887070894241333, "learning_rate": 0.00015892710357295496, "loss": 4.8805, "step": 560000 }, { "epoch": 3.377009442722591, "grad_norm": 1.986985206604004, "learning_rate": 0.00015890679059466802, "loss": 4.7695, "step": 560050 }, { "epoch": 3.3773109345039254, "grad_norm": 1.5097752809524536, "learning_rate": 0.00015888647745246432, "loss": 4.9566, "step": 560100 }, { "epoch": 3.3776124262852596, "grad_norm": 2.1181833744049072, "learning_rate": 0.00015886616414671767, "loss": 4.7239, "step": 560150 }, { "epoch": 3.3779139180665934, "grad_norm": 0.6986890435218811, "learning_rate": 0.00015884585067780192, "loss": 4.543, "step": 560200 }, { "epoch": 3.3782154098479276, "grad_norm": 2.214305877685547, "learning_rate": 0.00015882553704609092, "loss": 5.0856, "step": 560250 }, { "epoch": 3.3785169016292618, "grad_norm": 1.7315748929977417, "learning_rate": 0.00015880522325195846, "loss": 5.0621, "step": 560300 }, { "epoch": 3.3788183934105955, "grad_norm": 2.280644178390503, "learning_rate": 0.0001587849092957785, "loss": 5.073, "step": 560350 }, { "epoch": 3.3791198851919297, "grad_norm": 2.906066656112671, "learning_rate": 0.00015876459517792475, "loss": 4.6048, "step": 560400 }, { "epoch": 3.379421376973264, "grad_norm": 2.2327871322631836, "learning_rate": 0.0001587442808987712, "loss": 4.8798, "step": 560450 }, { "epoch": 3.3797228687545977, "grad_norm": 2.8066928386688232, "learning_rate": 0.00015872396645869163, "loss": 5.0785, "step": 560500 }, { "epoch": 3.380024360535932, "grad_norm": 1.7879031896591187, "learning_rate": 0.00015870365185805985, "loss": 4.9002, "step": 560550 }, { "epoch": 3.3803258523172657, "grad_norm": 2.8987555503845215, "learning_rate": 0.00015868333709724984, "loss": 5.2257, "step": 560600 }, { "epoch": 3.3806273440986, "grad_norm": 1.560755729675293, "learning_rate": 0.0001586630221766354, "loss": 4.8803, "step": 560650 }, { "epoch": 3.380928835879934, "grad_norm": 2.2796096801757812, "learning_rate": 0.00015864270709659036, "loss": 4.4968, "step": 560700 }, { "epoch": 3.381230327661268, "grad_norm": 1.8156447410583496, "learning_rate": 0.00015862239185748865, "loss": 4.6849, "step": 560750 }, { "epoch": 3.381531819442602, "grad_norm": 1.86711585521698, "learning_rate": 0.0001586020764597041, "loss": 4.8067, "step": 560800 }, { "epoch": 3.381833311223936, "grad_norm": 2.3853163719177246, "learning_rate": 0.00015858176090361063, "loss": 4.9385, "step": 560850 }, { "epoch": 3.38213480300527, "grad_norm": 3.165254592895508, "learning_rate": 0.00015856144518958213, "loss": 4.9922, "step": 560900 }, { "epoch": 3.3824362947866042, "grad_norm": 1.078372836112976, "learning_rate": 0.0001585411293179924, "loss": 4.6977, "step": 560950 }, { "epoch": 3.382737786567938, "grad_norm": 2.235837459564209, "learning_rate": 0.00015852081328921535, "loss": 5.1998, "step": 561000 }, { "epoch": 3.382737786567938, "eval_loss": 5.313165664672852, "eval_runtime": 39.1937, "eval_samples_per_second": 13.063, "eval_steps_per_second": 6.532, "eval_tts_loss": 7.576284133502027, "step": 561000 }, { "epoch": 3.383039278349272, "grad_norm": 1.7412481307983398, "learning_rate": 0.00015850049710362495, "loss": 5.1388, "step": 561050 }, { "epoch": 3.3833407701306064, "grad_norm": 1.7652288675308228, "learning_rate": 0.00015848018076159505, "loss": 5.1472, "step": 561100 }, { "epoch": 3.38364226191194, "grad_norm": 2.055445671081543, "learning_rate": 0.00015845986426349945, "loss": 4.6995, "step": 561150 }, { "epoch": 3.3839437536932744, "grad_norm": 2.1738972663879395, "learning_rate": 0.00015843954760971215, "loss": 5.0207, "step": 561200 }, { "epoch": 3.3842452454746086, "grad_norm": 3.242100477218628, "learning_rate": 0.00015841923080060704, "loss": 4.7816, "step": 561250 }, { "epoch": 3.3845467372559424, "grad_norm": 3.0920305252075195, "learning_rate": 0.00015839891383655802, "loss": 4.8417, "step": 561300 }, { "epoch": 3.3848482290372766, "grad_norm": 2.591970682144165, "learning_rate": 0.00015837859671793896, "loss": 4.3766, "step": 561350 }, { "epoch": 3.3851497208186103, "grad_norm": 0.758842945098877, "learning_rate": 0.00015835827944512372, "loss": 5.0451, "step": 561400 }, { "epoch": 3.3854512125999445, "grad_norm": 2.529629707336426, "learning_rate": 0.00015833796201848634, "loss": 4.9367, "step": 561450 }, { "epoch": 3.3857527043812787, "grad_norm": 1.8142120838165283, "learning_rate": 0.0001583176444384007, "loss": 4.1015, "step": 561500 }, { "epoch": 3.3860541961626125, "grad_norm": 1.463089108467102, "learning_rate": 0.00015829732670524058, "loss": 4.5535, "step": 561550 }, { "epoch": 3.3863556879439467, "grad_norm": 2.867548942565918, "learning_rate": 0.00015827700881938008, "loss": 4.8374, "step": 561600 }, { "epoch": 3.3866571797252805, "grad_norm": 1.8569830656051636, "learning_rate": 0.00015825669078119303, "loss": 4.6614, "step": 561650 }, { "epoch": 3.3869586715066147, "grad_norm": 1.9945868253707886, "learning_rate": 0.00015823637259105335, "loss": 5.1961, "step": 561700 }, { "epoch": 3.387260163287949, "grad_norm": 2.4886386394500732, "learning_rate": 0.000158216054249335, "loss": 5.3324, "step": 561750 }, { "epoch": 3.3875616550692826, "grad_norm": 1.4925537109375, "learning_rate": 0.0001581957357564119, "loss": 4.6329, "step": 561800 }, { "epoch": 3.387863146850617, "grad_norm": 1.4899966716766357, "learning_rate": 0.000158175417112658, "loss": 4.9955, "step": 561850 }, { "epoch": 3.388164638631951, "grad_norm": 3.8176379203796387, "learning_rate": 0.0001581550983184472, "loss": 4.807, "step": 561900 }, { "epoch": 3.388466130413285, "grad_norm": 2.2956008911132812, "learning_rate": 0.00015813477937415342, "loss": 4.8148, "step": 561950 }, { "epoch": 3.388767622194619, "grad_norm": 2.320122003555298, "learning_rate": 0.00015811446028015063, "loss": 5.3687, "step": 562000 }, { "epoch": 3.3890691139759532, "grad_norm": 1.8310045003890991, "learning_rate": 0.00015809414103681288, "loss": 4.7965, "step": 562050 }, { "epoch": 3.389370605757287, "grad_norm": 1.4070217609405518, "learning_rate": 0.00015807382164451393, "loss": 4.8865, "step": 562100 }, { "epoch": 3.389672097538621, "grad_norm": 2.4633142948150635, "learning_rate": 0.00015805350210362779, "loss": 5.1596, "step": 562150 }, { "epoch": 3.389973589319955, "grad_norm": 2.335397243499756, "learning_rate": 0.0001580331824145285, "loss": 5.0576, "step": 562200 }, { "epoch": 3.390275081101289, "grad_norm": 0.9123090505599976, "learning_rate": 0.00015801286257758994, "loss": 5.1502, "step": 562250 }, { "epoch": 3.3905765728826234, "grad_norm": 2.552335500717163, "learning_rate": 0.00015799254259318607, "loss": 4.9506, "step": 562300 }, { "epoch": 3.390878064663957, "grad_norm": 4.728635311126709, "learning_rate": 0.00015797222246169084, "loss": 4.9532, "step": 562350 }, { "epoch": 3.3911795564452913, "grad_norm": 1.8804550170898438, "learning_rate": 0.00015795190218347824, "loss": 5.0619, "step": 562400 }, { "epoch": 3.391481048226625, "grad_norm": 2.271597146987915, "learning_rate": 0.00015793158175892225, "loss": 4.7559, "step": 562450 }, { "epoch": 3.3917825400079593, "grad_norm": 2.2615814208984375, "learning_rate": 0.0001579112611883968, "loss": 5.4169, "step": 562500 }, { "epoch": 3.3920840317892935, "grad_norm": 2.5876567363739014, "learning_rate": 0.0001578909404722759, "loss": 5.0067, "step": 562550 }, { "epoch": 3.3923855235706273, "grad_norm": 2.228102207183838, "learning_rate": 0.00015787061961093345, "loss": 5.039, "step": 562600 }, { "epoch": 3.3926870153519615, "grad_norm": 2.6599082946777344, "learning_rate": 0.00015785029860474354, "loss": 4.7355, "step": 562650 }, { "epoch": 3.3929885071332957, "grad_norm": 1.808618426322937, "learning_rate": 0.00015782997745408006, "loss": 5.0137, "step": 562700 }, { "epoch": 3.3932899989146295, "grad_norm": 1.9924615621566772, "learning_rate": 0.000157809656159317, "loss": 4.6072, "step": 562750 }, { "epoch": 3.3935914906959637, "grad_norm": 2.1549911499023438, "learning_rate": 0.00015778933472082837, "loss": 5.0765, "step": 562800 }, { "epoch": 3.393892982477298, "grad_norm": 1.294675588607788, "learning_rate": 0.00015776901313898817, "loss": 4.8895, "step": 562850 }, { "epoch": 3.3941944742586316, "grad_norm": 2.0182929039001465, "learning_rate": 0.00015774869141417036, "loss": 4.5666, "step": 562900 }, { "epoch": 3.394495966039966, "grad_norm": 2.7636897563934326, "learning_rate": 0.00015772836954674896, "loss": 4.5745, "step": 562950 }, { "epoch": 3.3947974578212996, "grad_norm": 2.46773624420166, "learning_rate": 0.00015770804753709788, "loss": 4.8377, "step": 563000 }, { "epoch": 3.395098949602634, "grad_norm": 1.9544411897659302, "learning_rate": 0.00015768772538559127, "loss": 5.0866, "step": 563050 }, { "epoch": 3.395400441383968, "grad_norm": 1.6342272758483887, "learning_rate": 0.000157667403092603, "loss": 4.7707, "step": 563100 }, { "epoch": 3.395701933165302, "grad_norm": 1.119722843170166, "learning_rate": 0.00015764708065850713, "loss": 4.9778, "step": 563150 }, { "epoch": 3.396003424946636, "grad_norm": 2.9786014556884766, "learning_rate": 0.00015762675808367766, "loss": 5.2806, "step": 563200 }, { "epoch": 3.3963049167279697, "grad_norm": 2.6410884857177734, "learning_rate": 0.00015760643536848857, "loss": 4.8725, "step": 563250 }, { "epoch": 3.396606408509304, "grad_norm": 1.8300658464431763, "learning_rate": 0.0001575861125133139, "loss": 4.8259, "step": 563300 }, { "epoch": 3.396907900290638, "grad_norm": 2.0791590213775635, "learning_rate": 0.00015756578951852768, "loss": 4.5826, "step": 563350 }, { "epoch": 3.3972093920719724, "grad_norm": 1.6733765602111816, "learning_rate": 0.00015754546638450392, "loss": 4.9471, "step": 563400 }, { "epoch": 3.397510883853306, "grad_norm": 1.6301554441452026, "learning_rate": 0.00015752514311161655, "loss": 4.5956, "step": 563450 }, { "epoch": 3.3978123756346403, "grad_norm": 3.242464542388916, "learning_rate": 0.0001575048197002397, "loss": 4.7762, "step": 563500 }, { "epoch": 3.398113867415974, "grad_norm": 1.860192060470581, "learning_rate": 0.00015748449615074733, "loss": 4.6336, "step": 563550 }, { "epoch": 3.3984153591973083, "grad_norm": 2.0639426708221436, "learning_rate": 0.0001574641724635135, "loss": 4.4407, "step": 563600 }, { "epoch": 3.3987168509786425, "grad_norm": 2.710926055908203, "learning_rate": 0.00015744384863891227, "loss": 4.9886, "step": 563650 }, { "epoch": 3.3990183427599763, "grad_norm": 1.7293784618377686, "learning_rate": 0.00015742352467731758, "loss": 4.5703, "step": 563700 }, { "epoch": 3.3993198345413105, "grad_norm": 1.5647839307785034, "learning_rate": 0.0001574032005791035, "loss": 4.8835, "step": 563750 }, { "epoch": 3.3996213263226442, "grad_norm": 2.581260919570923, "learning_rate": 0.00015738287634464411, "loss": 5.0081, "step": 563800 }, { "epoch": 3.3999228181039785, "grad_norm": 2.4133999347686768, "learning_rate": 0.0001573625519743134, "loss": 4.5085, "step": 563850 }, { "epoch": 3.4002243098853127, "grad_norm": 2.2672181129455566, "learning_rate": 0.00015734222746848544, "loss": 4.8517, "step": 563900 }, { "epoch": 3.4005258016666464, "grad_norm": 2.927800178527832, "learning_rate": 0.00015732190282753426, "loss": 5.3837, "step": 563950 }, { "epoch": 3.4008272934479806, "grad_norm": 2.748206377029419, "learning_rate": 0.00015730157805183393, "loss": 4.7332, "step": 564000 }, { "epoch": 3.4008272934479806, "eval_loss": 5.318140029907227, "eval_runtime": 39.125, "eval_samples_per_second": 13.086, "eval_steps_per_second": 6.543, "eval_tts_loss": 7.53264911548952, "step": 564000 }, { "epoch": 3.401128785229315, "grad_norm": 2.7634286880493164, "learning_rate": 0.00015728125314175845, "loss": 4.3833, "step": 564050 }, { "epoch": 3.4014302770106486, "grad_norm": 1.8067153692245483, "learning_rate": 0.0001572609280976819, "loss": 4.9101, "step": 564100 }, { "epoch": 3.401731768791983, "grad_norm": 2.928004503250122, "learning_rate": 0.00015724060291997833, "loss": 4.3599, "step": 564150 }, { "epoch": 3.402033260573317, "grad_norm": 1.5280933380126953, "learning_rate": 0.00015722027760902175, "loss": 4.8125, "step": 564200 }, { "epoch": 3.4023347523546508, "grad_norm": 1.3475213050842285, "learning_rate": 0.00015719995216518636, "loss": 5.0464, "step": 564250 }, { "epoch": 3.402636244135985, "grad_norm": 1.9961743354797363, "learning_rate": 0.00015717962658884607, "loss": 5.0225, "step": 564300 }, { "epoch": 3.4029377359173187, "grad_norm": 2.16968035697937, "learning_rate": 0.00015715930088037498, "loss": 4.9789, "step": 564350 }, { "epoch": 3.403239227698653, "grad_norm": 1.807923436164856, "learning_rate": 0.00015713897504014717, "loss": 5.0885, "step": 564400 }, { "epoch": 3.403540719479987, "grad_norm": 1.959298014640808, "learning_rate": 0.00015711864906853675, "loss": 4.8738, "step": 564450 }, { "epoch": 3.403842211261321, "grad_norm": 2.916283369064331, "learning_rate": 0.00015709832296591773, "loss": 5.0365, "step": 564500 }, { "epoch": 3.404143703042655, "grad_norm": 2.2122819423675537, "learning_rate": 0.00015707799673266422, "loss": 4.697, "step": 564550 }, { "epoch": 3.404445194823989, "grad_norm": 2.487753391265869, "learning_rate": 0.00015705767036915027, "loss": 4.7879, "step": 564600 }, { "epoch": 3.404746686605323, "grad_norm": 3.066136598587036, "learning_rate": 0.00015703734387574999, "loss": 4.2776, "step": 564650 }, { "epoch": 3.4050481783866573, "grad_norm": 1.6482497453689575, "learning_rate": 0.00015701701725283744, "loss": 4.6001, "step": 564700 }, { "epoch": 3.405349670167991, "grad_norm": 1.9646810293197632, "learning_rate": 0.00015699669050078666, "loss": 4.8979, "step": 564750 }, { "epoch": 3.4056511619493253, "grad_norm": 2.8364474773406982, "learning_rate": 0.00015697636361997183, "loss": 5.2867, "step": 564800 }, { "epoch": 3.4059526537306595, "grad_norm": 1.2382644414901733, "learning_rate": 0.00015695603661076697, "loss": 4.9792, "step": 564850 }, { "epoch": 3.4062541455119932, "grad_norm": 1.7621333599090576, "learning_rate": 0.00015693570947354613, "loss": 4.8226, "step": 564900 }, { "epoch": 3.4065556372933274, "grad_norm": 1.636021375656128, "learning_rate": 0.0001569153822086835, "loss": 5.0128, "step": 564950 }, { "epoch": 3.4068571290746616, "grad_norm": 2.1594083309173584, "learning_rate": 0.00015689505481655312, "loss": 4.8897, "step": 565000 }, { "epoch": 3.4071586208559954, "grad_norm": 1.8307687044143677, "learning_rate": 0.00015687472729752913, "loss": 5.1257, "step": 565050 }, { "epoch": 3.4074601126373296, "grad_norm": 2.7524988651275635, "learning_rate": 0.00015685439965198555, "loss": 4.7558, "step": 565100 }, { "epoch": 3.4077616044186634, "grad_norm": 1.5286195278167725, "learning_rate": 0.00015683407188029654, "loss": 4.9877, "step": 565150 }, { "epoch": 3.4080630961999976, "grad_norm": 3.5463688373565674, "learning_rate": 0.00015681374398283617, "loss": 4.8559, "step": 565200 }, { "epoch": 3.408364587981332, "grad_norm": 1.820396065711975, "learning_rate": 0.0001567934159599786, "loss": 5.0584, "step": 565250 }, { "epoch": 3.4086660797626656, "grad_norm": 1.572643518447876, "learning_rate": 0.00015677308781209792, "loss": 5.0588, "step": 565300 }, { "epoch": 3.4089675715439998, "grad_norm": 2.197995662689209, "learning_rate": 0.00015675275953956817, "loss": 4.7361, "step": 565350 }, { "epoch": 3.4092690633253335, "grad_norm": 2.908815622329712, "learning_rate": 0.0001567324311427635, "loss": 5.027, "step": 565400 }, { "epoch": 3.4095705551066677, "grad_norm": 1.9001808166503906, "learning_rate": 0.00015671210262205814, "loss": 5.1012, "step": 565450 }, { "epoch": 3.409872046888002, "grad_norm": 1.7431730031967163, "learning_rate": 0.000156691773977826, "loss": 4.8286, "step": 565500 }, { "epoch": 3.4101735386693357, "grad_norm": 1.9908713102340698, "learning_rate": 0.0001566714452104414, "loss": 4.6181, "step": 565550 }, { "epoch": 3.41047503045067, "grad_norm": 1.694474697113037, "learning_rate": 0.00015665111632027834, "loss": 4.6913, "step": 565600 }, { "epoch": 3.410776522232004, "grad_norm": 2.462656259536743, "learning_rate": 0.000156630787307711, "loss": 5.0049, "step": 565650 }, { "epoch": 3.411078014013338, "grad_norm": 3.984642505645752, "learning_rate": 0.00015661045817311344, "loss": 5.267, "step": 565700 }, { "epoch": 3.411379505794672, "grad_norm": 2.5197978019714355, "learning_rate": 0.00015659012891685984, "loss": 4.8906, "step": 565750 }, { "epoch": 3.4116809975760063, "grad_norm": 1.5922421216964722, "learning_rate": 0.0001565697995393243, "loss": 4.6418, "step": 565800 }, { "epoch": 3.41198248935734, "grad_norm": 2.046872854232788, "learning_rate": 0.00015654947004088105, "loss": 4.9014, "step": 565850 }, { "epoch": 3.4122839811386743, "grad_norm": 1.7143898010253906, "learning_rate": 0.0001565291404219041, "loss": 4.7772, "step": 565900 }, { "epoch": 3.412585472920008, "grad_norm": 1.3442264795303345, "learning_rate": 0.0001565088106827676, "loss": 4.4759, "step": 565950 }, { "epoch": 3.4128869647013422, "grad_norm": 1.8848265409469604, "learning_rate": 0.00015648848082384576, "loss": 4.6379, "step": 566000 }, { "epoch": 3.4131884564826764, "grad_norm": 2.1057043075561523, "learning_rate": 0.0001564681508455127, "loss": 5.0474, "step": 566050 }, { "epoch": 3.41348994826401, "grad_norm": 1.6119272708892822, "learning_rate": 0.00015644782074814256, "loss": 4.6895, "step": 566100 }, { "epoch": 3.4137914400453444, "grad_norm": 1.7841472625732422, "learning_rate": 0.00015642749053210948, "loss": 4.6997, "step": 566150 }, { "epoch": 3.414092931826678, "grad_norm": 3.6742236614227295, "learning_rate": 0.00015640716019778755, "loss": 5.0367, "step": 566200 }, { "epoch": 3.4143944236080124, "grad_norm": 2.06886887550354, "learning_rate": 0.00015638682974555108, "loss": 5.0274, "step": 566250 }, { "epoch": 3.4146959153893466, "grad_norm": 0.4187009036540985, "learning_rate": 0.00015636649917577408, "loss": 4.2005, "step": 566300 }, { "epoch": 3.4149974071706803, "grad_norm": 2.394676923751831, "learning_rate": 0.0001563461684888307, "loss": 4.9871, "step": 566350 }, { "epoch": 3.4152988989520146, "grad_norm": 1.6671699285507202, "learning_rate": 0.00015632583768509522, "loss": 4.7664, "step": 566400 }, { "epoch": 3.4156003907333488, "grad_norm": 2.048729419708252, "learning_rate": 0.0001563055067649417, "loss": 5.0052, "step": 566450 }, { "epoch": 3.4159018825146825, "grad_norm": 2.2630555629730225, "learning_rate": 0.0001562851757287443, "loss": 4.8993, "step": 566500 }, { "epoch": 3.4162033742960167, "grad_norm": 3.2925281524658203, "learning_rate": 0.00015626484457687722, "loss": 4.6371, "step": 566550 }, { "epoch": 3.416504866077351, "grad_norm": 3.7714717388153076, "learning_rate": 0.00015624451330971466, "loss": 5.2738, "step": 566600 }, { "epoch": 3.4168063578586847, "grad_norm": 1.6719049215316772, "learning_rate": 0.00015622418192763068, "loss": 4.4758, "step": 566650 }, { "epoch": 3.417107849640019, "grad_norm": 2.9916248321533203, "learning_rate": 0.00015620385043099957, "loss": 4.9879, "step": 566700 }, { "epoch": 3.4174093414213527, "grad_norm": 2.1646971702575684, "learning_rate": 0.00015618351882019544, "loss": 4.7274, "step": 566750 }, { "epoch": 3.417710833202687, "grad_norm": 1.7775142192840576, "learning_rate": 0.00015616318709559243, "loss": 4.8469, "step": 566800 }, { "epoch": 3.418012324984021, "grad_norm": 2.0722289085388184, "learning_rate": 0.00015614285525756478, "loss": 4.7742, "step": 566850 }, { "epoch": 3.418313816765355, "grad_norm": 2.4782538414001465, "learning_rate": 0.00015612252330648662, "loss": 5.0632, "step": 566900 }, { "epoch": 3.418615308546689, "grad_norm": 1.8028488159179688, "learning_rate": 0.00015610219124273218, "loss": 5.2432, "step": 566950 }, { "epoch": 3.418916800328023, "grad_norm": 2.40704345703125, "learning_rate": 0.00015608185906667562, "loss": 4.5834, "step": 567000 }, { "epoch": 3.418916800328023, "eval_loss": 5.310447692871094, "eval_runtime": 39.0563, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.543785209595385, "step": 567000 }, { "epoch": 3.419218292109357, "grad_norm": 2.3191463947296143, "learning_rate": 0.00015606152677869115, "loss": 5.0754, "step": 567050 }, { "epoch": 3.4195197838906912, "grad_norm": 1.5085397958755493, "learning_rate": 0.00015604119437915288, "loss": 4.9411, "step": 567100 }, { "epoch": 3.4198212756720254, "grad_norm": 1.5546187162399292, "learning_rate": 0.0001560208618684351, "loss": 4.6827, "step": 567150 }, { "epoch": 3.420122767453359, "grad_norm": 2.156764030456543, "learning_rate": 0.0001560005292469119, "loss": 4.905, "step": 567200 }, { "epoch": 3.4204242592346934, "grad_norm": 1.6593397855758667, "learning_rate": 0.00015598019651495753, "loss": 5.367, "step": 567250 }, { "epoch": 3.420725751016027, "grad_norm": 2.5662214756011963, "learning_rate": 0.00015595986367294618, "loss": 4.5904, "step": 567300 }, { "epoch": 3.4210272427973614, "grad_norm": 2.988889694213867, "learning_rate": 0.00015593953072125206, "loss": 4.7444, "step": 567350 }, { "epoch": 3.4213287345786956, "grad_norm": 2.708301067352295, "learning_rate": 0.00015591919766024933, "loss": 4.7886, "step": 567400 }, { "epoch": 3.4216302263600293, "grad_norm": 1.6972856521606445, "learning_rate": 0.00015589886449031225, "loss": 4.6553, "step": 567450 }, { "epoch": 3.4219317181413635, "grad_norm": 1.5609383583068848, "learning_rate": 0.000155878531211815, "loss": 4.7264, "step": 567500 }, { "epoch": 3.4222332099226973, "grad_norm": 1.9747015237808228, "learning_rate": 0.0001558581978251317, "loss": 4.9568, "step": 567550 }, { "epoch": 3.4225347017040315, "grad_norm": 3.355811595916748, "learning_rate": 0.00015583786433063666, "loss": 4.691, "step": 567600 }, { "epoch": 3.4228361934853657, "grad_norm": 2.0129647254943848, "learning_rate": 0.0001558175307287041, "loss": 4.9992, "step": 567650 }, { "epoch": 3.4231376852666995, "grad_norm": 2.634812116622925, "learning_rate": 0.00015579719701970814, "loss": 4.4923, "step": 567700 }, { "epoch": 3.4234391770480337, "grad_norm": 0.8565313816070557, "learning_rate": 0.00015577686320402304, "loss": 4.9712, "step": 567750 }, { "epoch": 3.423740668829368, "grad_norm": 2.3452138900756836, "learning_rate": 0.00015575652928202303, "loss": 4.0755, "step": 567800 }, { "epoch": 3.4240421606107017, "grad_norm": 2.827721357345581, "learning_rate": 0.00015573619525408234, "loss": 4.6787, "step": 567850 }, { "epoch": 3.424343652392036, "grad_norm": 1.6287637948989868, "learning_rate": 0.00015571586112057514, "loss": 4.8061, "step": 567900 }, { "epoch": 3.42464514417337, "grad_norm": 4.269258975982666, "learning_rate": 0.00015569552688187566, "loss": 4.4756, "step": 567950 }, { "epoch": 3.424946635954704, "grad_norm": 1.7493890523910522, "learning_rate": 0.00015567519253835812, "loss": 5.0162, "step": 568000 }, { "epoch": 3.425248127736038, "grad_norm": 2.805344343185425, "learning_rate": 0.00015565485809039683, "loss": 4.8638, "step": 568050 }, { "epoch": 3.425549619517372, "grad_norm": 0.9623920917510986, "learning_rate": 0.00015563452353836588, "loss": 4.88, "step": 568100 }, { "epoch": 3.425851111298706, "grad_norm": 2.266435384750366, "learning_rate": 0.00015561418888263958, "loss": 4.5409, "step": 568150 }, { "epoch": 3.42615260308004, "grad_norm": 2.368950366973877, "learning_rate": 0.00015559385412359218, "loss": 4.8071, "step": 568200 }, { "epoch": 3.426454094861374, "grad_norm": 1.7317659854888916, "learning_rate": 0.00015557351926159785, "loss": 4.6281, "step": 568250 }, { "epoch": 3.426755586642708, "grad_norm": 2.1926560401916504, "learning_rate": 0.00015555318429703088, "loss": 5.2214, "step": 568300 }, { "epoch": 3.427057078424042, "grad_norm": 3.0136444568634033, "learning_rate": 0.00015553284923026543, "loss": 5.0113, "step": 568350 }, { "epoch": 3.427358570205376, "grad_norm": 2.9802722930908203, "learning_rate": 0.00015551251406167578, "loss": 5.0565, "step": 568400 }, { "epoch": 3.4276600619867104, "grad_norm": 0.6026069521903992, "learning_rate": 0.00015549217879163625, "loss": 4.5406, "step": 568450 }, { "epoch": 3.427961553768044, "grad_norm": 1.7777185440063477, "learning_rate": 0.00015547184342052097, "loss": 5.0316, "step": 568500 }, { "epoch": 3.4282630455493783, "grad_norm": 1.906175971031189, "learning_rate": 0.0001554515079487042, "loss": 4.8296, "step": 568550 }, { "epoch": 3.4285645373307125, "grad_norm": 1.3995795249938965, "learning_rate": 0.0001554311723765602, "loss": 4.9083, "step": 568600 }, { "epoch": 3.4288660291120463, "grad_norm": 1.950645089149475, "learning_rate": 0.00015541083670446326, "loss": 5.0934, "step": 568650 }, { "epoch": 3.4291675208933805, "grad_norm": 2.0192956924438477, "learning_rate": 0.0001553905009327876, "loss": 5.0238, "step": 568700 }, { "epoch": 3.4294690126747147, "grad_norm": 3.844454288482666, "learning_rate": 0.0001553701650619074, "loss": 4.7788, "step": 568750 }, { "epoch": 3.4297705044560485, "grad_norm": 2.2048373222351074, "learning_rate": 0.00015534982909219702, "loss": 5.0735, "step": 568800 }, { "epoch": 3.4300719962373827, "grad_norm": 0.7356613874435425, "learning_rate": 0.00015532949302403065, "loss": 4.8007, "step": 568850 }, { "epoch": 3.4303734880187164, "grad_norm": 3.380371570587158, "learning_rate": 0.00015530915685778258, "loss": 4.7293, "step": 568900 }, { "epoch": 3.4306749798000507, "grad_norm": 1.4407328367233276, "learning_rate": 0.00015528882059382708, "loss": 4.6172, "step": 568950 }, { "epoch": 3.430976471581385, "grad_norm": 3.1979546546936035, "learning_rate": 0.00015526848423253834, "loss": 4.9027, "step": 569000 }, { "epoch": 3.4312779633627186, "grad_norm": 1.9149097204208374, "learning_rate": 0.0001552481477742907, "loss": 4.8698, "step": 569050 }, { "epoch": 3.431579455144053, "grad_norm": 2.2390871047973633, "learning_rate": 0.00015522781121945836, "loss": 5.1524, "step": 569100 }, { "epoch": 3.4318809469253866, "grad_norm": 2.2110633850097656, "learning_rate": 0.00015520747456841563, "loss": 4.5656, "step": 569150 }, { "epoch": 3.432182438706721, "grad_norm": 2.028963565826416, "learning_rate": 0.00015518713782153675, "loss": 4.6622, "step": 569200 }, { "epoch": 3.432483930488055, "grad_norm": 2.9287967681884766, "learning_rate": 0.00015516680097919607, "loss": 4.9903, "step": 569250 }, { "epoch": 3.4327854222693888, "grad_norm": 2.3414487838745117, "learning_rate": 0.00015514646404176768, "loss": 4.8841, "step": 569300 }, { "epoch": 3.433086914050723, "grad_norm": 2.777588367462158, "learning_rate": 0.00015512612700962602, "loss": 4.881, "step": 569350 }, { "epoch": 3.433388405832057, "grad_norm": 3.3290629386901855, "learning_rate": 0.00015510578988314533, "loss": 5.0854, "step": 569400 }, { "epoch": 3.433689897613391, "grad_norm": 2.2237496376037598, "learning_rate": 0.00015508545266269981, "loss": 4.8833, "step": 569450 }, { "epoch": 3.433991389394725, "grad_norm": 2.4000957012176514, "learning_rate": 0.00015506511534866386, "loss": 4.9415, "step": 569500 }, { "epoch": 3.4342928811760594, "grad_norm": 3.448702335357666, "learning_rate": 0.0001550447779414116, "loss": 4.5537, "step": 569550 }, { "epoch": 3.434594372957393, "grad_norm": 2.903372287750244, "learning_rate": 0.00015502444044131745, "loss": 4.6407, "step": 569600 }, { "epoch": 3.4348958647387273, "grad_norm": 2.4141759872436523, "learning_rate": 0.00015500410284875567, "loss": 4.8238, "step": 569650 }, { "epoch": 3.435197356520061, "grad_norm": 2.1861236095428467, "learning_rate": 0.00015498376516410046, "loss": 5.2783, "step": 569700 }, { "epoch": 3.4354988483013953, "grad_norm": 2.8921895027160645, "learning_rate": 0.0001549634273877262, "loss": 4.9278, "step": 569750 }, { "epoch": 3.4358003400827295, "grad_norm": 2.288159132003784, "learning_rate": 0.00015494308952000718, "loss": 4.9845, "step": 569800 }, { "epoch": 3.4361018318640633, "grad_norm": 1.9542036056518555, "learning_rate": 0.0001549227515613176, "loss": 4.8843, "step": 569850 }, { "epoch": 3.4364033236453975, "grad_norm": 2.6563148498535156, "learning_rate": 0.0001549024135120318, "loss": 5.1366, "step": 569900 }, { "epoch": 3.4367048154267312, "grad_norm": 1.8635151386260986, "learning_rate": 0.00015488207537252413, "loss": 5.3014, "step": 569950 }, { "epoch": 3.4370063072080654, "grad_norm": 1.9277206659317017, "learning_rate": 0.00015486173714316872, "loss": 5.1946, "step": 570000 }, { "epoch": 3.4370063072080654, "eval_loss": 5.310383319854736, "eval_runtime": 39.2465, "eval_samples_per_second": 13.046, "eval_steps_per_second": 6.523, "eval_tts_loss": 7.516408598833219, "step": 570000 }, { "epoch": 3.4373077989893996, "grad_norm": 1.713112711906433, "learning_rate": 0.00015484139882434008, "loss": 4.9196, "step": 570050 }, { "epoch": 3.4376092907707334, "grad_norm": 2.591714382171631, "learning_rate": 0.0001548210604164124, "loss": 5.0872, "step": 570100 }, { "epoch": 3.4379107825520676, "grad_norm": 3.0885629653930664, "learning_rate": 0.0001548007219197599, "loss": 4.7329, "step": 570150 }, { "epoch": 3.438212274333402, "grad_norm": 1.8303271532058716, "learning_rate": 0.00015478038333475704, "loss": 5.0087, "step": 570200 }, { "epoch": 3.4385137661147356, "grad_norm": 2.977159261703491, "learning_rate": 0.00015476004466177804, "loss": 4.9016, "step": 570250 }, { "epoch": 3.43881525789607, "grad_norm": 2.201836109161377, "learning_rate": 0.00015473970590119715, "loss": 4.3848, "step": 570300 }, { "epoch": 3.439116749677404, "grad_norm": 2.7094857692718506, "learning_rate": 0.00015471936705338882, "loss": 4.7972, "step": 570350 }, { "epoch": 3.4394182414587378, "grad_norm": 1.4768489599227905, "learning_rate": 0.0001546990281187272, "loss": 4.9857, "step": 570400 }, { "epoch": 3.439719733240072, "grad_norm": 2.3773207664489746, "learning_rate": 0.00015467868909758672, "loss": 4.905, "step": 570450 }, { "epoch": 3.4400212250214057, "grad_norm": 2.3197357654571533, "learning_rate": 0.00015465834999034162, "loss": 4.7137, "step": 570500 }, { "epoch": 3.44032271680274, "grad_norm": 2.0690369606018066, "learning_rate": 0.00015463801079736628, "loss": 4.8987, "step": 570550 }, { "epoch": 3.440624208584074, "grad_norm": 1.1847882270812988, "learning_rate": 0.0001546176715190349, "loss": 4.5664, "step": 570600 }, { "epoch": 3.440925700365408, "grad_norm": 0.6159120798110962, "learning_rate": 0.00015459733215572195, "loss": 4.7922, "step": 570650 }, { "epoch": 3.441227192146742, "grad_norm": 2.0987415313720703, "learning_rate": 0.00015457699270780161, "loss": 4.6791, "step": 570700 }, { "epoch": 3.441528683928076, "grad_norm": 1.1031357049942017, "learning_rate": 0.00015455665317564823, "loss": 4.4685, "step": 570750 }, { "epoch": 3.44183017570941, "grad_norm": 1.9050630331039429, "learning_rate": 0.0001545363135596362, "loss": 5.1504, "step": 570800 }, { "epoch": 3.4421316674907443, "grad_norm": 3.357599973678589, "learning_rate": 0.00015451597386013978, "loss": 5.0485, "step": 570850 }, { "epoch": 3.4424331592720785, "grad_norm": 1.6487666368484497, "learning_rate": 0.00015449563407753328, "loss": 5.0121, "step": 570900 }, { "epoch": 3.4427346510534123, "grad_norm": 0.7258188724517822, "learning_rate": 0.00015447529421219107, "loss": 5.252, "step": 570950 }, { "epoch": 3.4430361428347465, "grad_norm": 3.082263708114624, "learning_rate": 0.00015445495426448744, "loss": 4.9393, "step": 571000 }, { "epoch": 3.4433376346160802, "grad_norm": 1.5572267770767212, "learning_rate": 0.0001544346142347968, "loss": 4.4275, "step": 571050 }, { "epoch": 3.4436391263974144, "grad_norm": 1.9098857641220093, "learning_rate": 0.00015441427412349334, "loss": 4.6615, "step": 571100 }, { "epoch": 3.4439406181787486, "grad_norm": 0.994833767414093, "learning_rate": 0.0001543939339309515, "loss": 5.0287, "step": 571150 }, { "epoch": 3.4442421099600824, "grad_norm": 3.0235769748687744, "learning_rate": 0.0001543735936575455, "loss": 4.8465, "step": 571200 }, { "epoch": 3.4445436017414166, "grad_norm": 2.762104034423828, "learning_rate": 0.00015435325330364984, "loss": 4.8274, "step": 571250 }, { "epoch": 3.4448450935227504, "grad_norm": 1.7350571155548096, "learning_rate": 0.00015433291286963874, "loss": 5.0524, "step": 571300 }, { "epoch": 3.4451465853040846, "grad_norm": 2.455066442489624, "learning_rate": 0.00015431257235588654, "loss": 4.7534, "step": 571350 }, { "epoch": 3.445448077085419, "grad_norm": 1.8259177207946777, "learning_rate": 0.00015429223176276763, "loss": 4.7411, "step": 571400 }, { "epoch": 3.4457495688667525, "grad_norm": 1.9396237134933472, "learning_rate": 0.00015427189109065633, "loss": 4.7576, "step": 571450 }, { "epoch": 3.4460510606480868, "grad_norm": 1.833455204963684, "learning_rate": 0.00015425155033992695, "loss": 5.1224, "step": 571500 }, { "epoch": 3.446352552429421, "grad_norm": 1.8612457513809204, "learning_rate": 0.00015423120951095384, "loss": 5.0249, "step": 571550 }, { "epoch": 3.4466540442107547, "grad_norm": 2.499206066131592, "learning_rate": 0.00015421086860411138, "loss": 4.3443, "step": 571600 }, { "epoch": 3.446955535992089, "grad_norm": 2.583594560623169, "learning_rate": 0.0001541905276197739, "loss": 4.914, "step": 571650 }, { "epoch": 3.447257027773423, "grad_norm": 2.8955078125, "learning_rate": 0.00015417018655831572, "loss": 4.8753, "step": 571700 }, { "epoch": 3.447558519554757, "grad_norm": 2.052600383758545, "learning_rate": 0.0001541498454201112, "loss": 5.0076, "step": 571750 }, { "epoch": 3.447860011336091, "grad_norm": 1.8146806955337524, "learning_rate": 0.0001541295042055347, "loss": 4.9725, "step": 571800 }, { "epoch": 3.448161503117425, "grad_norm": 0.504152774810791, "learning_rate": 0.0001541091629149606, "loss": 4.5958, "step": 571850 }, { "epoch": 3.448462994898759, "grad_norm": 2.2367377281188965, "learning_rate": 0.0001540888215487632, "loss": 5.2116, "step": 571900 }, { "epoch": 3.4487644866800933, "grad_norm": 1.2258819341659546, "learning_rate": 0.00015406848010731682, "loss": 4.8758, "step": 571950 }, { "epoch": 3.449065978461427, "grad_norm": 2.077636957168579, "learning_rate": 0.00015404813859099594, "loss": 4.773, "step": 572000 }, { "epoch": 3.4493674702427612, "grad_norm": 2.928683042526245, "learning_rate": 0.00015402779700017484, "loss": 4.9692, "step": 572050 }, { "epoch": 3.449668962024095, "grad_norm": 1.8219529390335083, "learning_rate": 0.0001540074553352279, "loss": 4.9505, "step": 572100 }, { "epoch": 3.449970453805429, "grad_norm": 2.658464193344116, "learning_rate": 0.00015398711359652941, "loss": 4.7216, "step": 572150 }, { "epoch": 3.4502719455867634, "grad_norm": 1.752777338027954, "learning_rate": 0.00015396677178445377, "loss": 4.314, "step": 572200 }, { "epoch": 3.450573437368097, "grad_norm": 2.1945250034332275, "learning_rate": 0.00015394642989937542, "loss": 4.6109, "step": 572250 }, { "epoch": 3.4508749291494314, "grad_norm": 2.192782163619995, "learning_rate": 0.00015392608794166865, "loss": 4.5554, "step": 572300 }, { "epoch": 3.4511764209307656, "grad_norm": 1.7428243160247803, "learning_rate": 0.00015390574591170778, "loss": 5.111, "step": 572350 }, { "epoch": 3.4514779127120994, "grad_norm": 2.2018487453460693, "learning_rate": 0.00015388540380986728, "loss": 4.7804, "step": 572400 }, { "epoch": 3.4517794044934336, "grad_norm": 3.548020124435425, "learning_rate": 0.0001538650616365215, "loss": 5.2223, "step": 572450 }, { "epoch": 3.4520808962747678, "grad_norm": 2.0750529766082764, "learning_rate": 0.0001538447193920447, "loss": 4.9518, "step": 572500 }, { "epoch": 3.4523823880561015, "grad_norm": 2.7075533866882324, "learning_rate": 0.00015382437707681137, "loss": 4.9162, "step": 572550 }, { "epoch": 3.4526838798374357, "grad_norm": 4.074883937835693, "learning_rate": 0.0001538040346911958, "loss": 4.9208, "step": 572600 }, { "epoch": 3.4529853716187695, "grad_norm": 1.8318344354629517, "learning_rate": 0.00015378369223557244, "loss": 5.1272, "step": 572650 }, { "epoch": 3.4532868634001037, "grad_norm": 1.4441734552383423, "learning_rate": 0.00015376334971031562, "loss": 4.8909, "step": 572700 }, { "epoch": 3.453588355181438, "grad_norm": 2.45483660697937, "learning_rate": 0.0001537430071157997, "loss": 4.9373, "step": 572750 }, { "epoch": 3.4538898469627717, "grad_norm": 2.245894193649292, "learning_rate": 0.0001537226644523991, "loss": 4.8302, "step": 572800 }, { "epoch": 3.454191338744106, "grad_norm": 2.4584760665893555, "learning_rate": 0.00015370232172048817, "loss": 5.0494, "step": 572850 }, { "epoch": 3.4544928305254397, "grad_norm": 2.761199712753296, "learning_rate": 0.00015368197892044128, "loss": 5.3426, "step": 572900 }, { "epoch": 3.454794322306774, "grad_norm": 0.5719131231307983, "learning_rate": 0.0001536616360526328, "loss": 5.2302, "step": 572950 }, { "epoch": 3.455095814088108, "grad_norm": 3.278090000152588, "learning_rate": 0.00015364129311743716, "loss": 4.9767, "step": 573000 }, { "epoch": 3.455095814088108, "eval_loss": 5.30024528503418, "eval_runtime": 39.265, "eval_samples_per_second": 13.04, "eval_steps_per_second": 6.52, "eval_tts_loss": 7.56272570134117, "step": 573000 }, { "epoch": 3.455397305869442, "grad_norm": 1.7695643901824951, "learning_rate": 0.00015362095011522872, "loss": 4.6939, "step": 573050 }, { "epoch": 3.455698797650776, "grad_norm": 3.019076347351074, "learning_rate": 0.00015360060704638185, "loss": 4.7985, "step": 573100 }, { "epoch": 3.4560002894321102, "grad_norm": 2.562635660171509, "learning_rate": 0.00015358026391127096, "loss": 4.4402, "step": 573150 }, { "epoch": 3.456301781213444, "grad_norm": 2.4828267097473145, "learning_rate": 0.00015355992071027042, "loss": 4.8742, "step": 573200 }, { "epoch": 3.456603272994778, "grad_norm": 2.122727870941162, "learning_rate": 0.0001535395774437546, "loss": 4.8491, "step": 573250 }, { "epoch": 3.4569047647761124, "grad_norm": 2.8257415294647217, "learning_rate": 0.00015351923411209794, "loss": 5.2772, "step": 573300 }, { "epoch": 3.457206256557446, "grad_norm": 2.6766178607940674, "learning_rate": 0.00015349889071567473, "loss": 5.4932, "step": 573350 }, { "epoch": 3.4575077483387804, "grad_norm": 2.8324756622314453, "learning_rate": 0.00015347854725485948, "loss": 4.9218, "step": 573400 }, { "epoch": 3.457809240120114, "grad_norm": 1.8764714002609253, "learning_rate": 0.00015345820373002656, "loss": 4.8233, "step": 573450 }, { "epoch": 3.4581107319014484, "grad_norm": 2.105234146118164, "learning_rate": 0.0001534378601415503, "loss": 4.5209, "step": 573500 }, { "epoch": 3.4584122236827826, "grad_norm": 1.9994436502456665, "learning_rate": 0.00015341751648980515, "loss": 4.7695, "step": 573550 }, { "epoch": 3.4587137154641163, "grad_norm": 1.914555311203003, "learning_rate": 0.00015339717277516545, "loss": 5.0728, "step": 573600 }, { "epoch": 3.4590152072454505, "grad_norm": 2.509326934814453, "learning_rate": 0.00015337682899800568, "loss": 4.896, "step": 573650 }, { "epoch": 3.4593166990267843, "grad_norm": 3.1758296489715576, "learning_rate": 0.00015335648515870016, "loss": 4.8275, "step": 573700 }, { "epoch": 3.4596181908081185, "grad_norm": 2.1415770053863525, "learning_rate": 0.0001533361412576233, "loss": 4.4846, "step": 573750 }, { "epoch": 3.4599196825894527, "grad_norm": 2.860419988632202, "learning_rate": 0.00015331579729514958, "loss": 4.6531, "step": 573800 }, { "epoch": 3.4602211743707865, "grad_norm": 2.7260775566101074, "learning_rate": 0.0001532954532716533, "loss": 4.7335, "step": 573850 }, { "epoch": 3.4605226661521207, "grad_norm": 3.6674411296844482, "learning_rate": 0.00015327510918750891, "loss": 5.3308, "step": 573900 }, { "epoch": 3.460824157933455, "grad_norm": 1.5566514730453491, "learning_rate": 0.0001532547650430908, "loss": 4.8042, "step": 573950 }, { "epoch": 3.4611256497147886, "grad_norm": 1.9347695112228394, "learning_rate": 0.00015323442083877342, "loss": 4.9862, "step": 574000 }, { "epoch": 3.461427141496123, "grad_norm": 2.164823055267334, "learning_rate": 0.00015321407657493114, "loss": 4.8055, "step": 574050 }, { "epoch": 3.461728633277457, "grad_norm": 2.0933382511138916, "learning_rate": 0.00015319373225193835, "loss": 4.9513, "step": 574100 }, { "epoch": 3.462030125058791, "grad_norm": 1.9353502988815308, "learning_rate": 0.00015317338787016943, "loss": 4.9284, "step": 574150 }, { "epoch": 3.462331616840125, "grad_norm": 2.0212559700012207, "learning_rate": 0.00015315304342999888, "loss": 4.5537, "step": 574200 }, { "epoch": 3.462633108621459, "grad_norm": 3.689751148223877, "learning_rate": 0.00015313269893180108, "loss": 4.8377, "step": 574250 }, { "epoch": 3.462934600402793, "grad_norm": 2.780531167984009, "learning_rate": 0.00015311235437595044, "loss": 4.3822, "step": 574300 }, { "epoch": 3.463236092184127, "grad_norm": 2.1594552993774414, "learning_rate": 0.00015309200976282132, "loss": 4.8714, "step": 574350 }, { "epoch": 3.463537583965461, "grad_norm": 1.998472809791565, "learning_rate": 0.0001530716650927882, "loss": 4.7028, "step": 574400 }, { "epoch": 3.463839075746795, "grad_norm": 2.003480911254883, "learning_rate": 0.00015305132036622543, "loss": 4.6757, "step": 574450 }, { "epoch": 3.4641405675281294, "grad_norm": 2.042597770690918, "learning_rate": 0.0001530309755835075, "loss": 4.6649, "step": 574500 }, { "epoch": 3.464442059309463, "grad_norm": 1.9205389022827148, "learning_rate": 0.00015301063074500874, "loss": 4.5021, "step": 574550 }, { "epoch": 3.4647435510907973, "grad_norm": 2.2384817600250244, "learning_rate": 0.00015299028585110365, "loss": 4.8413, "step": 574600 }, { "epoch": 3.4650450428721316, "grad_norm": 2.12558913230896, "learning_rate": 0.00015296994090216662, "loss": 4.9808, "step": 574650 }, { "epoch": 3.4653465346534653, "grad_norm": 2.611314296722412, "learning_rate": 0.00015294959589857207, "loss": 5.0973, "step": 574700 }, { "epoch": 3.4656480264347995, "grad_norm": 2.054981231689453, "learning_rate": 0.0001529292508406944, "loss": 4.3996, "step": 574750 }, { "epoch": 3.4659495182161333, "grad_norm": 2.1116864681243896, "learning_rate": 0.00015290890572890807, "loss": 4.9447, "step": 574800 }, { "epoch": 3.4662510099974675, "grad_norm": 1.6189608573913574, "learning_rate": 0.00015288856056358747, "loss": 4.8921, "step": 574850 }, { "epoch": 3.4665525017788017, "grad_norm": 2.992522716522217, "learning_rate": 0.00015286821534510705, "loss": 5.0538, "step": 574900 }, { "epoch": 3.4668539935601355, "grad_norm": 2.0434982776641846, "learning_rate": 0.00015284787007384116, "loss": 4.8372, "step": 574950 }, { "epoch": 3.4671554853414697, "grad_norm": 2.2773172855377197, "learning_rate": 0.00015282752475016435, "loss": 5.0607, "step": 575000 }, { "epoch": 3.4674569771228034, "grad_norm": 2.1401357650756836, "learning_rate": 0.00015280717937445098, "loss": 4.9766, "step": 575050 }, { "epoch": 3.4677584689041376, "grad_norm": 1.7542849779129028, "learning_rate": 0.00015278683394707543, "loss": 5.2441, "step": 575100 }, { "epoch": 3.468059960685472, "grad_norm": 1.9259129762649536, "learning_rate": 0.0001527664884684122, "loss": 4.6103, "step": 575150 }, { "epoch": 3.4683614524668056, "grad_norm": 2.5028233528137207, "learning_rate": 0.00015274614293883573, "loss": 4.764, "step": 575200 }, { "epoch": 3.46866294424814, "grad_norm": 1.7163400650024414, "learning_rate": 0.0001527257973587204, "loss": 4.9394, "step": 575250 }, { "epoch": 3.468964436029474, "grad_norm": 1.7661248445510864, "learning_rate": 0.00015270545172844065, "loss": 4.8889, "step": 575300 }, { "epoch": 3.469265927810808, "grad_norm": 1.784015417098999, "learning_rate": 0.0001526851060483709, "loss": 5.1697, "step": 575350 }, { "epoch": 3.469567419592142, "grad_norm": 2.0405287742614746, "learning_rate": 0.00015266476031888568, "loss": 4.8568, "step": 575400 }, { "epoch": 3.469868911373476, "grad_norm": 2.5362889766693115, "learning_rate": 0.0001526444145403593, "loss": 4.3871, "step": 575450 }, { "epoch": 3.47017040315481, "grad_norm": 1.723375678062439, "learning_rate": 0.00015262406871316623, "loss": 5.1592, "step": 575500 }, { "epoch": 3.470471894936144, "grad_norm": 1.9661428928375244, "learning_rate": 0.0001526037228376809, "loss": 4.6716, "step": 575550 }, { "epoch": 3.470773386717478, "grad_norm": 2.4476125240325928, "learning_rate": 0.0001525833769142778, "loss": 5.3067, "step": 575600 }, { "epoch": 3.471074878498812, "grad_norm": 1.5041090250015259, "learning_rate": 0.00015256303094333138, "loss": 5.2111, "step": 575650 }, { "epoch": 3.4713763702801463, "grad_norm": 1.2426352500915527, "learning_rate": 0.00015254268492521597, "loss": 4.7437, "step": 575700 }, { "epoch": 3.47167786206148, "grad_norm": 1.8460973501205444, "learning_rate": 0.00015252233886030607, "loss": 4.9874, "step": 575750 }, { "epoch": 3.4719793538428143, "grad_norm": 2.110992670059204, "learning_rate": 0.0001525019927489762, "loss": 5.1647, "step": 575800 }, { "epoch": 3.472280845624148, "grad_norm": 2.474165916442871, "learning_rate": 0.00015248164659160065, "loss": 4.6215, "step": 575850 }, { "epoch": 3.4725823374054823, "grad_norm": 2.1901605129241943, "learning_rate": 0.00015246130038855396, "loss": 4.6731, "step": 575900 }, { "epoch": 3.4728838291868165, "grad_norm": 2.2809784412384033, "learning_rate": 0.00015244095414021052, "loss": 5.0097, "step": 575950 }, { "epoch": 3.4731853209681502, "grad_norm": 1.9544445276260376, "learning_rate": 0.00015242060784694482, "loss": 4.7485, "step": 576000 }, { "epoch": 3.4731853209681502, "eval_loss": 5.3045244216918945, "eval_runtime": 39.0767, "eval_samples_per_second": 13.102, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.512238689303739, "step": 576000 }, { "epoch": 3.4734868127494845, "grad_norm": 2.031341791152954, "learning_rate": 0.00015240026150913132, "loss": 5.2144, "step": 576050 }, { "epoch": 3.4737883045308187, "grad_norm": 1.3740479946136475, "learning_rate": 0.0001523799151271444, "loss": 4.5492, "step": 576100 }, { "epoch": 3.4740897963121524, "grad_norm": 1.9430803060531616, "learning_rate": 0.00015235956870135852, "loss": 4.9812, "step": 576150 }, { "epoch": 3.4743912880934866, "grad_norm": 2.8981120586395264, "learning_rate": 0.00015233922223214816, "loss": 5.1741, "step": 576200 }, { "epoch": 3.474692779874821, "grad_norm": 2.3677818775177, "learning_rate": 0.00015231887571988783, "loss": 5.2439, "step": 576250 }, { "epoch": 3.4749942716561546, "grad_norm": 1.7801339626312256, "learning_rate": 0.00015229852916495178, "loss": 4.8389, "step": 576300 }, { "epoch": 3.475295763437489, "grad_norm": 2.586190938949585, "learning_rate": 0.00015227818256771463, "loss": 5.1181, "step": 576350 }, { "epoch": 3.4755972552188226, "grad_norm": 1.7264363765716553, "learning_rate": 0.0001522578359285508, "loss": 4.9076, "step": 576400 }, { "epoch": 3.4758987470001568, "grad_norm": 1.8043650388717651, "learning_rate": 0.0001522374892478347, "loss": 4.9421, "step": 576450 }, { "epoch": 3.476200238781491, "grad_norm": 2.4552295207977295, "learning_rate": 0.00015221714252594085, "loss": 4.6856, "step": 576500 }, { "epoch": 3.4765017305628247, "grad_norm": 2.993420124053955, "learning_rate": 0.00015219679576324355, "loss": 5.129, "step": 576550 }, { "epoch": 3.476803222344159, "grad_norm": 1.3638246059417725, "learning_rate": 0.00015217644896011747, "loss": 4.8172, "step": 576600 }, { "epoch": 3.4771047141254927, "grad_norm": 2.7255754470825195, "learning_rate": 0.0001521561021169369, "loss": 4.7372, "step": 576650 }, { "epoch": 3.477406205906827, "grad_norm": 1.6195058822631836, "learning_rate": 0.00015213575523407633, "loss": 4.8334, "step": 576700 }, { "epoch": 3.477707697688161, "grad_norm": 1.81043541431427, "learning_rate": 0.00015211540831191027, "loss": 4.9816, "step": 576750 }, { "epoch": 3.478009189469495, "grad_norm": 1.1545677185058594, "learning_rate": 0.00015209506135081312, "loss": 5.0336, "step": 576800 }, { "epoch": 3.478310681250829, "grad_norm": 2.531677007675171, "learning_rate": 0.00015207471435115937, "loss": 4.5611, "step": 576850 }, { "epoch": 3.4786121730321633, "grad_norm": 1.917659044265747, "learning_rate": 0.00015205436731332347, "loss": 4.6381, "step": 576900 }, { "epoch": 3.478913664813497, "grad_norm": 2.2035884857177734, "learning_rate": 0.00015203402023767986, "loss": 5.0411, "step": 576950 }, { "epoch": 3.4792151565948313, "grad_norm": 2.503218173980713, "learning_rate": 0.000152013673124603, "loss": 4.8935, "step": 577000 }, { "epoch": 3.4795166483761655, "grad_norm": 1.7667295932769775, "learning_rate": 0.00015199332597446737, "loss": 5.0054, "step": 577050 }, { "epoch": 3.4798181401574992, "grad_norm": 3.641355514526367, "learning_rate": 0.00015197297878764742, "loss": 4.8814, "step": 577100 }, { "epoch": 3.4801196319388334, "grad_norm": 3.1733827590942383, "learning_rate": 0.0001519526315645176, "loss": 4.3159, "step": 577150 }, { "epoch": 3.480421123720167, "grad_norm": 2.1942191123962402, "learning_rate": 0.0001519322843054524, "loss": 4.5899, "step": 577200 }, { "epoch": 3.4807226155015014, "grad_norm": 1.802143931388855, "learning_rate": 0.0001519119370108263, "loss": 5.0276, "step": 577250 }, { "epoch": 3.4810241072828356, "grad_norm": 6.020052433013916, "learning_rate": 0.00015189158968101363, "loss": 5.1669, "step": 577300 }, { "epoch": 3.4813255990641694, "grad_norm": 3.8050763607025146, "learning_rate": 0.000151871242316389, "loss": 4.517, "step": 577350 }, { "epoch": 3.4816270908455036, "grad_norm": 2.0157053470611572, "learning_rate": 0.00015185089491732683, "loss": 4.9093, "step": 577400 }, { "epoch": 3.4819285826268374, "grad_norm": 2.5504157543182373, "learning_rate": 0.0001518305474842016, "loss": 4.47, "step": 577450 }, { "epoch": 3.4822300744081716, "grad_norm": 1.1399621963500977, "learning_rate": 0.00015181020001738774, "loss": 4.8808, "step": 577500 }, { "epoch": 3.4825315661895058, "grad_norm": 0.778329610824585, "learning_rate": 0.00015178985251725974, "loss": 4.7134, "step": 577550 }, { "epoch": 3.4828330579708395, "grad_norm": 2.701050281524658, "learning_rate": 0.00015176950498419205, "loss": 4.9247, "step": 577600 }, { "epoch": 3.4831345497521737, "grad_norm": 2.2359445095062256, "learning_rate": 0.00015174915741855913, "loss": 4.8381, "step": 577650 }, { "epoch": 3.483436041533508, "grad_norm": 1.9188079833984375, "learning_rate": 0.0001517288098207355, "loss": 4.8803, "step": 577700 }, { "epoch": 3.4837375333148417, "grad_norm": 1.787710428237915, "learning_rate": 0.00015170846219109553, "loss": 5.1829, "step": 577750 }, { "epoch": 3.484039025096176, "grad_norm": 2.3265583515167236, "learning_rate": 0.00015168811453001378, "loss": 5.186, "step": 577800 }, { "epoch": 3.48434051687751, "grad_norm": 2.0731799602508545, "learning_rate": 0.00015166776683786477, "loss": 4.5648, "step": 577850 }, { "epoch": 3.484642008658844, "grad_norm": 2.292104482650757, "learning_rate": 0.00015164741911502278, "loss": 5.012, "step": 577900 }, { "epoch": 3.484943500440178, "grad_norm": 2.008298873901367, "learning_rate": 0.00015162707136186243, "loss": 4.5574, "step": 577950 }, { "epoch": 3.485244992221512, "grad_norm": 2.5231242179870605, "learning_rate": 0.00015160672357875815, "loss": 4.755, "step": 578000 }, { "epoch": 3.485546484002846, "grad_norm": 5.310060501098633, "learning_rate": 0.00015158637576608442, "loss": 4.9039, "step": 578050 }, { "epoch": 3.4858479757841803, "grad_norm": 2.739854574203491, "learning_rate": 0.00015156602792421576, "loss": 4.9861, "step": 578100 }, { "epoch": 3.486149467565514, "grad_norm": 0.8156861066818237, "learning_rate": 0.0001515456800535265, "loss": 5.013, "step": 578150 }, { "epoch": 3.4864509593468482, "grad_norm": 2.316558599472046, "learning_rate": 0.00015152533215439127, "loss": 4.9462, "step": 578200 }, { "epoch": 3.4867524511281824, "grad_norm": 2.257267713546753, "learning_rate": 0.00015150498422718446, "loss": 4.4647, "step": 578250 }, { "epoch": 3.487053942909516, "grad_norm": 2.8025147914886475, "learning_rate": 0.00015148463627228054, "loss": 4.8596, "step": 578300 }, { "epoch": 3.4873554346908504, "grad_norm": 3.306666612625122, "learning_rate": 0.00015146428829005405, "loss": 5.2052, "step": 578350 }, { "epoch": 3.4876569264721846, "grad_norm": 2.316453695297241, "learning_rate": 0.0001514439402808794, "loss": 4.7416, "step": 578400 }, { "epoch": 3.4879584182535184, "grad_norm": 0.620675265789032, "learning_rate": 0.0001514235922451311, "loss": 5.1696, "step": 578450 }, { "epoch": 3.4882599100348526, "grad_norm": 1.973557710647583, "learning_rate": 0.0001514032441831836, "loss": 5.3712, "step": 578500 }, { "epoch": 3.4885614018161863, "grad_norm": 3.2282776832580566, "learning_rate": 0.0001513828960954114, "loss": 4.918, "step": 578550 }, { "epoch": 3.4888628935975206, "grad_norm": 2.2039883136749268, "learning_rate": 0.00015136254798218897, "loss": 4.6774, "step": 578600 }, { "epoch": 3.4891643853788548, "grad_norm": 1.8238054513931274, "learning_rate": 0.0001513421998438908, "loss": 4.7997, "step": 578650 }, { "epoch": 3.4894658771601885, "grad_norm": 2.304396867752075, "learning_rate": 0.00015132185168089136, "loss": 4.5528, "step": 578700 }, { "epoch": 3.4897673689415227, "grad_norm": 2.1414589881896973, "learning_rate": 0.0001513015034935651, "loss": 5.0501, "step": 578750 }, { "epoch": 3.4900688607228565, "grad_norm": 1.7455182075500488, "learning_rate": 0.00015128115528228656, "loss": 5.0927, "step": 578800 }, { "epoch": 3.4903703525041907, "grad_norm": 1.7908743619918823, "learning_rate": 0.00015126080704743017, "loss": 4.9683, "step": 578850 }, { "epoch": 3.490671844285525, "grad_norm": 2.1183254718780518, "learning_rate": 0.00015124045878937037, "loss": 4.7451, "step": 578900 }, { "epoch": 3.4909733360668587, "grad_norm": 2.8899881839752197, "learning_rate": 0.00015122011050848174, "loss": 4.8771, "step": 578950 }, { "epoch": 3.491274827848193, "grad_norm": 2.983694076538086, "learning_rate": 0.00015119976220513874, "loss": 5.1094, "step": 579000 }, { "epoch": 3.491274827848193, "eval_loss": 5.304281234741211, "eval_runtime": 38.8769, "eval_samples_per_second": 13.17, "eval_steps_per_second": 6.585, "eval_tts_loss": 7.564632987473256, "step": 579000 }, { "epoch": 3.491576319629527, "grad_norm": 1.7459535598754883, "learning_rate": 0.00015117941387971583, "loss": 5.0659, "step": 579050 }, { "epoch": 3.491877811410861, "grad_norm": 2.5645155906677246, "learning_rate": 0.0001511590655325875, "loss": 4.8265, "step": 579100 }, { "epoch": 3.492179303192195, "grad_norm": 2.1181464195251465, "learning_rate": 0.00015113871716412817, "loss": 4.4189, "step": 579150 }, { "epoch": 3.4924807949735293, "grad_norm": 1.8436352014541626, "learning_rate": 0.00015111836877471243, "loss": 4.5103, "step": 579200 }, { "epoch": 3.492782286754863, "grad_norm": 2.2255799770355225, "learning_rate": 0.00015109802036471467, "loss": 4.6893, "step": 579250 }, { "epoch": 3.4930837785361972, "grad_norm": 1.7951698303222656, "learning_rate": 0.00015107767193450942, "loss": 4.7662, "step": 579300 }, { "epoch": 3.493385270317531, "grad_norm": 1.3229963779449463, "learning_rate": 0.00015105732348447113, "loss": 5.1861, "step": 579350 }, { "epoch": 3.493686762098865, "grad_norm": 2.1510860919952393, "learning_rate": 0.00015103697501497437, "loss": 4.9892, "step": 579400 }, { "epoch": 3.4939882538801994, "grad_norm": 1.844804048538208, "learning_rate": 0.00015101662652639356, "loss": 4.9091, "step": 579450 }, { "epoch": 3.494289745661533, "grad_norm": 2.110811233520508, "learning_rate": 0.00015099627801910313, "loss": 5.0502, "step": 579500 }, { "epoch": 3.4945912374428674, "grad_norm": 3.165330648422241, "learning_rate": 0.00015097592949347765, "loss": 4.9845, "step": 579550 }, { "epoch": 3.494892729224201, "grad_norm": 1.9408930540084839, "learning_rate": 0.00015095558094989158, "loss": 4.6094, "step": 579600 }, { "epoch": 3.4951942210055353, "grad_norm": 1.4798226356506348, "learning_rate": 0.00015093523238871942, "loss": 4.187, "step": 579650 }, { "epoch": 3.4954957127868695, "grad_norm": 1.4805030822753906, "learning_rate": 0.00015091488381033567, "loss": 4.6844, "step": 579700 }, { "epoch": 3.4957972045682033, "grad_norm": 2.943112850189209, "learning_rate": 0.0001508945352151147, "loss": 4.3194, "step": 579750 }, { "epoch": 3.4960986963495375, "grad_norm": 2.7980268001556396, "learning_rate": 0.0001508741866034312, "loss": 4.8525, "step": 579800 }, { "epoch": 3.4964001881308717, "grad_norm": 1.8958226442337036, "learning_rate": 0.00015085383797565948, "loss": 4.879, "step": 579850 }, { "epoch": 3.4967016799122055, "grad_norm": 2.404177188873291, "learning_rate": 0.00015083348933217408, "loss": 5.0963, "step": 579900 }, { "epoch": 3.4970031716935397, "grad_norm": 2.4553382396698, "learning_rate": 0.00015081314067334948, "loss": 4.7566, "step": 579950 }, { "epoch": 3.497304663474874, "grad_norm": 2.5593581199645996, "learning_rate": 0.00015079279199956027, "loss": 4.7547, "step": 580000 }, { "epoch": 3.4976061552562077, "grad_norm": 1.0016651153564453, "learning_rate": 0.0001507724433111808, "loss": 4.6289, "step": 580050 }, { "epoch": 3.497907647037542, "grad_norm": 2.553210496902466, "learning_rate": 0.00015075209460858562, "loss": 4.8459, "step": 580100 }, { "epoch": 3.4982091388188756, "grad_norm": 2.0114762783050537, "learning_rate": 0.00015073174589214918, "loss": 4.8657, "step": 580150 }, { "epoch": 3.49851063060021, "grad_norm": 1.9867379665374756, "learning_rate": 0.00015071139716224603, "loss": 5.0307, "step": 580200 }, { "epoch": 3.498812122381544, "grad_norm": 2.432142496109009, "learning_rate": 0.00015069104841925065, "loss": 4.6218, "step": 580250 }, { "epoch": 3.499113614162878, "grad_norm": 2.4741315841674805, "learning_rate": 0.00015067069966353752, "loss": 4.552, "step": 580300 }, { "epoch": 3.499415105944212, "grad_norm": 2.241514205932617, "learning_rate": 0.00015065035089548105, "loss": 4.8336, "step": 580350 }, { "epoch": 3.4997165977255458, "grad_norm": 2.1126458644866943, "learning_rate": 0.00015063000211545585, "loss": 4.8224, "step": 580400 }, { "epoch": 3.50001808950688, "grad_norm": 1.9213961362838745, "learning_rate": 0.0001506096533238364, "loss": 4.8611, "step": 580450 }, { "epoch": 3.500319581288214, "grad_norm": 2.4439384937286377, "learning_rate": 0.00015058930452099706, "loss": 4.467, "step": 580500 }, { "epoch": 3.5006210730695484, "grad_norm": 1.743458867073059, "learning_rate": 0.00015056895570731246, "loss": 4.8735, "step": 580550 }, { "epoch": 3.500922564850882, "grad_norm": 3.175896167755127, "learning_rate": 0.00015054860688315703, "loss": 4.9905, "step": 580600 }, { "epoch": 3.5012240566322164, "grad_norm": 1.091936707496643, "learning_rate": 0.0001505282580489053, "loss": 5.0508, "step": 580650 }, { "epoch": 3.50152554841355, "grad_norm": 1.7422306537628174, "learning_rate": 0.0001505079092049317, "loss": 4.7676, "step": 580700 }, { "epoch": 3.5018270401948843, "grad_norm": 2.4983136653900146, "learning_rate": 0.00015048756035161075, "loss": 4.9133, "step": 580750 }, { "epoch": 3.5021285319762185, "grad_norm": 2.1430258750915527, "learning_rate": 0.00015046721148931702, "loss": 5.162, "step": 580800 }, { "epoch": 3.5024300237575523, "grad_norm": 1.8009815216064453, "learning_rate": 0.00015044686261842487, "loss": 4.8125, "step": 580850 }, { "epoch": 3.5027315155388865, "grad_norm": 1.7307697534561157, "learning_rate": 0.00015042651373930885, "loss": 4.6083, "step": 580900 }, { "epoch": 3.5030330073202203, "grad_norm": 2.59614896774292, "learning_rate": 0.00015040616485234343, "loss": 5.0218, "step": 580950 }, { "epoch": 3.5033344991015545, "grad_norm": 1.599149465560913, "learning_rate": 0.0001503858159579032, "loss": 4.7871, "step": 581000 }, { "epoch": 3.5036359908828887, "grad_norm": 0.539177417755127, "learning_rate": 0.00015036546705636254, "loss": 4.4483, "step": 581050 }, { "epoch": 3.5039374826642224, "grad_norm": 3.121891498565674, "learning_rate": 0.00015034511814809593, "loss": 5.085, "step": 581100 }, { "epoch": 3.5042389744455567, "grad_norm": 0.689150333404541, "learning_rate": 0.00015032476923347796, "loss": 4.7646, "step": 581150 }, { "epoch": 3.5045404662268904, "grad_norm": 2.035635232925415, "learning_rate": 0.00015030442031288308, "loss": 4.9212, "step": 581200 }, { "epoch": 3.5048419580082246, "grad_norm": 3.0515263080596924, "learning_rate": 0.00015028407138668577, "loss": 4.9317, "step": 581250 }, { "epoch": 3.505143449789559, "grad_norm": 1.465239405632019, "learning_rate": 0.00015026372245526053, "loss": 4.5423, "step": 581300 }, { "epoch": 3.505444941570893, "grad_norm": 2.7222793102264404, "learning_rate": 0.00015024337351898185, "loss": 5.0271, "step": 581350 }, { "epoch": 3.505746433352227, "grad_norm": 1.8391350507736206, "learning_rate": 0.00015022302457822428, "loss": 5.0053, "step": 581400 }, { "epoch": 3.506047925133561, "grad_norm": 1.8474278450012207, "learning_rate": 0.00015020267563336221, "loss": 5.0389, "step": 581450 }, { "epoch": 3.5063494169148948, "grad_norm": 2.736783504486084, "learning_rate": 0.00015018232668477014, "loss": 4.7246, "step": 581500 }, { "epoch": 3.506650908696229, "grad_norm": 1.880911946296692, "learning_rate": 0.00015016197773282268, "loss": 5.2153, "step": 581550 }, { "epoch": 3.506952400477563, "grad_norm": 2.7442996501922607, "learning_rate": 0.00015014162877789426, "loss": 4.6931, "step": 581600 }, { "epoch": 3.507253892258897, "grad_norm": 2.586209297180176, "learning_rate": 0.0001501212798203593, "loss": 4.6965, "step": 581650 }, { "epoch": 3.507555384040231, "grad_norm": 2.473187208175659, "learning_rate": 0.00015010093086059243, "loss": 5.429, "step": 581700 }, { "epoch": 3.507856875821565, "grad_norm": 1.7580628395080566, "learning_rate": 0.00015008058189896803, "loss": 4.6637, "step": 581750 }, { "epoch": 3.508158367602899, "grad_norm": 2.786762237548828, "learning_rate": 0.0001500602329358606, "loss": 5.1398, "step": 581800 }, { "epoch": 3.5084598593842333, "grad_norm": 2.0547666549682617, "learning_rate": 0.00015003988397164472, "loss": 4.8116, "step": 581850 }, { "epoch": 3.508761351165567, "grad_norm": 2.273481607437134, "learning_rate": 0.00015001953500669483, "loss": 5.0875, "step": 581900 }, { "epoch": 3.5090628429469013, "grad_norm": 1.6394052505493164, "learning_rate": 0.00014999918604138542, "loss": 4.8991, "step": 581950 }, { "epoch": 3.509364334728235, "grad_norm": 1.8844276666641235, "learning_rate": 0.000149978837076091, "loss": 4.5356, "step": 582000 }, { "epoch": 3.509364334728235, "eval_loss": 5.315556526184082, "eval_runtime": 39.2768, "eval_samples_per_second": 13.036, "eval_steps_per_second": 6.518, "eval_tts_loss": 7.535357301329812, "step": 582000 }, { "epoch": 3.5096658265095693, "grad_norm": 2.186310052871704, "learning_rate": 0.000149958488111186, "loss": 4.7943, "step": 582050 }, { "epoch": 3.5099673182909035, "grad_norm": 1.9533696174621582, "learning_rate": 0.00014993813914704506, "loss": 4.9306, "step": 582100 }, { "epoch": 3.5102688100722377, "grad_norm": 0.9213695526123047, "learning_rate": 0.00014991779018404247, "loss": 4.8851, "step": 582150 }, { "epoch": 3.5105703018535714, "grad_norm": 0.7110292911529541, "learning_rate": 0.0001498974412225529, "loss": 4.8629, "step": 582200 }, { "epoch": 3.5108717936349056, "grad_norm": 3.587799072265625, "learning_rate": 0.0001498770922629508, "loss": 5.3815, "step": 582250 }, { "epoch": 3.5111732854162394, "grad_norm": 2.3780040740966797, "learning_rate": 0.00014985674330561058, "loss": 5.0025, "step": 582300 }, { "epoch": 3.5114747771975736, "grad_norm": 2.733018636703491, "learning_rate": 0.0001498363943509068, "loss": 4.3316, "step": 582350 }, { "epoch": 3.511776268978908, "grad_norm": 1.4952195882797241, "learning_rate": 0.00014981604539921403, "loss": 4.7756, "step": 582400 }, { "epoch": 3.5120777607602416, "grad_norm": 2.236062526702881, "learning_rate": 0.00014979569645090661, "loss": 5.016, "step": 582450 }, { "epoch": 3.512379252541576, "grad_norm": 2.163412570953369, "learning_rate": 0.0001497753475063591, "loss": 4.8231, "step": 582500 }, { "epoch": 3.5126807443229096, "grad_norm": 1.915449857711792, "learning_rate": 0.00014975499856594607, "loss": 5.0155, "step": 582550 }, { "epoch": 3.5129822361042438, "grad_norm": 2.8328640460968018, "learning_rate": 0.0001497346496300419, "loss": 4.5826, "step": 582600 }, { "epoch": 3.513283727885578, "grad_norm": 1.9472756385803223, "learning_rate": 0.00014971430069902106, "loss": 4.9686, "step": 582650 }, { "epoch": 3.5135852196669117, "grad_norm": 2.386547088623047, "learning_rate": 0.00014969395177325822, "loss": 4.9171, "step": 582700 }, { "epoch": 3.513886711448246, "grad_norm": 2.0430850982666016, "learning_rate": 0.00014967360285312765, "loss": 5.0042, "step": 582750 }, { "epoch": 3.5141882032295797, "grad_norm": 2.0655834674835205, "learning_rate": 0.00014965325393900402, "loss": 4.4321, "step": 582800 }, { "epoch": 3.514489695010914, "grad_norm": 1.9036530256271362, "learning_rate": 0.00014963290503126179, "loss": 4.7778, "step": 582850 }, { "epoch": 3.514791186792248, "grad_norm": 2.77401065826416, "learning_rate": 0.00014961255613027535, "loss": 4.7268, "step": 582900 }, { "epoch": 3.5150926785735823, "grad_norm": 1.921255111694336, "learning_rate": 0.00014959220723641926, "loss": 5.1061, "step": 582950 }, { "epoch": 3.515394170354916, "grad_norm": 1.7756452560424805, "learning_rate": 0.00014957185835006806, "loss": 4.9176, "step": 583000 }, { "epoch": 3.5156956621362503, "grad_norm": 2.7662999629974365, "learning_rate": 0.0001495515094715962, "loss": 4.7002, "step": 583050 }, { "epoch": 3.515997153917584, "grad_norm": 2.774881362915039, "learning_rate": 0.00014953116060137812, "loss": 4.3747, "step": 583100 }, { "epoch": 3.5162986456989183, "grad_norm": 1.9435803890228271, "learning_rate": 0.0001495108117397884, "loss": 5.0617, "step": 583150 }, { "epoch": 3.5166001374802525, "grad_norm": 1.734418511390686, "learning_rate": 0.00014949046288720148, "loss": 4.7341, "step": 583200 }, { "epoch": 3.5169016292615862, "grad_norm": 2.4042181968688965, "learning_rate": 0.00014947011404399182, "loss": 5.2432, "step": 583250 }, { "epoch": 3.5172031210429204, "grad_norm": 2.2342159748077393, "learning_rate": 0.00014944976521053406, "loss": 4.7706, "step": 583300 }, { "epoch": 3.517504612824254, "grad_norm": 2.6999337673187256, "learning_rate": 0.0001494294163872025, "loss": 5.002, "step": 583350 }, { "epoch": 3.5178061046055884, "grad_norm": 2.7999119758605957, "learning_rate": 0.00014940906757437172, "loss": 5.0811, "step": 583400 }, { "epoch": 3.5181075963869226, "grad_norm": 2.4914944171905518, "learning_rate": 0.00014938871877241626, "loss": 5.1312, "step": 583450 }, { "epoch": 3.5184090881682564, "grad_norm": 1.6652673482894897, "learning_rate": 0.0001493683699817105, "loss": 4.7445, "step": 583500 }, { "epoch": 3.5187105799495906, "grad_norm": 1.934350609779358, "learning_rate": 0.00014934802120262897, "loss": 5.1344, "step": 583550 }, { "epoch": 3.5190120717309243, "grad_norm": 2.789064407348633, "learning_rate": 0.00014932767243554625, "loss": 5.3437, "step": 583600 }, { "epoch": 3.5193135635122585, "grad_norm": 1.8720450401306152, "learning_rate": 0.00014930732368083674, "loss": 4.7895, "step": 583650 }, { "epoch": 3.5196150552935928, "grad_norm": 2.311894416809082, "learning_rate": 0.00014928697493887493, "loss": 4.5395, "step": 583700 }, { "epoch": 3.519916547074927, "grad_norm": 1.8806508779525757, "learning_rate": 0.00014926662621003535, "loss": 4.8353, "step": 583750 }, { "epoch": 3.5202180388562607, "grad_norm": 2.6782734394073486, "learning_rate": 0.00014924627749469247, "loss": 4.5649, "step": 583800 }, { "epoch": 3.520519530637595, "grad_norm": 2.5694050788879395, "learning_rate": 0.00014922592879322073, "loss": 4.4226, "step": 583850 }, { "epoch": 3.5208210224189287, "grad_norm": 1.474596381187439, "learning_rate": 0.00014920558010599475, "loss": 4.9583, "step": 583900 }, { "epoch": 3.521122514200263, "grad_norm": 2.608254909515381, "learning_rate": 0.00014918523143338884, "loss": 5.0515, "step": 583950 }, { "epoch": 3.521424005981597, "grad_norm": 1.7172390222549438, "learning_rate": 0.00014916488277577766, "loss": 4.7008, "step": 584000 }, { "epoch": 3.521725497762931, "grad_norm": 2.3860578536987305, "learning_rate": 0.00014914453413353553, "loss": 5.3346, "step": 584050 }, { "epoch": 3.522026989544265, "grad_norm": 2.2115120887756348, "learning_rate": 0.00014912418550703713, "loss": 4.9928, "step": 584100 }, { "epoch": 3.522328481325599, "grad_norm": 2.7306976318359375, "learning_rate": 0.00014910383689665678, "loss": 4.7613, "step": 584150 }, { "epoch": 3.522629973106933, "grad_norm": 2.302534341812134, "learning_rate": 0.00014908348830276906, "loss": 5.3762, "step": 584200 }, { "epoch": 3.5229314648882673, "grad_norm": 2.5534732341766357, "learning_rate": 0.00014906313972574847, "loss": 4.7562, "step": 584250 }, { "epoch": 3.5232329566696015, "grad_norm": 2.858684539794922, "learning_rate": 0.00014904279116596938, "loss": 4.5812, "step": 584300 }, { "epoch": 3.523534448450935, "grad_norm": 3.376953601837158, "learning_rate": 0.00014902244262380638, "loss": 4.7161, "step": 584350 }, { "epoch": 3.5238359402322694, "grad_norm": 1.0422641038894653, "learning_rate": 0.00014900209409963395, "loss": 5.107, "step": 584400 }, { "epoch": 3.524137432013603, "grad_norm": 2.358736515045166, "learning_rate": 0.00014898174559382652, "loss": 4.6596, "step": 584450 }, { "epoch": 3.5244389237949374, "grad_norm": 1.8622541427612305, "learning_rate": 0.0001489613971067586, "loss": 4.6687, "step": 584500 }, { "epoch": 3.5247404155762716, "grad_norm": 2.0174319744110107, "learning_rate": 0.00014894104863880474, "loss": 4.9662, "step": 584550 }, { "epoch": 3.5250419073576054, "grad_norm": 1.5807031393051147, "learning_rate": 0.00014892070019033936, "loss": 4.7052, "step": 584600 }, { "epoch": 3.5253433991389396, "grad_norm": 3.4832005500793457, "learning_rate": 0.00014890035176173689, "loss": 4.8355, "step": 584650 }, { "epoch": 3.5256448909202733, "grad_norm": 2.4302284717559814, "learning_rate": 0.00014888000335337197, "loss": 4.8316, "step": 584700 }, { "epoch": 3.5259463827016075, "grad_norm": 2.9356184005737305, "learning_rate": 0.0001488596549656189, "loss": 4.5711, "step": 584750 }, { "epoch": 3.5262478744829417, "grad_norm": 2.3666346073150635, "learning_rate": 0.0001488393065988523, "loss": 4.6045, "step": 584800 }, { "epoch": 3.5265493662642755, "grad_norm": 2.234927177429199, "learning_rate": 0.00014881895825344664, "loss": 5.0028, "step": 584850 }, { "epoch": 3.5268508580456097, "grad_norm": 3.038342237472534, "learning_rate": 0.0001487986099297763, "loss": 5.3313, "step": 584900 }, { "epoch": 3.5271523498269435, "grad_norm": 3.0568647384643555, "learning_rate": 0.00014877826162821583, "loss": 4.7499, "step": 584950 }, { "epoch": 3.5274538416082777, "grad_norm": 1.6828134059906006, "learning_rate": 0.0001487579133491398, "loss": 4.8951, "step": 585000 }, { "epoch": 3.5274538416082777, "eval_loss": 5.292641639709473, "eval_runtime": 38.9963, "eval_samples_per_second": 13.129, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.546426512844554, "step": 585000 }, { "epoch": 3.527755333389612, "grad_norm": 1.9554404020309448, "learning_rate": 0.00014873756509292248, "loss": 4.9401, "step": 585050 }, { "epoch": 3.528056825170946, "grad_norm": 2.523483991622925, "learning_rate": 0.00014871721685993851, "loss": 4.8787, "step": 585100 }, { "epoch": 3.52835831695228, "grad_norm": 1.2545679807662964, "learning_rate": 0.0001486968686505624, "loss": 4.4755, "step": 585150 }, { "epoch": 3.528659808733614, "grad_norm": 1.611392617225647, "learning_rate": 0.00014867652046516853, "loss": 4.8276, "step": 585200 }, { "epoch": 3.528961300514948, "grad_norm": 2.2624664306640625, "learning_rate": 0.00014865617230413135, "loss": 4.7611, "step": 585250 }, { "epoch": 3.529262792296282, "grad_norm": 2.0825960636138916, "learning_rate": 0.0001486358241678255, "loss": 5.1203, "step": 585300 }, { "epoch": 3.5295642840776162, "grad_norm": 2.5580177307128906, "learning_rate": 0.00014861547605662527, "loss": 5.4602, "step": 585350 }, { "epoch": 3.52986577585895, "grad_norm": 2.4732635021209717, "learning_rate": 0.00014859512797090526, "loss": 4.4744, "step": 585400 }, { "epoch": 3.530167267640284, "grad_norm": 1.9317348003387451, "learning_rate": 0.00014857477991103997, "loss": 5.062, "step": 585450 }, { "epoch": 3.530468759421618, "grad_norm": 2.3282268047332764, "learning_rate": 0.00014855443187740375, "loss": 5.1413, "step": 585500 }, { "epoch": 3.530770251202952, "grad_norm": 1.963469386100769, "learning_rate": 0.00014853408387037118, "loss": 4.8588, "step": 585550 }, { "epoch": 3.5310717429842864, "grad_norm": 2.3162667751312256, "learning_rate": 0.0001485137358903167, "loss": 4.4813, "step": 585600 }, { "epoch": 3.53137323476562, "grad_norm": 2.61655330657959, "learning_rate": 0.00014849338793761476, "loss": 4.3549, "step": 585650 }, { "epoch": 3.5316747265469544, "grad_norm": 2.574172258377075, "learning_rate": 0.0001484730400126399, "loss": 4.6797, "step": 585700 }, { "epoch": 3.531976218328288, "grad_norm": 2.926191806793213, "learning_rate": 0.00014845269211576656, "loss": 4.9977, "step": 585750 }, { "epoch": 3.5322777101096223, "grad_norm": 2.373305559158325, "learning_rate": 0.0001484323442473692, "loss": 5.2181, "step": 585800 }, { "epoch": 3.5325792018909565, "grad_norm": 0.9993284344673157, "learning_rate": 0.00014841199640782229, "loss": 4.4094, "step": 585850 }, { "epoch": 3.5328806936722907, "grad_norm": 1.9550172090530396, "learning_rate": 0.00014839164859750038, "loss": 5.1256, "step": 585900 }, { "epoch": 3.5331821854536245, "grad_norm": 2.8455491065979004, "learning_rate": 0.0001483713008167778, "loss": 5.041, "step": 585950 }, { "epoch": 3.5334836772349587, "grad_norm": 2.8207404613494873, "learning_rate": 0.00014835095306602917, "loss": 5.0816, "step": 586000 }, { "epoch": 3.5337851690162925, "grad_norm": 2.8025553226470947, "learning_rate": 0.00014833060534562891, "loss": 4.8139, "step": 586050 }, { "epoch": 3.5340866607976267, "grad_norm": 1.9136674404144287, "learning_rate": 0.0001483102576559514, "loss": 4.6799, "step": 586100 }, { "epoch": 3.534388152578961, "grad_norm": 1.8283613920211792, "learning_rate": 0.00014828990999737123, "loss": 4.5314, "step": 586150 }, { "epoch": 3.5346896443602946, "grad_norm": 1.5980361700057983, "learning_rate": 0.00014826956237026282, "loss": 5.0298, "step": 586200 }, { "epoch": 3.534991136141629, "grad_norm": 1.3718065023422241, "learning_rate": 0.00014824921477500066, "loss": 4.7283, "step": 586250 }, { "epoch": 3.5352926279229626, "grad_norm": 1.1840007305145264, "learning_rate": 0.00014822886721195915, "loss": 4.6458, "step": 586300 }, { "epoch": 3.535594119704297, "grad_norm": 2.2251603603363037, "learning_rate": 0.0001482085196815129, "loss": 4.8012, "step": 586350 }, { "epoch": 3.535895611485631, "grad_norm": 2.3975448608398438, "learning_rate": 0.00014818817218403623, "loss": 4.964, "step": 586400 }, { "epoch": 3.536197103266965, "grad_norm": 3.205742835998535, "learning_rate": 0.00014816782471990366, "loss": 5.1428, "step": 586450 }, { "epoch": 3.536498595048299, "grad_norm": 1.9458644390106201, "learning_rate": 0.00014814747728948973, "loss": 5.2238, "step": 586500 }, { "epoch": 3.5368000868296328, "grad_norm": 1.678775429725647, "learning_rate": 0.00014812712989316874, "loss": 5.0088, "step": 586550 }, { "epoch": 3.537101578610967, "grad_norm": 2.741048574447632, "learning_rate": 0.0001481067825313153, "loss": 4.8089, "step": 586600 }, { "epoch": 3.537403070392301, "grad_norm": 1.1829322576522827, "learning_rate": 0.00014808643520430385, "loss": 4.8608, "step": 586650 }, { "epoch": 3.5377045621736354, "grad_norm": 2.4303362369537354, "learning_rate": 0.0001480660879125088, "loss": 5.1968, "step": 586700 }, { "epoch": 3.538006053954969, "grad_norm": 1.6171740293502808, "learning_rate": 0.0001480457406563046, "loss": 4.6445, "step": 586750 }, { "epoch": 3.5383075457363034, "grad_norm": 1.9516698122024536, "learning_rate": 0.00014802539343606582, "loss": 4.9165, "step": 586800 }, { "epoch": 3.538609037517637, "grad_norm": 1.709934949874878, "learning_rate": 0.00014800504625216683, "loss": 4.678, "step": 586850 }, { "epoch": 3.5389105292989713, "grad_norm": 2.137068748474121, "learning_rate": 0.0001479846991049821, "loss": 4.8694, "step": 586900 }, { "epoch": 3.5392120210803055, "grad_norm": 1.8672304153442383, "learning_rate": 0.00014796435199488617, "loss": 4.6554, "step": 586950 }, { "epoch": 3.5395135128616393, "grad_norm": 2.9037559032440186, "learning_rate": 0.0001479440049222534, "loss": 4.8965, "step": 587000 }, { "epoch": 3.5398150046429735, "grad_norm": 2.132246732711792, "learning_rate": 0.00014792365788745824, "loss": 4.633, "step": 587050 }, { "epoch": 3.5401164964243073, "grad_norm": 1.9357792139053345, "learning_rate": 0.00014790331089087526, "loss": 5.2259, "step": 587100 }, { "epoch": 3.5404179882056415, "grad_norm": 1.6973944902420044, "learning_rate": 0.0001478829639328788, "loss": 5.0797, "step": 587150 }, { "epoch": 3.5407194799869757, "grad_norm": 1.6902981996536255, "learning_rate": 0.00014786261701384338, "loss": 4.6459, "step": 587200 }, { "epoch": 3.5410209717683094, "grad_norm": 1.648021936416626, "learning_rate": 0.00014784227013414351, "loss": 4.1165, "step": 587250 }, { "epoch": 3.5413224635496436, "grad_norm": 1.117089867591858, "learning_rate": 0.00014782192329415347, "loss": 4.7277, "step": 587300 }, { "epoch": 3.5416239553309774, "grad_norm": 2.829540729522705, "learning_rate": 0.00014780157649424785, "loss": 5.1768, "step": 587350 }, { "epoch": 3.5419254471123116, "grad_norm": 2.700181484222412, "learning_rate": 0.00014778122973480113, "loss": 5.0388, "step": 587400 }, { "epoch": 3.542226938893646, "grad_norm": 0.5830956697463989, "learning_rate": 0.0001477608830161877, "loss": 4.8833, "step": 587450 }, { "epoch": 3.54252843067498, "grad_norm": 2.441335916519165, "learning_rate": 0.000147740536338782, "loss": 4.578, "step": 587500 }, { "epoch": 3.542829922456314, "grad_norm": 2.9304208755493164, "learning_rate": 0.0001477201897029585, "loss": 5.0608, "step": 587550 }, { "epoch": 3.543131414237648, "grad_norm": 1.7963401079177856, "learning_rate": 0.0001476998431090917, "loss": 5.2715, "step": 587600 }, { "epoch": 3.5434329060189818, "grad_norm": 1.9483975172042847, "learning_rate": 0.00014767949655755593, "loss": 4.7893, "step": 587650 }, { "epoch": 3.543734397800316, "grad_norm": 1.7366589307785034, "learning_rate": 0.00014765915004872574, "loss": 5.0807, "step": 587700 }, { "epoch": 3.54403588958165, "grad_norm": 2.1981372833251953, "learning_rate": 0.00014763880358297562, "loss": 4.8599, "step": 587750 }, { "epoch": 3.544337381362984, "grad_norm": 2.6423704624176025, "learning_rate": 0.00014761845716067989, "loss": 4.7538, "step": 587800 }, { "epoch": 3.544638873144318, "grad_norm": 2.1392908096313477, "learning_rate": 0.00014759811078221305, "loss": 4.708, "step": 587850 }, { "epoch": 3.544940364925652, "grad_norm": 3.5971155166625977, "learning_rate": 0.0001475777644479496, "loss": 4.7951, "step": 587900 }, { "epoch": 3.545241856706986, "grad_norm": 1.9217965602874756, "learning_rate": 0.00014755741815826386, "loss": 4.7798, "step": 587950 }, { "epoch": 3.5455433484883203, "grad_norm": 1.2872475385665894, "learning_rate": 0.0001475370719135304, "loss": 4.7663, "step": 588000 }, { "epoch": 3.5455433484883203, "eval_loss": 5.293297290802002, "eval_runtime": 39.0672, "eval_samples_per_second": 13.106, "eval_steps_per_second": 6.553, "eval_tts_loss": 7.501764914603489, "step": 588000 }, { "epoch": 3.5458448402696545, "grad_norm": 2.2725648880004883, "learning_rate": 0.00014751672571412368, "loss": 5.1934, "step": 588050 }, { "epoch": 3.5461463320509883, "grad_norm": 2.0366477966308594, "learning_rate": 0.00014749637956041798, "loss": 4.7712, "step": 588100 }, { "epoch": 3.5464478238323225, "grad_norm": 2.799604892730713, "learning_rate": 0.00014747603345278788, "loss": 4.4207, "step": 588150 }, { "epoch": 3.5467493156136563, "grad_norm": 2.324998140335083, "learning_rate": 0.00014745568739160783, "loss": 4.8361, "step": 588200 }, { "epoch": 3.5470508073949905, "grad_norm": 1.7282911539077759, "learning_rate": 0.00014743534137725214, "loss": 4.7294, "step": 588250 }, { "epoch": 3.5473522991763247, "grad_norm": 3.656815767288208, "learning_rate": 0.00014741499541009534, "loss": 5.1531, "step": 588300 }, { "epoch": 3.5476537909576584, "grad_norm": 2.943812370300293, "learning_rate": 0.0001473946494905119, "loss": 5.0576, "step": 588350 }, { "epoch": 3.5479552827389926, "grad_norm": 1.7926597595214844, "learning_rate": 0.00014737430361887622, "loss": 4.6779, "step": 588400 }, { "epoch": 3.5482567745203264, "grad_norm": 1.9886480569839478, "learning_rate": 0.0001473539577955627, "loss": 5.0408, "step": 588450 }, { "epoch": 3.5485582663016606, "grad_norm": 2.193214178085327, "learning_rate": 0.00014733361202094587, "loss": 4.5888, "step": 588500 }, { "epoch": 3.548859758082995, "grad_norm": 1.1437623500823975, "learning_rate": 0.00014731326629540003, "loss": 4.7796, "step": 588550 }, { "epoch": 3.5491612498643286, "grad_norm": 2.051604747772217, "learning_rate": 0.00014729292061929972, "loss": 4.9284, "step": 588600 }, { "epoch": 3.5494627416456628, "grad_norm": 1.8392221927642822, "learning_rate": 0.0001472725749930194, "loss": 4.914, "step": 588650 }, { "epoch": 3.5497642334269965, "grad_norm": 1.9206405878067017, "learning_rate": 0.00014725222941693335, "loss": 5.2824, "step": 588700 }, { "epoch": 3.5500657252083307, "grad_norm": 0.7040712237358093, "learning_rate": 0.00014723188389141614, "loss": 4.6729, "step": 588750 }, { "epoch": 3.550367216989665, "grad_norm": 1.5431699752807617, "learning_rate": 0.0001472115384168422, "loss": 4.4779, "step": 588800 }, { "epoch": 3.550668708770999, "grad_norm": 2.5360491275787354, "learning_rate": 0.00014719119299358584, "loss": 4.6501, "step": 588850 }, { "epoch": 3.550970200552333, "grad_norm": 2.125086784362793, "learning_rate": 0.00014717084762202158, "loss": 4.8352, "step": 588900 }, { "epoch": 3.551271692333667, "grad_norm": 2.096287488937378, "learning_rate": 0.00014715050230252387, "loss": 5.0282, "step": 588950 }, { "epoch": 3.551573184115001, "grad_norm": 2.071479082107544, "learning_rate": 0.00014713015703546708, "loss": 5.1506, "step": 589000 }, { "epoch": 3.551874675896335, "grad_norm": 1.860582709312439, "learning_rate": 0.00014710981182122562, "loss": 5.1505, "step": 589050 }, { "epoch": 3.5521761676776693, "grad_norm": 1.6456050872802734, "learning_rate": 0.00014708946666017403, "loss": 4.2167, "step": 589100 }, { "epoch": 3.552477659459003, "grad_norm": 3.1462771892547607, "learning_rate": 0.00014706912155268656, "loss": 4.4239, "step": 589150 }, { "epoch": 3.5527791512403373, "grad_norm": 2.936295986175537, "learning_rate": 0.00014704877649913777, "loss": 5.1353, "step": 589200 }, { "epoch": 3.553080643021671, "grad_norm": 2.474255084991455, "learning_rate": 0.00014702843149990208, "loss": 4.7046, "step": 589250 }, { "epoch": 3.5533821348030052, "grad_norm": 1.9049694538116455, "learning_rate": 0.0001470080865553538, "loss": 4.8285, "step": 589300 }, { "epoch": 3.5536836265843395, "grad_norm": 1.835187554359436, "learning_rate": 0.00014698774166586743, "loss": 5.0272, "step": 589350 }, { "epoch": 3.553985118365673, "grad_norm": 1.5294429063796997, "learning_rate": 0.00014696739683181742, "loss": 4.7622, "step": 589400 }, { "epoch": 3.5542866101470074, "grad_norm": 2.287743330001831, "learning_rate": 0.00014694705205357808, "loss": 4.7868, "step": 589450 }, { "epoch": 3.554588101928341, "grad_norm": 0.8648072481155396, "learning_rate": 0.00014692670733152391, "loss": 5.1087, "step": 589500 }, { "epoch": 3.5548895937096754, "grad_norm": 2.318251848220825, "learning_rate": 0.00014690636266602935, "loss": 4.9742, "step": 589550 }, { "epoch": 3.5551910854910096, "grad_norm": 2.4218435287475586, "learning_rate": 0.00014688601805746873, "loss": 5.0634, "step": 589600 }, { "epoch": 3.555492577272344, "grad_norm": 2.580094575881958, "learning_rate": 0.0001468656735062165, "loss": 5.0595, "step": 589650 }, { "epoch": 3.5557940690536776, "grad_norm": 2.2053210735321045, "learning_rate": 0.0001468453290126471, "loss": 5.1223, "step": 589700 }, { "epoch": 3.5560955608350118, "grad_norm": 2.0307114124298096, "learning_rate": 0.00014682498457713487, "loss": 4.4889, "step": 589750 }, { "epoch": 3.5563970526163455, "grad_norm": 1.6455453634262085, "learning_rate": 0.0001468046402000543, "loss": 4.6983, "step": 589800 }, { "epoch": 3.5566985443976797, "grad_norm": 1.9251114130020142, "learning_rate": 0.0001467842958817798, "loss": 4.5374, "step": 589850 }, { "epoch": 3.557000036179014, "grad_norm": 2.3121604919433594, "learning_rate": 0.0001467639516226857, "loss": 4.6927, "step": 589900 }, { "epoch": 3.5573015279603477, "grad_norm": 1.8089585304260254, "learning_rate": 0.00014674360742314646, "loss": 4.8285, "step": 589950 }, { "epoch": 3.557603019741682, "grad_norm": 3.396347999572754, "learning_rate": 0.00014672326328353652, "loss": 4.9109, "step": 590000 }, { "epoch": 3.5579045115230157, "grad_norm": 2.3880879878997803, "learning_rate": 0.00014670291920423017, "loss": 4.5303, "step": 590050 }, { "epoch": 3.55820600330435, "grad_norm": 1.505596399307251, "learning_rate": 0.00014668257518560192, "loss": 5.0967, "step": 590100 }, { "epoch": 3.558507495085684, "grad_norm": 3.0262258052825928, "learning_rate": 0.00014666223122802616, "loss": 5.1032, "step": 590150 }, { "epoch": 3.558808986867018, "grad_norm": 2.1355385780334473, "learning_rate": 0.00014664188733187725, "loss": 4.6533, "step": 590200 }, { "epoch": 3.559110478648352, "grad_norm": 1.8614269495010376, "learning_rate": 0.0001466215434975296, "loss": 4.4635, "step": 590250 }, { "epoch": 3.559411970429686, "grad_norm": 2.3803577423095703, "learning_rate": 0.00014660119972535767, "loss": 5.0197, "step": 590300 }, { "epoch": 3.55971346221102, "grad_norm": 2.2773404121398926, "learning_rate": 0.00014658085601573575, "loss": 4.8665, "step": 590350 }, { "epoch": 3.5600149539923542, "grad_norm": 2.883737087249756, "learning_rate": 0.00014656051236903832, "loss": 4.9654, "step": 590400 }, { "epoch": 3.5603164457736884, "grad_norm": 2.0761911869049072, "learning_rate": 0.00014654016878563978, "loss": 4.8443, "step": 590450 }, { "epoch": 3.560617937555022, "grad_norm": 3.11447811126709, "learning_rate": 0.00014651982526591445, "loss": 4.9144, "step": 590500 }, { "epoch": 3.5609194293363564, "grad_norm": 2.6644270420074463, "learning_rate": 0.0001464994818102368, "loss": 4.8668, "step": 590550 }, { "epoch": 3.56122092111769, "grad_norm": 2.29988956451416, "learning_rate": 0.00014647913841898117, "loss": 4.8287, "step": 590600 }, { "epoch": 3.5615224128990244, "grad_norm": 2.156928300857544, "learning_rate": 0.00014645879509252199, "loss": 4.9877, "step": 590650 }, { "epoch": 3.5618239046803586, "grad_norm": 2.4187369346618652, "learning_rate": 0.0001464384518312336, "loss": 4.6533, "step": 590700 }, { "epoch": 3.5621253964616924, "grad_norm": 2.23879075050354, "learning_rate": 0.00014641810863549043, "loss": 4.8006, "step": 590750 }, { "epoch": 3.5624268882430266, "grad_norm": 2.391329526901245, "learning_rate": 0.00014639776550566693, "loss": 4.8128, "step": 590800 }, { "epoch": 3.5627283800243603, "grad_norm": 3.451347589492798, "learning_rate": 0.0001463774224421373, "loss": 4.5578, "step": 590850 }, { "epoch": 3.5630298718056945, "grad_norm": 0.9204000234603882, "learning_rate": 0.0001463570794452761, "loss": 4.6867, "step": 590900 }, { "epoch": 3.5633313635870287, "grad_norm": 2.1447324752807617, "learning_rate": 0.00014633673651545767, "loss": 5.0547, "step": 590950 }, { "epoch": 3.5636328553683625, "grad_norm": 1.6165066957473755, "learning_rate": 0.00014631639365305637, "loss": 4.9388, "step": 591000 }, { "epoch": 3.5636328553683625, "eval_loss": 5.286226272583008, "eval_runtime": 38.9539, "eval_samples_per_second": 13.144, "eval_steps_per_second": 6.572, "eval_tts_loss": 7.556283968062257, "step": 591000 }, { "epoch": 3.5639343471496967, "grad_norm": 1.7222257852554321, "learning_rate": 0.00014629605085844657, "loss": 5.0469, "step": 591050 }, { "epoch": 3.5642358389310305, "grad_norm": 2.408935308456421, "learning_rate": 0.0001462757081320027, "loss": 4.6721, "step": 591100 }, { "epoch": 3.5645373307123647, "grad_norm": 3.0726850032806396, "learning_rate": 0.0001462553654740991, "loss": 4.6903, "step": 591150 }, { "epoch": 3.564838822493699, "grad_norm": 2.2374045848846436, "learning_rate": 0.0001462350228851101, "loss": 5.1722, "step": 591200 }, { "epoch": 3.565140314275033, "grad_norm": 2.8220012187957764, "learning_rate": 0.0001462146803654102, "loss": 5.115, "step": 591250 }, { "epoch": 3.565441806056367, "grad_norm": 1.8557398319244385, "learning_rate": 0.00014619433791537368, "loss": 5.0858, "step": 591300 }, { "epoch": 3.565743297837701, "grad_norm": 3.2859201431274414, "learning_rate": 0.0001461739955353749, "loss": 4.8329, "step": 591350 }, { "epoch": 3.566044789619035, "grad_norm": 2.5226194858551025, "learning_rate": 0.00014615365322578837, "loss": 4.5077, "step": 591400 }, { "epoch": 3.566346281400369, "grad_norm": 2.364335060119629, "learning_rate": 0.00014613331098698825, "loss": 5.1908, "step": 591450 }, { "epoch": 3.5666477731817032, "grad_norm": 2.0489284992218018, "learning_rate": 0.00014611296881934905, "loss": 4.9421, "step": 591500 }, { "epoch": 3.566949264963037, "grad_norm": 2.952951669692993, "learning_rate": 0.00014609262672324517, "loss": 4.8037, "step": 591550 }, { "epoch": 3.567250756744371, "grad_norm": 4.700255870819092, "learning_rate": 0.0001460722846990509, "loss": 4.8938, "step": 591600 }, { "epoch": 3.567552248525705, "grad_norm": 0.6532846689224243, "learning_rate": 0.0001460519427471406, "loss": 4.6452, "step": 591650 }, { "epoch": 3.567853740307039, "grad_norm": 2.1284871101379395, "learning_rate": 0.0001460316008678887, "loss": 5.0369, "step": 591700 }, { "epoch": 3.5681552320883734, "grad_norm": 2.717838764190674, "learning_rate": 0.00014601125906166948, "loss": 4.7198, "step": 591750 }, { "epoch": 3.5684567238697076, "grad_norm": 1.3692759275436401, "learning_rate": 0.00014599091732885733, "loss": 4.8424, "step": 591800 }, { "epoch": 3.5687582156510413, "grad_norm": 2.582070827484131, "learning_rate": 0.00014597057566982668, "loss": 5.0088, "step": 591850 }, { "epoch": 3.5690597074323755, "grad_norm": 4.014963150024414, "learning_rate": 0.00014595023408495178, "loss": 5.1772, "step": 591900 }, { "epoch": 3.5693611992137093, "grad_norm": 2.379335403442383, "learning_rate": 0.0001459298925746071, "loss": 4.6918, "step": 591950 }, { "epoch": 3.5696626909950435, "grad_norm": 2.0587756633758545, "learning_rate": 0.00014590955113916694, "loss": 4.821, "step": 592000 }, { "epoch": 3.5699641827763777, "grad_norm": 0.7212017178535461, "learning_rate": 0.0001458892097790056, "loss": 4.92, "step": 592050 }, { "epoch": 3.5702656745577115, "grad_norm": 3.2308385372161865, "learning_rate": 0.00014586886849449746, "loss": 4.4546, "step": 592100 }, { "epoch": 3.5705671663390457, "grad_norm": 1.943210482597351, "learning_rate": 0.000145848527286017, "loss": 4.7201, "step": 592150 }, { "epoch": 3.5708686581203795, "grad_norm": 2.0136613845825195, "learning_rate": 0.0001458281861539384, "loss": 4.7034, "step": 592200 }, { "epoch": 3.5711701499017137, "grad_norm": 1.8767480850219727, "learning_rate": 0.00014580784509863609, "loss": 4.8082, "step": 592250 }, { "epoch": 3.571471641683048, "grad_norm": 2.176225423812866, "learning_rate": 0.00014578750412048446, "loss": 4.7659, "step": 592300 }, { "epoch": 3.5717731334643816, "grad_norm": 1.152917504310608, "learning_rate": 0.00014576716321985773, "loss": 4.892, "step": 592350 }, { "epoch": 3.572074625245716, "grad_norm": 2.349698781967163, "learning_rate": 0.00014574682239713036, "loss": 4.6558, "step": 592400 }, { "epoch": 3.5723761170270496, "grad_norm": 1.5328892469406128, "learning_rate": 0.00014572648165267667, "loss": 4.5149, "step": 592450 }, { "epoch": 3.572677608808384, "grad_norm": 1.1935771703720093, "learning_rate": 0.00014570614098687094, "loss": 4.6032, "step": 592500 }, { "epoch": 3.572979100589718, "grad_norm": 1.8658051490783691, "learning_rate": 0.0001456858004000876, "loss": 4.8883, "step": 592550 }, { "epoch": 3.5732805923710522, "grad_norm": 2.904883861541748, "learning_rate": 0.00014566545989270097, "loss": 4.6999, "step": 592600 }, { "epoch": 3.573582084152386, "grad_norm": 2.0327792167663574, "learning_rate": 0.0001456451194650853, "loss": 4.7758, "step": 592650 }, { "epoch": 3.57388357593372, "grad_norm": 2.3461532592773438, "learning_rate": 0.00014562477911761497, "loss": 4.9903, "step": 592700 }, { "epoch": 3.574185067715054, "grad_norm": 2.2635498046875, "learning_rate": 0.0001456044388506644, "loss": 4.8893, "step": 592750 }, { "epoch": 3.574486559496388, "grad_norm": 2.1691362857818604, "learning_rate": 0.00014558409866460788, "loss": 4.6568, "step": 592800 }, { "epoch": 3.5747880512777224, "grad_norm": 1.6641660928726196, "learning_rate": 0.00014556375855981964, "loss": 4.7059, "step": 592850 }, { "epoch": 3.575089543059056, "grad_norm": 1.8256943225860596, "learning_rate": 0.0001455434185366742, "loss": 4.9197, "step": 592900 }, { "epoch": 3.5753910348403903, "grad_norm": 1.9027435779571533, "learning_rate": 0.00014552307859554568, "loss": 4.8168, "step": 592950 }, { "epoch": 3.575692526621724, "grad_norm": 2.1607882976531982, "learning_rate": 0.00014550273873680853, "loss": 4.2999, "step": 593000 }, { "epoch": 3.5759940184030583, "grad_norm": 2.41967511177063, "learning_rate": 0.00014548239896083713, "loss": 5.0702, "step": 593050 }, { "epoch": 3.5762955101843925, "grad_norm": 2.0490920543670654, "learning_rate": 0.00014546205926800563, "loss": 4.4838, "step": 593100 }, { "epoch": 3.5765970019657263, "grad_norm": 2.832568883895874, "learning_rate": 0.00014544171965868852, "loss": 5.0278, "step": 593150 }, { "epoch": 3.5768984937470605, "grad_norm": 2.317357301712036, "learning_rate": 0.00014542138013326006, "loss": 4.8738, "step": 593200 }, { "epoch": 3.5771999855283942, "grad_norm": 2.4065561294555664, "learning_rate": 0.00014540104069209452, "loss": 4.804, "step": 593250 }, { "epoch": 3.5775014773097285, "grad_norm": 0.5634401440620422, "learning_rate": 0.00014538070133556625, "loss": 4.581, "step": 593300 }, { "epoch": 3.5778029690910627, "grad_norm": 2.435441255569458, "learning_rate": 0.00014536036206404964, "loss": 5.0018, "step": 593350 }, { "epoch": 3.578104460872397, "grad_norm": 2.447068452835083, "learning_rate": 0.00014534002287791892, "loss": 5.0894, "step": 593400 }, { "epoch": 3.5784059526537306, "grad_norm": 3.2176105976104736, "learning_rate": 0.0001453196837775484, "loss": 4.7965, "step": 593450 }, { "epoch": 3.578707444435065, "grad_norm": 2.4268569946289062, "learning_rate": 0.00014529934476331248, "loss": 5.2524, "step": 593500 }, { "epoch": 3.5790089362163986, "grad_norm": 2.7636184692382812, "learning_rate": 0.00014527900583558535, "loss": 4.7682, "step": 593550 }, { "epoch": 3.579310427997733, "grad_norm": 1.9426261186599731, "learning_rate": 0.0001452586669947414, "loss": 4.6602, "step": 593600 }, { "epoch": 3.579611919779067, "grad_norm": 1.681286334991455, "learning_rate": 0.00014523832824115497, "loss": 5.0273, "step": 593650 }, { "epoch": 3.5799134115604008, "grad_norm": 1.9988915920257568, "learning_rate": 0.00014521798957520024, "loss": 4.9722, "step": 593700 }, { "epoch": 3.580214903341735, "grad_norm": 4.988100051879883, "learning_rate": 0.00014519765099725162, "loss": 4.9223, "step": 593750 }, { "epoch": 3.5805163951230687, "grad_norm": 2.805989980697632, "learning_rate": 0.0001451773125076834, "loss": 5.1473, "step": 593800 }, { "epoch": 3.580817886904403, "grad_norm": 2.0151398181915283, "learning_rate": 0.00014515697410686984, "loss": 5.0261, "step": 593850 }, { "epoch": 3.581119378685737, "grad_norm": 2.456751585006714, "learning_rate": 0.00014513663579518522, "loss": 4.7648, "step": 593900 }, { "epoch": 3.581420870467071, "grad_norm": 2.522130250930786, "learning_rate": 0.00014511629757300393, "loss": 4.892, "step": 593950 }, { "epoch": 3.581722362248405, "grad_norm": 1.8926970958709717, "learning_rate": 0.0001450959594407002, "loss": 5.1051, "step": 594000 }, { "epoch": 3.581722362248405, "eval_loss": 5.278932571411133, "eval_runtime": 39.1767, "eval_samples_per_second": 13.069, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.530736580674174, "step": 594000 }, { "epoch": 3.582023854029739, "grad_norm": 1.7361359596252441, "learning_rate": 0.00014507562139864832, "loss": 5.328, "step": 594050 }, { "epoch": 3.582325345811073, "grad_norm": 1.9838130474090576, "learning_rate": 0.0001450552834472226, "loss": 4.7079, "step": 594100 }, { "epoch": 3.5826268375924073, "grad_norm": 2.221719741821289, "learning_rate": 0.00014503494558679736, "loss": 4.7205, "step": 594150 }, { "epoch": 3.5829283293737415, "grad_norm": 1.8674010038375854, "learning_rate": 0.00014501460781774688, "loss": 5.0004, "step": 594200 }, { "epoch": 3.5832298211550753, "grad_norm": 1.703411340713501, "learning_rate": 0.00014499427014044537, "loss": 5.2179, "step": 594250 }, { "epoch": 3.5835313129364095, "grad_norm": 2.3063225746154785, "learning_rate": 0.00014497393255526723, "loss": 4.3782, "step": 594300 }, { "epoch": 3.5838328047177432, "grad_norm": 1.7161856889724731, "learning_rate": 0.00014495359506258668, "loss": 4.5172, "step": 594350 }, { "epoch": 3.5841342964990774, "grad_norm": 2.424853801727295, "learning_rate": 0.00014493325766277794, "loss": 5.1195, "step": 594400 }, { "epoch": 3.5844357882804116, "grad_norm": 2.5186071395874023, "learning_rate": 0.00014491292035621547, "loss": 5.111, "step": 594450 }, { "epoch": 3.5847372800617454, "grad_norm": 2.522350311279297, "learning_rate": 0.00014489258314327334, "loss": 4.6709, "step": 594500 }, { "epoch": 3.5850387718430796, "grad_norm": 1.8913377523422241, "learning_rate": 0.00014487224602432598, "loss": 5.1165, "step": 594550 }, { "epoch": 3.5853402636244134, "grad_norm": 0.8649100661277771, "learning_rate": 0.00014485190899974764, "loss": 4.5804, "step": 594600 }, { "epoch": 3.5856417554057476, "grad_norm": 2.8083505630493164, "learning_rate": 0.00014483157206991251, "loss": 4.7938, "step": 594650 }, { "epoch": 3.585943247187082, "grad_norm": 2.6741786003112793, "learning_rate": 0.0001448112352351949, "loss": 4.9333, "step": 594700 }, { "epoch": 3.5862447389684156, "grad_norm": 3.7708933353424072, "learning_rate": 0.00014479089849596922, "loss": 4.7353, "step": 594750 }, { "epoch": 3.5865462307497498, "grad_norm": 1.7459726333618164, "learning_rate": 0.00014477056185260956, "loss": 4.9243, "step": 594800 }, { "epoch": 3.5868477225310835, "grad_norm": 2.276808023452759, "learning_rate": 0.0001447502253054902, "loss": 4.6454, "step": 594850 }, { "epoch": 3.5871492143124177, "grad_norm": 2.543083429336548, "learning_rate": 0.0001447298888549855, "loss": 4.7143, "step": 594900 }, { "epoch": 3.587450706093752, "grad_norm": 1.6790003776550293, "learning_rate": 0.00014470955250146968, "loss": 4.6155, "step": 594950 }, { "epoch": 3.587752197875086, "grad_norm": 2.0825860500335693, "learning_rate": 0.00014468921624531696, "loss": 5.4018, "step": 595000 }, { "epoch": 3.58805368965642, "grad_norm": 1.8176888227462769, "learning_rate": 0.00014466888008690166, "loss": 4.7816, "step": 595050 }, { "epoch": 3.588355181437754, "grad_norm": 1.3987303972244263, "learning_rate": 0.00014464854402659798, "loss": 4.8769, "step": 595100 }, { "epoch": 3.588656673219088, "grad_norm": 2.1469197273254395, "learning_rate": 0.00014462820806478025, "loss": 4.8649, "step": 595150 }, { "epoch": 3.588958165000422, "grad_norm": 1.6747970581054688, "learning_rate": 0.00014460787220182272, "loss": 4.4833, "step": 595200 }, { "epoch": 3.5892596567817563, "grad_norm": 0.7121274471282959, "learning_rate": 0.0001445875364380995, "loss": 4.4599, "step": 595250 }, { "epoch": 3.58956114856309, "grad_norm": 2.045389175415039, "learning_rate": 0.00014456720077398502, "loss": 4.8762, "step": 595300 }, { "epoch": 3.5898626403444243, "grad_norm": 2.9963510036468506, "learning_rate": 0.00014454686520985345, "loss": 4.6484, "step": 595350 }, { "epoch": 3.590164132125758, "grad_norm": 2.863081455230713, "learning_rate": 0.00014452652974607905, "loss": 4.3813, "step": 595400 }, { "epoch": 3.5904656239070922, "grad_norm": 1.905411720275879, "learning_rate": 0.000144506194383036, "loss": 4.701, "step": 595450 }, { "epoch": 3.5907671156884264, "grad_norm": 2.337311267852783, "learning_rate": 0.0001444858591210987, "loss": 4.8049, "step": 595500 }, { "epoch": 3.5910686074697606, "grad_norm": 2.036417245864868, "learning_rate": 0.00014446552396064124, "loss": 4.8589, "step": 595550 }, { "epoch": 3.5913700992510944, "grad_norm": 1.7317183017730713, "learning_rate": 0.00014444518890203788, "loss": 4.7965, "step": 595600 }, { "epoch": 3.5916715910324286, "grad_norm": 3.073516607284546, "learning_rate": 0.00014442485394566297, "loss": 4.8284, "step": 595650 }, { "epoch": 3.5919730828137624, "grad_norm": 2.5638933181762695, "learning_rate": 0.00014440451909189061, "loss": 4.4562, "step": 595700 }, { "epoch": 3.5922745745950966, "grad_norm": 3.0575881004333496, "learning_rate": 0.0001443841843410951, "loss": 4.6671, "step": 595750 }, { "epoch": 3.592576066376431, "grad_norm": 2.6776621341705322, "learning_rate": 0.0001443638496936507, "loss": 4.8847, "step": 595800 }, { "epoch": 3.5928775581577646, "grad_norm": 2.2742815017700195, "learning_rate": 0.00014434351514993154, "loss": 4.6377, "step": 595850 }, { "epoch": 3.5931790499390988, "grad_norm": 2.06807017326355, "learning_rate": 0.00014432318071031192, "loss": 4.8348, "step": 595900 }, { "epoch": 3.5934805417204325, "grad_norm": 1.7132999897003174, "learning_rate": 0.00014430284637516612, "loss": 5.1323, "step": 595950 }, { "epoch": 3.5937820335017667, "grad_norm": 2.18764591217041, "learning_rate": 0.00014428251214486826, "loss": 4.7907, "step": 596000 }, { "epoch": 3.594083525283101, "grad_norm": 1.3622630834579468, "learning_rate": 0.00014426217801979257, "loss": 4.7185, "step": 596050 }, { "epoch": 3.5943850170644347, "grad_norm": 3.943657875061035, "learning_rate": 0.00014424184400031338, "loss": 5.0544, "step": 596100 }, { "epoch": 3.594686508845769, "grad_norm": 1.9562956094741821, "learning_rate": 0.0001442215100868048, "loss": 5.1148, "step": 596150 }, { "epoch": 3.5949880006271027, "grad_norm": 2.6245996952056885, "learning_rate": 0.00014420117627964106, "loss": 5.1382, "step": 596200 }, { "epoch": 3.595289492408437, "grad_norm": 1.7987905740737915, "learning_rate": 0.00014418084257919646, "loss": 4.9374, "step": 596250 }, { "epoch": 3.595590984189771, "grad_norm": 2.1248936653137207, "learning_rate": 0.00014416050898584506, "loss": 4.9634, "step": 596300 }, { "epoch": 3.5958924759711053, "grad_norm": 2.1499197483062744, "learning_rate": 0.0001441401754999612, "loss": 4.7707, "step": 596350 }, { "epoch": 3.596193967752439, "grad_norm": 2.1429648399353027, "learning_rate": 0.00014411984212191908, "loss": 4.4169, "step": 596400 }, { "epoch": 3.5964954595337733, "grad_norm": 2.6389386653900146, "learning_rate": 0.0001440995088520928, "loss": 4.5669, "step": 596450 }, { "epoch": 3.596796951315107, "grad_norm": 3.0157928466796875, "learning_rate": 0.00014407917569085667, "loss": 4.5301, "step": 596500 }, { "epoch": 3.5970984430964412, "grad_norm": 2.330214738845825, "learning_rate": 0.0001440588426385849, "loss": 4.7308, "step": 596550 }, { "epoch": 3.5973999348777754, "grad_norm": 1.967700481414795, "learning_rate": 0.00014403850969565163, "loss": 4.7975, "step": 596600 }, { "epoch": 3.597701426659109, "grad_norm": 2.4337286949157715, "learning_rate": 0.00014401817686243102, "loss": 4.7152, "step": 596650 }, { "epoch": 3.5980029184404434, "grad_norm": 3.0427162647247314, "learning_rate": 0.00014399784413929745, "loss": 4.436, "step": 596700 }, { "epoch": 3.598304410221777, "grad_norm": 3.253941535949707, "learning_rate": 0.00014397751152662493, "loss": 4.8651, "step": 596750 }, { "epoch": 3.5986059020031114, "grad_norm": 1.5827600955963135, "learning_rate": 0.00014395717902478767, "loss": 4.9383, "step": 596800 }, { "epoch": 3.5989073937844456, "grad_norm": 2.287919521331787, "learning_rate": 0.00014393684663416, "loss": 5.2492, "step": 596850 }, { "epoch": 3.5992088855657793, "grad_norm": 2.1524314880371094, "learning_rate": 0.00014391651435511594, "loss": 4.6859, "step": 596900 }, { "epoch": 3.5995103773471135, "grad_norm": 0.38600343465805054, "learning_rate": 0.00014389618218802976, "loss": 5.085, "step": 596950 }, { "epoch": 3.5998118691284473, "grad_norm": 2.0686559677124023, "learning_rate": 0.0001438758501332757, "loss": 4.5438, "step": 597000 }, { "epoch": 3.5998118691284473, "eval_loss": 5.285881042480469, "eval_runtime": 39.1683, "eval_samples_per_second": 13.072, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.417315262907849, "step": 597000 }, { "epoch": 3.6001133609097815, "grad_norm": 3.7460849285125732, "learning_rate": 0.0001438555181912278, "loss": 4.7678, "step": 597050 }, { "epoch": 3.6004148526911157, "grad_norm": 1.4628320932388306, "learning_rate": 0.00014383518636226032, "loss": 4.7459, "step": 597100 }, { "epoch": 3.60071634447245, "grad_norm": 1.6387602090835571, "learning_rate": 0.0001438148546467475, "loss": 4.8896, "step": 597150 }, { "epoch": 3.6010178362537837, "grad_norm": 1.8552134037017822, "learning_rate": 0.00014379452304506344, "loss": 4.9974, "step": 597200 }, { "epoch": 3.601319328035118, "grad_norm": 2.3704288005828857, "learning_rate": 0.00014377419155758227, "loss": 4.7714, "step": 597250 }, { "epoch": 3.6016208198164517, "grad_norm": 2.389198064804077, "learning_rate": 0.0001437538601846783, "loss": 5.2913, "step": 597300 }, { "epoch": 3.601922311597786, "grad_norm": 2.4789741039276123, "learning_rate": 0.00014373352892672555, "loss": 4.9061, "step": 597350 }, { "epoch": 3.60222380337912, "grad_norm": 1.8223397731781006, "learning_rate": 0.00014371319778409827, "loss": 5.0029, "step": 597400 }, { "epoch": 3.602525295160454, "grad_norm": 0.5690242052078247, "learning_rate": 0.00014369286675717063, "loss": 4.8496, "step": 597450 }, { "epoch": 3.602826786941788, "grad_norm": 3.196721315383911, "learning_rate": 0.00014367253584631677, "loss": 4.83, "step": 597500 }, { "epoch": 3.603128278723122, "grad_norm": 2.09965443611145, "learning_rate": 0.0001436522050519109, "loss": 4.8277, "step": 597550 }, { "epoch": 3.603429770504456, "grad_norm": 2.4449777603149414, "learning_rate": 0.00014363187437432706, "loss": 5.3035, "step": 597600 }, { "epoch": 3.60373126228579, "grad_norm": 2.871486186981201, "learning_rate": 0.00014361154381393955, "loss": 4.9008, "step": 597650 }, { "epoch": 3.604032754067124, "grad_norm": 3.138023614883423, "learning_rate": 0.0001435912133711224, "loss": 5.3914, "step": 597700 }, { "epoch": 3.604334245848458, "grad_norm": 2.5347042083740234, "learning_rate": 0.00014357088304624987, "loss": 5.0747, "step": 597750 }, { "epoch": 3.604635737629792, "grad_norm": 1.2383720874786377, "learning_rate": 0.00014355055283969608, "loss": 5.1223, "step": 597800 }, { "epoch": 3.604937229411126, "grad_norm": 1.985948920249939, "learning_rate": 0.0001435302227518351, "loss": 4.9053, "step": 597850 }, { "epoch": 3.6052387211924604, "grad_norm": 1.026909351348877, "learning_rate": 0.00014350989278304118, "loss": 4.4902, "step": 597900 }, { "epoch": 3.6055402129737946, "grad_norm": 2.2593138217926025, "learning_rate": 0.00014348956293368841, "loss": 4.9451, "step": 597950 }, { "epoch": 3.6058417047551283, "grad_norm": 2.0270533561706543, "learning_rate": 0.00014346923320415094, "loss": 5.1715, "step": 598000 }, { "epoch": 3.6061431965364625, "grad_norm": 1.9442027807235718, "learning_rate": 0.0001434489035948029, "loss": 4.7839, "step": 598050 }, { "epoch": 3.6064446883177963, "grad_norm": 1.5831717252731323, "learning_rate": 0.0001434285741060185, "loss": 4.8313, "step": 598100 }, { "epoch": 3.6067461800991305, "grad_norm": 1.5372576713562012, "learning_rate": 0.00014340824473817177, "loss": 4.7782, "step": 598150 }, { "epoch": 3.6070476718804647, "grad_norm": 1.017267107963562, "learning_rate": 0.00014338791549163685, "loss": 4.5761, "step": 598200 }, { "epoch": 3.6073491636617985, "grad_norm": 2.740004777908325, "learning_rate": 0.00014336758636678798, "loss": 5.0856, "step": 598250 }, { "epoch": 3.6076506554431327, "grad_norm": 3.34191632270813, "learning_rate": 0.00014334725736399917, "loss": 5.019, "step": 598300 }, { "epoch": 3.6079521472244664, "grad_norm": 2.8273651599884033, "learning_rate": 0.00014332692848364459, "loss": 4.7722, "step": 598350 }, { "epoch": 3.6082536390058007, "grad_norm": 1.1740739345550537, "learning_rate": 0.00014330659972609845, "loss": 5.1743, "step": 598400 }, { "epoch": 3.608555130787135, "grad_norm": 2.3971850872039795, "learning_rate": 0.00014328627109173468, "loss": 4.6719, "step": 598450 }, { "epoch": 3.6088566225684686, "grad_norm": 1.8908122777938843, "learning_rate": 0.00014326594258092754, "loss": 4.9409, "step": 598500 }, { "epoch": 3.609158114349803, "grad_norm": 2.657447576522827, "learning_rate": 0.00014324561419405116, "loss": 5.203, "step": 598550 }, { "epoch": 3.6094596061311366, "grad_norm": 3.0476157665252686, "learning_rate": 0.0001432252859314796, "loss": 4.8781, "step": 598600 }, { "epoch": 3.609761097912471, "grad_norm": 2.1554455757141113, "learning_rate": 0.00014320495779358695, "loss": 4.7732, "step": 598650 }, { "epoch": 3.610062589693805, "grad_norm": 2.162959098815918, "learning_rate": 0.0001431846297807474, "loss": 4.6803, "step": 598700 }, { "epoch": 3.610364081475139, "grad_norm": 2.394798994064331, "learning_rate": 0.000143164301893335, "loss": 4.8256, "step": 598750 }, { "epoch": 3.610665573256473, "grad_norm": 1.83539617061615, "learning_rate": 0.00014314397413172383, "loss": 5.1398, "step": 598800 }, { "epoch": 3.610967065037807, "grad_norm": 2.810314893722534, "learning_rate": 0.0001431236464962881, "loss": 4.9718, "step": 598850 }, { "epoch": 3.611268556819141, "grad_norm": 2.7943317890167236, "learning_rate": 0.0001431033189874018, "loss": 5.0769, "step": 598900 }, { "epoch": 3.611570048600475, "grad_norm": 2.3212485313415527, "learning_rate": 0.00014308299160543906, "loss": 5.0515, "step": 598950 }, { "epoch": 3.6118715403818094, "grad_norm": 2.3277673721313477, "learning_rate": 0.00014306266435077403, "loss": 5.0601, "step": 599000 }, { "epoch": 3.612173032163143, "grad_norm": 2.323596715927124, "learning_rate": 0.0001430423372237807, "loss": 5.0783, "step": 599050 }, { "epoch": 3.6124745239444773, "grad_norm": 2.175095319747925, "learning_rate": 0.00014302201022483324, "loss": 5.4061, "step": 599100 }, { "epoch": 3.612776015725811, "grad_norm": 1.9961026906967163, "learning_rate": 0.00014300168335430578, "loss": 5.2789, "step": 599150 }, { "epoch": 3.6130775075071453, "grad_norm": 2.878730297088623, "learning_rate": 0.00014298135661257232, "loss": 5.0883, "step": 599200 }, { "epoch": 3.6133789992884795, "grad_norm": 2.1547329425811768, "learning_rate": 0.00014296103000000694, "loss": 4.8441, "step": 599250 }, { "epoch": 3.6136804910698137, "grad_norm": 2.323140859603882, "learning_rate": 0.00014294070351698383, "loss": 4.5324, "step": 599300 }, { "epoch": 3.6139819828511475, "grad_norm": 1.2194111347198486, "learning_rate": 0.00014292037716387695, "loss": 5.0289, "step": 599350 }, { "epoch": 3.6142834746324817, "grad_norm": 2.061596393585205, "learning_rate": 0.0001429000509410604, "loss": 4.8464, "step": 599400 }, { "epoch": 3.6145849664138154, "grad_norm": 2.0722591876983643, "learning_rate": 0.00014287972484890835, "loss": 4.4281, "step": 599450 }, { "epoch": 3.6148864581951496, "grad_norm": 1.2519935369491577, "learning_rate": 0.00014285939888779474, "loss": 4.5081, "step": 599500 }, { "epoch": 3.615187949976484, "grad_norm": 1.731449842453003, "learning_rate": 0.00014283907305809372, "loss": 5.1137, "step": 599550 }, { "epoch": 3.6154894417578176, "grad_norm": 2.3230512142181396, "learning_rate": 0.0001428187473601794, "loss": 4.8359, "step": 599600 }, { "epoch": 3.615790933539152, "grad_norm": 2.5450901985168457, "learning_rate": 0.0001427984217944257, "loss": 4.9201, "step": 599650 }, { "epoch": 3.6160924253204856, "grad_norm": 1.013922929763794, "learning_rate": 0.00014277809636120675, "loss": 4.4693, "step": 599700 }, { "epoch": 3.61639391710182, "grad_norm": 2.4874939918518066, "learning_rate": 0.00014275777106089675, "loss": 4.3454, "step": 599750 }, { "epoch": 3.616695408883154, "grad_norm": 2.540699005126953, "learning_rate": 0.00014273744589386958, "loss": 4.8744, "step": 599800 }, { "epoch": 3.6169969006644878, "grad_norm": 2.1998841762542725, "learning_rate": 0.0001427171208604993, "loss": 5.0162, "step": 599850 }, { "epoch": 3.617298392445822, "grad_norm": 1.62966787815094, "learning_rate": 0.00014269679596116012, "loss": 4.6289, "step": 599900 }, { "epoch": 3.6175998842271557, "grad_norm": 1.8066543340682983, "learning_rate": 0.00014267647119622592, "loss": 4.7989, "step": 599950 }, { "epoch": 3.61790137600849, "grad_norm": 2.0776968002319336, "learning_rate": 0.0001426561465660708, "loss": 4.4707, "step": 600000 }, { "epoch": 3.61790137600849, "eval_loss": 5.281982421875, "eval_runtime": 39.1582, "eval_samples_per_second": 13.075, "eval_steps_per_second": 6.538, "eval_tts_loss": 7.562631792103554, "step": 600000 }, { "epoch": 3.618202867789824, "grad_norm": 1.9488883018493652, "learning_rate": 0.0001426358220710689, "loss": 4.4368, "step": 600050 }, { "epoch": 3.6185043595711583, "grad_norm": 2.1175315380096436, "learning_rate": 0.0001426154977115941, "loss": 4.9382, "step": 600100 }, { "epoch": 3.618805851352492, "grad_norm": 2.138643503189087, "learning_rate": 0.00014259517348802058, "loss": 4.5987, "step": 600150 }, { "epoch": 3.6191073431338263, "grad_norm": 2.7708375453948975, "learning_rate": 0.00014257484940072233, "loss": 4.7437, "step": 600200 }, { "epoch": 3.61940883491516, "grad_norm": 2.694244861602783, "learning_rate": 0.00014255452545007335, "loss": 5.2004, "step": 600250 }, { "epoch": 3.6197103266964943, "grad_norm": 1.9435127973556519, "learning_rate": 0.00014253420163644765, "loss": 4.7381, "step": 600300 }, { "epoch": 3.6200118184778285, "grad_norm": 1.8736357688903809, "learning_rate": 0.00014251387796021943, "loss": 5.0106, "step": 600350 }, { "epoch": 3.6203133102591623, "grad_norm": 1.578931212425232, "learning_rate": 0.00014249355442176256, "loss": 4.6167, "step": 600400 }, { "epoch": 3.6206148020404965, "grad_norm": 2.138183116912842, "learning_rate": 0.00014247323102145105, "loss": 5.0959, "step": 600450 }, { "epoch": 3.6209162938218302, "grad_norm": 1.9812928438186646, "learning_rate": 0.00014245290775965908, "loss": 4.5347, "step": 600500 }, { "epoch": 3.6212177856031644, "grad_norm": 2.6535892486572266, "learning_rate": 0.00014243258463676053, "loss": 4.5878, "step": 600550 }, { "epoch": 3.6215192773844986, "grad_norm": 1.2107806205749512, "learning_rate": 0.0001424122616531294, "loss": 4.6889, "step": 600600 }, { "epoch": 3.6218207691658324, "grad_norm": 1.977651596069336, "learning_rate": 0.00014239193880913987, "loss": 4.6535, "step": 600650 }, { "epoch": 3.6221222609471666, "grad_norm": 2.477976083755493, "learning_rate": 0.00014237161610516577, "loss": 4.5952, "step": 600700 }, { "epoch": 3.6224237527285004, "grad_norm": 1.9173954725265503, "learning_rate": 0.00014235129354158124, "loss": 4.6855, "step": 600750 }, { "epoch": 3.6227252445098346, "grad_norm": 2.022534132003784, "learning_rate": 0.00014233097111876017, "loss": 4.9717, "step": 600800 }, { "epoch": 3.623026736291169, "grad_norm": 2.424381971359253, "learning_rate": 0.00014231064883707672, "loss": 4.9554, "step": 600850 }, { "epoch": 3.623328228072503, "grad_norm": 3.718538522720337, "learning_rate": 0.00014229032669690474, "loss": 4.8299, "step": 600900 }, { "epoch": 3.6236297198538368, "grad_norm": 2.2210402488708496, "learning_rate": 0.0001422700046986183, "loss": 4.8315, "step": 600950 }, { "epoch": 3.623931211635171, "grad_norm": 2.3367717266082764, "learning_rate": 0.00014224968284259144, "loss": 5.0447, "step": 601000 }, { "epoch": 3.6242327034165047, "grad_norm": 1.602226972579956, "learning_rate": 0.00014222936112919806, "loss": 5.2633, "step": 601050 }, { "epoch": 3.624534195197839, "grad_norm": 3.1570351123809814, "learning_rate": 0.00014220903955881222, "loss": 4.9649, "step": 601100 }, { "epoch": 3.624835686979173, "grad_norm": 2.8416080474853516, "learning_rate": 0.00014218871813180791, "loss": 4.8144, "step": 601150 }, { "epoch": 3.625137178760507, "grad_norm": 1.7704423666000366, "learning_rate": 0.00014216839684855906, "loss": 4.8455, "step": 601200 }, { "epoch": 3.625438670541841, "grad_norm": 2.17303204536438, "learning_rate": 0.00014214807570943965, "loss": 4.4285, "step": 601250 }, { "epoch": 3.625740162323175, "grad_norm": 2.1486589908599854, "learning_rate": 0.0001421277547148238, "loss": 4.7744, "step": 601300 }, { "epoch": 3.626041654104509, "grad_norm": 1.7095388174057007, "learning_rate": 0.00014210743386508534, "loss": 4.5486, "step": 601350 }, { "epoch": 3.6263431458858433, "grad_norm": 2.619718074798584, "learning_rate": 0.00014208711316059827, "loss": 4.922, "step": 601400 }, { "epoch": 3.626644637667177, "grad_norm": 1.0690195560455322, "learning_rate": 0.00014206679260173665, "loss": 4.6499, "step": 601450 }, { "epoch": 3.6269461294485112, "grad_norm": 1.9556177854537964, "learning_rate": 0.00014204647218887432, "loss": 4.8747, "step": 601500 }, { "epoch": 3.627247621229845, "grad_norm": 2.1743104457855225, "learning_rate": 0.00014202615192238536, "loss": 4.7337, "step": 601550 }, { "epoch": 3.627549113011179, "grad_norm": 2.1555516719818115, "learning_rate": 0.00014200583180264372, "loss": 4.7878, "step": 601600 }, { "epoch": 3.6278506047925134, "grad_norm": 1.9062387943267822, "learning_rate": 0.0001419855118300233, "loss": 4.4294, "step": 601650 }, { "epoch": 3.6281520965738476, "grad_norm": 2.0703518390655518, "learning_rate": 0.0001419651920048981, "loss": 4.8812, "step": 601700 }, { "epoch": 3.6284535883551814, "grad_norm": 1.8122764825820923, "learning_rate": 0.00014194487232764214, "loss": 4.9641, "step": 601750 }, { "epoch": 3.6287550801365156, "grad_norm": 2.086225748062134, "learning_rate": 0.0001419245527986292, "loss": 4.9705, "step": 601800 }, { "epoch": 3.6290565719178494, "grad_norm": 1.6558480262756348, "learning_rate": 0.00014190423341823335, "loss": 5.0473, "step": 601850 }, { "epoch": 3.6293580636991836, "grad_norm": 1.4146316051483154, "learning_rate": 0.00014188391418682862, "loss": 4.4559, "step": 601900 }, { "epoch": 3.6296595554805178, "grad_norm": 2.252791404724121, "learning_rate": 0.00014186359510478882, "loss": 4.7691, "step": 601950 }, { "epoch": 3.6299610472618515, "grad_norm": 1.635374665260315, "learning_rate": 0.00014184327617248793, "loss": 5.0978, "step": 602000 }, { "epoch": 3.6302625390431857, "grad_norm": 2.950056791305542, "learning_rate": 0.00014182295739029993, "loss": 4.9793, "step": 602050 }, { "epoch": 3.6305640308245195, "grad_norm": 2.8927693367004395, "learning_rate": 0.00014180263875859868, "loss": 5.0424, "step": 602100 }, { "epoch": 3.6308655226058537, "grad_norm": 2.0306313037872314, "learning_rate": 0.0001417823202777582, "loss": 5.1282, "step": 602150 }, { "epoch": 3.631167014387188, "grad_norm": 1.4861586093902588, "learning_rate": 0.00014176200194815244, "loss": 4.6921, "step": 602200 }, { "epoch": 3.6314685061685217, "grad_norm": 1.4697908163070679, "learning_rate": 0.00014174168377015519, "loss": 4.7765, "step": 602250 }, { "epoch": 3.631769997949856, "grad_norm": 2.491675615310669, "learning_rate": 0.0001417213657441405, "loss": 4.906, "step": 602300 }, { "epoch": 3.6320714897311897, "grad_norm": 0.40746814012527466, "learning_rate": 0.00014170104787048228, "loss": 5.1733, "step": 602350 }, { "epoch": 3.632372981512524, "grad_norm": 1.9916186332702637, "learning_rate": 0.00014168073014955438, "loss": 4.8386, "step": 602400 }, { "epoch": 3.632674473293858, "grad_norm": 2.0230817794799805, "learning_rate": 0.00014166041258173077, "loss": 4.7278, "step": 602450 }, { "epoch": 3.6329759650751923, "grad_norm": 1.9494119882583618, "learning_rate": 0.00014164009516738543, "loss": 4.5913, "step": 602500 }, { "epoch": 3.633277456856526, "grad_norm": 2.895730972290039, "learning_rate": 0.00014161977790689217, "loss": 4.594, "step": 602550 }, { "epoch": 3.6335789486378602, "grad_norm": 2.0814785957336426, "learning_rate": 0.0001415994608006249, "loss": 4.5812, "step": 602600 }, { "epoch": 3.633880440419194, "grad_norm": 4.787227630615234, "learning_rate": 0.00014157914384895764, "loss": 4.6193, "step": 602650 }, { "epoch": 3.634181932200528, "grad_norm": 2.4920408725738525, "learning_rate": 0.00014155882705226417, "loss": 4.8512, "step": 602700 }, { "epoch": 3.6344834239818624, "grad_norm": 3.5469930171966553, "learning_rate": 0.00014153851041091847, "loss": 4.7582, "step": 602750 }, { "epoch": 3.634784915763196, "grad_norm": 1.7898192405700684, "learning_rate": 0.00014151819392529442, "loss": 5.2074, "step": 602800 }, { "epoch": 3.6350864075445304, "grad_norm": 2.9117071628570557, "learning_rate": 0.00014149787759576586, "loss": 4.6465, "step": 602850 }, { "epoch": 3.635387899325864, "grad_norm": 1.867037296295166, "learning_rate": 0.00014147756142270676, "loss": 4.9813, "step": 602900 }, { "epoch": 3.6356893911071984, "grad_norm": 3.5052950382232666, "learning_rate": 0.000141457245406491, "loss": 4.9697, "step": 602950 }, { "epoch": 3.6359908828885326, "grad_norm": 1.753382682800293, "learning_rate": 0.0001414369295474924, "loss": 4.8285, "step": 603000 }, { "epoch": 3.6359908828885326, "eval_loss": 5.280771732330322, "eval_runtime": 39.2044, "eval_samples_per_second": 13.06, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.53185281044594, "step": 603000 }, { "epoch": 3.6362923746698668, "grad_norm": 1.0043156147003174, "learning_rate": 0.00014141661384608489, "loss": 4.8849, "step": 603050 }, { "epoch": 3.6365938664512005, "grad_norm": 1.784505844116211, "learning_rate": 0.00014139629830264242, "loss": 4.5798, "step": 603100 }, { "epoch": 3.6368953582325347, "grad_norm": 2.2679169178009033, "learning_rate": 0.00014137598291753874, "loss": 4.9286, "step": 603150 }, { "epoch": 3.6371968500138685, "grad_norm": 2.681427240371704, "learning_rate": 0.0001413556676911478, "loss": 5.0552, "step": 603200 }, { "epoch": 3.6374983417952027, "grad_norm": 2.003040313720703, "learning_rate": 0.00014133535262384348, "loss": 4.6627, "step": 603250 }, { "epoch": 3.637799833576537, "grad_norm": 2.0931241512298584, "learning_rate": 0.0001413150377159996, "loss": 4.6798, "step": 603300 }, { "epoch": 3.6381013253578707, "grad_norm": 1.8549174070358276, "learning_rate": 0.00014129472296799006, "loss": 4.7684, "step": 603350 }, { "epoch": 3.638402817139205, "grad_norm": 2.475480318069458, "learning_rate": 0.00014127440838018879, "loss": 4.8033, "step": 603400 }, { "epoch": 3.6387043089205386, "grad_norm": 2.1931331157684326, "learning_rate": 0.00014125409395296948, "loss": 4.7732, "step": 603450 }, { "epoch": 3.639005800701873, "grad_norm": 1.9589570760726929, "learning_rate": 0.00014123377968670614, "loss": 5.0637, "step": 603500 }, { "epoch": 3.639307292483207, "grad_norm": 2.3099966049194336, "learning_rate": 0.00014121346558177258, "loss": 5.1351, "step": 603550 }, { "epoch": 3.639608784264541, "grad_norm": 1.2501707077026367, "learning_rate": 0.00014119315163854264, "loss": 4.6331, "step": 603600 }, { "epoch": 3.639910276045875, "grad_norm": 2.297065258026123, "learning_rate": 0.0001411728378573901, "loss": 4.7578, "step": 603650 }, { "epoch": 3.640211767827209, "grad_norm": 2.0342020988464355, "learning_rate": 0.00014115252423868902, "loss": 5.0636, "step": 603700 }, { "epoch": 3.640513259608543, "grad_norm": 2.119058609008789, "learning_rate": 0.00014113221078281305, "loss": 5.0596, "step": 603750 }, { "epoch": 3.640814751389877, "grad_norm": 2.602567672729492, "learning_rate": 0.00014111189749013603, "loss": 4.7715, "step": 603800 }, { "epoch": 3.6411162431712114, "grad_norm": 2.002916097640991, "learning_rate": 0.00014109158436103193, "loss": 4.8626, "step": 603850 }, { "epoch": 3.641417734952545, "grad_norm": 2.4993295669555664, "learning_rate": 0.00014107127139587446, "loss": 5.1109, "step": 603900 }, { "epoch": 3.6417192267338794, "grad_norm": 1.766196370124817, "learning_rate": 0.0001410509585950375, "loss": 5.1264, "step": 603950 }, { "epoch": 3.642020718515213, "grad_norm": 2.004511594772339, "learning_rate": 0.00014103064595889493, "loss": 4.4279, "step": 604000 }, { "epoch": 3.6423222102965473, "grad_norm": 0.8775911331176758, "learning_rate": 0.00014101033348782045, "loss": 4.5624, "step": 604050 }, { "epoch": 3.6426237020778816, "grad_norm": 2.585066318511963, "learning_rate": 0.00014099002118218794, "loss": 4.8787, "step": 604100 }, { "epoch": 3.6429251938592153, "grad_norm": 2.046640396118164, "learning_rate": 0.00014096970904237132, "loss": 4.9053, "step": 604150 }, { "epoch": 3.6432266856405495, "grad_norm": 1.4869718551635742, "learning_rate": 0.00014094939706874428, "loss": 4.7498, "step": 604200 }, { "epoch": 3.6435281774218833, "grad_norm": 1.9693398475646973, "learning_rate": 0.00014092908526168064, "loss": 4.843, "step": 604250 }, { "epoch": 3.6438296692032175, "grad_norm": 3.3616943359375, "learning_rate": 0.00014090877362155427, "loss": 4.9331, "step": 604300 }, { "epoch": 3.6441311609845517, "grad_norm": 2.4337098598480225, "learning_rate": 0.00014088846214873897, "loss": 4.8276, "step": 604350 }, { "epoch": 3.6444326527658855, "grad_norm": 1.9255229234695435, "learning_rate": 0.00014086815084360846, "loss": 4.9823, "step": 604400 }, { "epoch": 3.6447341445472197, "grad_norm": 3.0084896087646484, "learning_rate": 0.0001408478397065366, "loss": 4.9491, "step": 604450 }, { "epoch": 3.6450356363285534, "grad_norm": 3.2795021533966064, "learning_rate": 0.00014082752873789727, "loss": 4.569, "step": 604500 }, { "epoch": 3.6453371281098876, "grad_norm": 2.323791027069092, "learning_rate": 0.00014080721793806414, "loss": 5.2828, "step": 604550 }, { "epoch": 3.645638619891222, "grad_norm": 1.896728515625, "learning_rate": 0.00014078690730741104, "loss": 5.0054, "step": 604600 }, { "epoch": 3.645940111672556, "grad_norm": 1.8774877786636353, "learning_rate": 0.00014076659684631178, "loss": 4.9935, "step": 604650 }, { "epoch": 3.64624160345389, "grad_norm": 2.6795289516448975, "learning_rate": 0.0001407462865551401, "loss": 4.8123, "step": 604700 }, { "epoch": 3.646543095235224, "grad_norm": 2.680063009262085, "learning_rate": 0.0001407259764342698, "loss": 4.7064, "step": 604750 }, { "epoch": 3.646844587016558, "grad_norm": 1.3878018856048584, "learning_rate": 0.00014070566648407475, "loss": 5.0668, "step": 604800 }, { "epoch": 3.647146078797892, "grad_norm": 1.7356212139129639, "learning_rate": 0.00014068535670492855, "loss": 4.6093, "step": 604850 }, { "epoch": 3.647447570579226, "grad_norm": 2.3560736179351807, "learning_rate": 0.00014066504709720512, "loss": 5.0784, "step": 604900 }, { "epoch": 3.64774906236056, "grad_norm": 2.0444703102111816, "learning_rate": 0.00014064473766127822, "loss": 5.0758, "step": 604950 }, { "epoch": 3.648050554141894, "grad_norm": 2.087239980697632, "learning_rate": 0.00014062442839752148, "loss": 4.9595, "step": 605000 }, { "epoch": 3.648352045923228, "grad_norm": 2.4124295711517334, "learning_rate": 0.00014060411930630874, "loss": 4.9154, "step": 605050 }, { "epoch": 3.648653537704562, "grad_norm": 2.2441279888153076, "learning_rate": 0.00014058381038801386, "loss": 4.954, "step": 605100 }, { "epoch": 3.6489550294858963, "grad_norm": 1.2550854682922363, "learning_rate": 0.00014056350164301049, "loss": 4.4803, "step": 605150 }, { "epoch": 3.64925652126723, "grad_norm": 2.6659679412841797, "learning_rate": 0.00014054319307167236, "loss": 4.9401, "step": 605200 }, { "epoch": 3.6495580130485643, "grad_norm": 1.5388373136520386, "learning_rate": 0.00014052288467437333, "loss": 5.1836, "step": 605250 }, { "epoch": 3.649859504829898, "grad_norm": 2.6068005561828613, "learning_rate": 0.00014050257645148703, "loss": 5.1267, "step": 605300 }, { "epoch": 3.6501609966112323, "grad_norm": 2.0362844467163086, "learning_rate": 0.00014048226840338729, "loss": 4.8303, "step": 605350 }, { "epoch": 3.6504624883925665, "grad_norm": 2.060683250427246, "learning_rate": 0.00014046196053044783, "loss": 4.8869, "step": 605400 }, { "epoch": 3.6507639801739007, "grad_norm": 2.706183433532715, "learning_rate": 0.00014044165283304234, "loss": 4.686, "step": 605450 }, { "epoch": 3.6510654719552345, "grad_norm": 1.086942195892334, "learning_rate": 0.0001404213453115446, "loss": 4.6012, "step": 605500 }, { "epoch": 3.6513669637365687, "grad_norm": 1.8585002422332764, "learning_rate": 0.00014040103796632836, "loss": 4.5947, "step": 605550 }, { "epoch": 3.6516684555179024, "grad_norm": 2.4666483402252197, "learning_rate": 0.00014038073079776725, "loss": 4.9915, "step": 605600 }, { "epoch": 3.6519699472992366, "grad_norm": 1.6431196928024292, "learning_rate": 0.00014036042380623508, "loss": 4.622, "step": 605650 }, { "epoch": 3.652271439080571, "grad_norm": 1.8038288354873657, "learning_rate": 0.00014034011699210562, "loss": 4.6434, "step": 605700 }, { "epoch": 3.6525729308619046, "grad_norm": 1.8589882850646973, "learning_rate": 0.0001403198103557525, "loss": 5.0204, "step": 605750 }, { "epoch": 3.652874422643239, "grad_norm": 2.774874687194824, "learning_rate": 0.0001402995038975494, "loss": 4.7502, "step": 605800 }, { "epoch": 3.6531759144245726, "grad_norm": 2.532569169998169, "learning_rate": 0.00014027919761787014, "loss": 4.4371, "step": 605850 }, { "epoch": 3.6534774062059068, "grad_norm": 2.6130166053771973, "learning_rate": 0.00014025889151708833, "loss": 4.4223, "step": 605900 }, { "epoch": 3.653778897987241, "grad_norm": 2.124739646911621, "learning_rate": 0.00014023858559557775, "loss": 4.6048, "step": 605950 }, { "epoch": 3.6540803897685747, "grad_norm": 2.3057210445404053, "learning_rate": 0.0001402182798537121, "loss": 4.7304, "step": 606000 }, { "epoch": 3.6540803897685747, "eval_loss": 5.29068660736084, "eval_runtime": 39.1373, "eval_samples_per_second": 13.082, "eval_steps_per_second": 6.541, "eval_tts_loss": 7.555292674833535, "step": 606000 }, { "epoch": 3.654381881549909, "grad_norm": 1.9559332132339478, "learning_rate": 0.00014019797429186499, "loss": 5.1961, "step": 606050 }, { "epoch": 3.6546833733312427, "grad_norm": 2.537101984024048, "learning_rate": 0.0001401776689104102, "loss": 5.0496, "step": 606100 }, { "epoch": 3.654984865112577, "grad_norm": 1.5447750091552734, "learning_rate": 0.00014015736370972143, "loss": 4.9919, "step": 606150 }, { "epoch": 3.655286356893911, "grad_norm": 2.147279739379883, "learning_rate": 0.00014013705869017227, "loss": 4.9303, "step": 606200 }, { "epoch": 3.6555878486752453, "grad_norm": 2.596881628036499, "learning_rate": 0.00014011675385213647, "loss": 5.0836, "step": 606250 }, { "epoch": 3.655889340456579, "grad_norm": 3.8554463386535645, "learning_rate": 0.0001400964491959878, "loss": 4.6407, "step": 606300 }, { "epoch": 3.6561908322379133, "grad_norm": 1.236157774925232, "learning_rate": 0.0001400761447220998, "loss": 4.722, "step": 606350 }, { "epoch": 3.656492324019247, "grad_norm": 3.5483455657958984, "learning_rate": 0.00014005584043084616, "loss": 4.8009, "step": 606400 }, { "epoch": 3.6567938158005813, "grad_norm": 2.948859453201294, "learning_rate": 0.00014003553632260064, "loss": 5.1245, "step": 606450 }, { "epoch": 3.6570953075819155, "grad_norm": 2.0427114963531494, "learning_rate": 0.0001400152323977368, "loss": 4.8315, "step": 606500 }, { "epoch": 3.6573967993632492, "grad_norm": 2.9186694622039795, "learning_rate": 0.0001399949286566284, "loss": 4.9265, "step": 606550 }, { "epoch": 3.6576982911445834, "grad_norm": 3.0799787044525146, "learning_rate": 0.00013997462509964908, "loss": 4.5658, "step": 606600 }, { "epoch": 3.657999782925917, "grad_norm": 2.13952374458313, "learning_rate": 0.0001399543217271724, "loss": 4.8077, "step": 606650 }, { "epoch": 3.6583012747072514, "grad_norm": 2.247169017791748, "learning_rate": 0.00013993401853957215, "loss": 4.6994, "step": 606700 }, { "epoch": 3.6586027664885856, "grad_norm": 2.0537939071655273, "learning_rate": 0.00013991371553722195, "loss": 4.6279, "step": 606750 }, { "epoch": 3.65890425826992, "grad_norm": 2.1433603763580322, "learning_rate": 0.00013989341272049535, "loss": 4.5189, "step": 606800 }, { "epoch": 3.6592057500512536, "grad_norm": 0.8473440408706665, "learning_rate": 0.00013987311008976607, "loss": 4.6816, "step": 606850 }, { "epoch": 3.659507241832588, "grad_norm": 1.9406826496124268, "learning_rate": 0.00013985280764540782, "loss": 4.5868, "step": 606900 }, { "epoch": 3.6598087336139216, "grad_norm": 2.5540151596069336, "learning_rate": 0.0001398325053877941, "loss": 4.655, "step": 606950 }, { "epoch": 3.6601102253952558, "grad_norm": 3.1172029972076416, "learning_rate": 0.00013981220331729863, "loss": 5.1364, "step": 607000 }, { "epoch": 3.66041171717659, "grad_norm": 3.9183974266052246, "learning_rate": 0.00013979190143429505, "loss": 5.2849, "step": 607050 }, { "epoch": 3.6607132089579237, "grad_norm": 2.571378231048584, "learning_rate": 0.0001397715997391569, "loss": 4.353, "step": 607100 }, { "epoch": 3.661014700739258, "grad_norm": 2.4280524253845215, "learning_rate": 0.00013975129823225787, "loss": 5.0929, "step": 607150 }, { "epoch": 3.6613161925205917, "grad_norm": 2.7316951751708984, "learning_rate": 0.00013973099691397164, "loss": 4.6493, "step": 607200 }, { "epoch": 3.661617684301926, "grad_norm": 0.9975784420967102, "learning_rate": 0.00013971069578467166, "loss": 4.99, "step": 607250 }, { "epoch": 3.66191917608326, "grad_norm": 1.9896399974822998, "learning_rate": 0.00013969039484473168, "loss": 4.8811, "step": 607300 }, { "epoch": 3.662220667864594, "grad_norm": 0.7077226042747498, "learning_rate": 0.00013967009409452533, "loss": 4.5867, "step": 607350 }, { "epoch": 3.662522159645928, "grad_norm": 2.45564603805542, "learning_rate": 0.00013964979353442607, "loss": 4.9409, "step": 607400 }, { "epoch": 3.662823651427262, "grad_norm": 3.069898843765259, "learning_rate": 0.00013962949316480758, "loss": 4.8637, "step": 607450 }, { "epoch": 3.663125143208596, "grad_norm": 2.7274351119995117, "learning_rate": 0.00013960919298604353, "loss": 5.0887, "step": 607500 }, { "epoch": 3.6634266349899303, "grad_norm": 1.041351079940796, "learning_rate": 0.00013958889299850745, "loss": 4.5409, "step": 607550 }, { "epoch": 3.6637281267712645, "grad_norm": 2.0025312900543213, "learning_rate": 0.00013956859320257288, "loss": 4.8585, "step": 607600 }, { "epoch": 3.6640296185525982, "grad_norm": 2.1617543697357178, "learning_rate": 0.0001395482935986135, "loss": 5.1817, "step": 607650 }, { "epoch": 3.6643311103339324, "grad_norm": 1.1119879484176636, "learning_rate": 0.00013952799418700287, "loss": 4.6821, "step": 607700 }, { "epoch": 3.664632602115266, "grad_norm": 2.2885754108428955, "learning_rate": 0.00013950769496811455, "loss": 4.7834, "step": 607750 }, { "epoch": 3.6649340938966004, "grad_norm": 2.0017735958099365, "learning_rate": 0.00013948739594232213, "loss": 4.9594, "step": 607800 }, { "epoch": 3.6652355856779346, "grad_norm": 3.5547635555267334, "learning_rate": 0.00013946709710999924, "loss": 4.7436, "step": 607850 }, { "epoch": 3.6655370774592684, "grad_norm": 2.9183859825134277, "learning_rate": 0.00013944679847151934, "loss": 5.1071, "step": 607900 }, { "epoch": 3.6658385692406026, "grad_norm": 1.004065752029419, "learning_rate": 0.00013942650002725604, "loss": 4.6549, "step": 607950 }, { "epoch": 3.6661400610219363, "grad_norm": 1.5183265209197998, "learning_rate": 0.00013940620177758296, "loss": 4.6148, "step": 608000 }, { "epoch": 3.6664415528032706, "grad_norm": 2.4698445796966553, "learning_rate": 0.0001393859037228736, "loss": 4.8079, "step": 608050 }, { "epoch": 3.6667430445846048, "grad_norm": 2.237917184829712, "learning_rate": 0.00013936560586350153, "loss": 4.9378, "step": 608100 }, { "epoch": 3.6670445363659385, "grad_norm": 1.7851853370666504, "learning_rate": 0.00013934530819984035, "loss": 4.6451, "step": 608150 }, { "epoch": 3.6673460281472727, "grad_norm": 0.5026610493659973, "learning_rate": 0.0001393250107322635, "loss": 4.5358, "step": 608200 }, { "epoch": 3.6676475199286065, "grad_norm": 4.150226593017578, "learning_rate": 0.00013930471346114458, "loss": 5.0218, "step": 608250 }, { "epoch": 3.6679490117099407, "grad_norm": 1.4216737747192383, "learning_rate": 0.00013928441638685725, "loss": 4.8419, "step": 608300 }, { "epoch": 3.668250503491275, "grad_norm": 0.8809646368026733, "learning_rate": 0.00013926411950977488, "loss": 4.8331, "step": 608350 }, { "epoch": 3.668551995272609, "grad_norm": 2.521407127380371, "learning_rate": 0.00013924382283027104, "loss": 4.7634, "step": 608400 }, { "epoch": 3.668853487053943, "grad_norm": 2.0282461643218994, "learning_rate": 0.00013922352634871936, "loss": 4.6808, "step": 608450 }, { "epoch": 3.669154978835277, "grad_norm": 2.380004405975342, "learning_rate": 0.00013920323006549328, "loss": 4.3581, "step": 608500 }, { "epoch": 3.669456470616611, "grad_norm": 3.369656801223755, "learning_rate": 0.00013918293398096628, "loss": 4.7415, "step": 608550 }, { "epoch": 3.669757962397945, "grad_norm": 2.232241153717041, "learning_rate": 0.00013916263809551205, "loss": 4.8969, "step": 608600 }, { "epoch": 3.6700594541792793, "grad_norm": 1.3178317546844482, "learning_rate": 0.0001391423424095039, "loss": 4.8363, "step": 608650 }, { "epoch": 3.670360945960613, "grad_norm": 2.3620195388793945, "learning_rate": 0.00013912204692331548, "loss": 5.1415, "step": 608700 }, { "epoch": 3.6706624377419472, "grad_norm": 1.7405588626861572, "learning_rate": 0.00013910175163732032, "loss": 4.1583, "step": 608750 }, { "epoch": 3.670963929523281, "grad_norm": 3.460843324661255, "learning_rate": 0.00013908145655189177, "loss": 4.9132, "step": 608800 }, { "epoch": 3.671265421304615, "grad_norm": 1.6592755317687988, "learning_rate": 0.00013906116166740345, "loss": 4.9402, "step": 608850 }, { "epoch": 3.6715669130859494, "grad_norm": 2.0624351501464844, "learning_rate": 0.0001390408669842289, "loss": 5.139, "step": 608900 }, { "epoch": 3.671868404867283, "grad_norm": 1.65823495388031, "learning_rate": 0.00013902057250274154, "loss": 4.6291, "step": 608950 }, { "epoch": 3.6721698966486174, "grad_norm": 2.2876832485198975, "learning_rate": 0.0001390002782233148, "loss": 5.158, "step": 609000 }, { "epoch": 3.6721698966486174, "eval_loss": 5.279341697692871, "eval_runtime": 39.0394, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.508117918741824, "step": 609000 }, { "epoch": 3.672471388429951, "grad_norm": 2.6307883262634277, "learning_rate": 0.00013897998414632234, "loss": 5.0548, "step": 609050 }, { "epoch": 3.6727728802112853, "grad_norm": 1.684024691581726, "learning_rate": 0.00013895969027213746, "loss": 4.7444, "step": 609100 }, { "epoch": 3.6730743719926195, "grad_norm": 2.586719036102295, "learning_rate": 0.00013893939660113378, "loss": 4.9581, "step": 609150 }, { "epoch": 3.6733758637739538, "grad_norm": 1.6179438829421997, "learning_rate": 0.00013891910313368474, "loss": 4.4496, "step": 609200 }, { "epoch": 3.6736773555552875, "grad_norm": 1.6792337894439697, "learning_rate": 0.00013889880987016373, "loss": 5.0695, "step": 609250 }, { "epoch": 3.6739788473366217, "grad_norm": 2.4939043521881104, "learning_rate": 0.0001388785168109443, "loss": 5.0027, "step": 609300 }, { "epoch": 3.6742803391179555, "grad_norm": 1.6376841068267822, "learning_rate": 0.00013885822395639995, "loss": 4.8172, "step": 609350 }, { "epoch": 3.6745818308992897, "grad_norm": 0.5878002047538757, "learning_rate": 0.00013883793130690402, "loss": 4.5266, "step": 609400 }, { "epoch": 3.674883322680624, "grad_norm": 2.5077061653137207, "learning_rate": 0.00013881763886283005, "loss": 4.665, "step": 609450 }, { "epoch": 3.6751848144619577, "grad_norm": 0.9632259607315063, "learning_rate": 0.0001387973466245515, "loss": 4.6846, "step": 609500 }, { "epoch": 3.675486306243292, "grad_norm": 3.3365087509155273, "learning_rate": 0.00013877705459244182, "loss": 4.9105, "step": 609550 }, { "epoch": 3.6757877980246256, "grad_norm": 1.7183923721313477, "learning_rate": 0.00013875676276687437, "loss": 5.3382, "step": 609600 }, { "epoch": 3.67608928980596, "grad_norm": 2.0819156169891357, "learning_rate": 0.0001387364711482227, "loss": 4.3464, "step": 609650 }, { "epoch": 3.676390781587294, "grad_norm": 2.8726601600646973, "learning_rate": 0.00013871617973686018, "loss": 5.2258, "step": 609700 }, { "epoch": 3.676692273368628, "grad_norm": 2.0269057750701904, "learning_rate": 0.00013869588853316028, "loss": 4.7246, "step": 609750 }, { "epoch": 3.676993765149962, "grad_norm": 3.8701937198638916, "learning_rate": 0.00013867559753749648, "loss": 4.6637, "step": 609800 }, { "epoch": 3.677295256931296, "grad_norm": 2.8051135540008545, "learning_rate": 0.00013865530675024205, "loss": 5.1196, "step": 609850 }, { "epoch": 3.67759674871263, "grad_norm": 2.9080002307891846, "learning_rate": 0.00013863501617177054, "loss": 4.4469, "step": 609900 }, { "epoch": 3.677898240493964, "grad_norm": 1.9943889379501343, "learning_rate": 0.0001386147258024554, "loss": 4.6934, "step": 609950 }, { "epoch": 3.6781997322752984, "grad_norm": 1.1511245965957642, "learning_rate": 0.00013859443564266988, "loss": 4.6076, "step": 610000 }, { "epoch": 3.678501224056632, "grad_norm": 1.6328256130218506, "learning_rate": 0.00013857414569278754, "loss": 4.7941, "step": 610050 }, { "epoch": 3.6788027158379664, "grad_norm": 1.3499915599822998, "learning_rate": 0.00013855385595318176, "loss": 4.8822, "step": 610100 }, { "epoch": 3.6791042076193, "grad_norm": 1.2418653964996338, "learning_rate": 0.00013853356642422592, "loss": 4.1839, "step": 610150 }, { "epoch": 3.6794056994006343, "grad_norm": 1.8918348550796509, "learning_rate": 0.0001385132771062934, "loss": 4.6985, "step": 610200 }, { "epoch": 3.6797071911819685, "grad_norm": 2.801795721054077, "learning_rate": 0.0001384929879997577, "loss": 5.0172, "step": 610250 }, { "epoch": 3.6800086829633023, "grad_norm": 3.732018232345581, "learning_rate": 0.00013847269910499205, "loss": 5.0073, "step": 610300 }, { "epoch": 3.6803101747446365, "grad_norm": 2.8140294551849365, "learning_rate": 0.00013845241042236995, "loss": 4.5801, "step": 610350 }, { "epoch": 3.6806116665259703, "grad_norm": 4.180272102355957, "learning_rate": 0.0001384321219522648, "loss": 5.2083, "step": 610400 }, { "epoch": 3.6809131583073045, "grad_norm": 1.6751525402069092, "learning_rate": 0.00013841183369504986, "loss": 4.9255, "step": 610450 }, { "epoch": 3.6812146500886387, "grad_norm": 3.549326181411743, "learning_rate": 0.0001383915456510986, "loss": 5.0604, "step": 610500 }, { "epoch": 3.681516141869973, "grad_norm": 2.053597927093506, "learning_rate": 0.00013837125782078445, "loss": 5.0562, "step": 610550 }, { "epoch": 3.6818176336513067, "grad_norm": 1.902752161026001, "learning_rate": 0.00013835097020448064, "loss": 4.7302, "step": 610600 }, { "epoch": 3.682119125432641, "grad_norm": 1.7586315870285034, "learning_rate": 0.00013833068280256057, "loss": 4.789, "step": 610650 }, { "epoch": 3.6824206172139746, "grad_norm": 2.225637197494507, "learning_rate": 0.0001383103956153977, "loss": 4.8863, "step": 610700 }, { "epoch": 3.682722108995309, "grad_norm": 3.0876059532165527, "learning_rate": 0.00013829010864336528, "loss": 4.9038, "step": 610750 }, { "epoch": 3.683023600776643, "grad_norm": 3.5369303226470947, "learning_rate": 0.0001382698218868367, "loss": 5.1685, "step": 610800 }, { "epoch": 3.683325092557977, "grad_norm": 2.4578919410705566, "learning_rate": 0.00013824953534618533, "loss": 5.1229, "step": 610850 }, { "epoch": 3.683626584339311, "grad_norm": 1.9168041944503784, "learning_rate": 0.00013822924902178444, "loss": 4.5845, "step": 610900 }, { "epoch": 3.6839280761206448, "grad_norm": 2.4632644653320312, "learning_rate": 0.00013820896291400747, "loss": 4.5463, "step": 610950 }, { "epoch": 3.684229567901979, "grad_norm": 1.392192006111145, "learning_rate": 0.00013818867702322768, "loss": 4.8783, "step": 611000 }, { "epoch": 3.684531059683313, "grad_norm": 2.3147785663604736, "learning_rate": 0.0001381683913498185, "loss": 4.927, "step": 611050 }, { "epoch": 3.684832551464647, "grad_norm": 2.1598634719848633, "learning_rate": 0.00013814810589415314, "loss": 4.9048, "step": 611100 }, { "epoch": 3.685134043245981, "grad_norm": 3.558774471282959, "learning_rate": 0.00013812782065660497, "loss": 5.0928, "step": 611150 }, { "epoch": 3.685435535027315, "grad_norm": 2.0684194564819336, "learning_rate": 0.00013810753563754736, "loss": 4.7474, "step": 611200 }, { "epoch": 3.685737026808649, "grad_norm": 3.020352840423584, "learning_rate": 0.00013808725083735353, "loss": 5.1833, "step": 611250 }, { "epoch": 3.6860385185899833, "grad_norm": 3.022963762283325, "learning_rate": 0.0001380669662563969, "loss": 5.378, "step": 611300 }, { "epoch": 3.6863400103713175, "grad_norm": 0.9192133545875549, "learning_rate": 0.00013804668189505075, "loss": 4.9832, "step": 611350 }, { "epoch": 3.6866415021526513, "grad_norm": 2.350778341293335, "learning_rate": 0.00013802639775368827, "loss": 4.8024, "step": 611400 }, { "epoch": 3.6869429939339855, "grad_norm": 1.033347725868225, "learning_rate": 0.00013800611383268287, "loss": 4.7111, "step": 611450 }, { "epoch": 3.6872444857153193, "grad_norm": 2.557236909866333, "learning_rate": 0.00013798583013240788, "loss": 4.7294, "step": 611500 }, { "epoch": 3.6875459774966535, "grad_norm": 3.181957960128784, "learning_rate": 0.00013796554665323652, "loss": 5.3476, "step": 611550 }, { "epoch": 3.6878474692779877, "grad_norm": 2.259755849838257, "learning_rate": 0.00013794526339554208, "loss": 4.7666, "step": 611600 }, { "epoch": 3.6881489610593214, "grad_norm": 2.0017480850219727, "learning_rate": 0.00013792498035969793, "loss": 4.7874, "step": 611650 }, { "epoch": 3.6884504528406556, "grad_norm": 1.9504013061523438, "learning_rate": 0.00013790469754607721, "loss": 5.1677, "step": 611700 }, { "epoch": 3.6887519446219894, "grad_norm": 5.153165340423584, "learning_rate": 0.0001378844149550533, "loss": 4.8103, "step": 611750 }, { "epoch": 3.6890534364033236, "grad_norm": 2.659633159637451, "learning_rate": 0.00013786413258699948, "loss": 5.0299, "step": 611800 }, { "epoch": 3.689354928184658, "grad_norm": 0.5185667872428894, "learning_rate": 0.0001378438504422889, "loss": 4.6411, "step": 611850 }, { "epoch": 3.6896564199659916, "grad_norm": 2.885793924331665, "learning_rate": 0.00013782356852129495, "loss": 5.2397, "step": 611900 }, { "epoch": 3.689957911747326, "grad_norm": 2.1494152545928955, "learning_rate": 0.00013780328682439085, "loss": 4.673, "step": 611950 }, { "epoch": 3.6902594035286596, "grad_norm": 1.895846962928772, "learning_rate": 0.00013778300535194983, "loss": 5.2046, "step": 612000 }, { "epoch": 3.6902594035286596, "eval_loss": 5.277981758117676, "eval_runtime": 38.982, "eval_samples_per_second": 13.134, "eval_steps_per_second": 6.567, "eval_tts_loss": 7.537391385631022, "step": 612000 }, { "epoch": 3.6905608953099938, "grad_norm": 2.7870235443115234, "learning_rate": 0.00013776272410434513, "loss": 5.1296, "step": 612050 }, { "epoch": 3.690862387091328, "grad_norm": 2.5064444541931152, "learning_rate": 0.00013774244308195011, "loss": 4.9963, "step": 612100 }, { "epoch": 3.691163878872662, "grad_norm": 2.4081690311431885, "learning_rate": 0.0001377221622851379, "loss": 4.9336, "step": 612150 }, { "epoch": 3.691465370653996, "grad_norm": 3.2063066959381104, "learning_rate": 0.00013770188171428175, "loss": 5.1868, "step": 612200 }, { "epoch": 3.69176686243533, "grad_norm": 1.946374535560608, "learning_rate": 0.00013768160136975495, "loss": 5.0866, "step": 612250 }, { "epoch": 3.692068354216664, "grad_norm": 1.9727240800857544, "learning_rate": 0.0001376613212519307, "loss": 4.8141, "step": 612300 }, { "epoch": 3.692369845997998, "grad_norm": 1.3129044771194458, "learning_rate": 0.00013764104136118217, "loss": 4.4727, "step": 612350 }, { "epoch": 3.6926713377793323, "grad_norm": 2.8575503826141357, "learning_rate": 0.0001376207616978827, "loss": 4.879, "step": 612400 }, { "epoch": 3.692972829560666, "grad_norm": 2.420708179473877, "learning_rate": 0.0001376004822624054, "loss": 5.0434, "step": 612450 }, { "epoch": 3.6932743213420003, "grad_norm": 2.6916472911834717, "learning_rate": 0.00013758020305512354, "loss": 4.8474, "step": 612500 }, { "epoch": 3.693575813123334, "grad_norm": 3.668260097503662, "learning_rate": 0.00013755992407641035, "loss": 5.232, "step": 612550 }, { "epoch": 3.6938773049046683, "grad_norm": 1.8892536163330078, "learning_rate": 0.00013753964532663897, "loss": 4.8241, "step": 612600 }, { "epoch": 3.6941787966860025, "grad_norm": 1.6226139068603516, "learning_rate": 0.0001375193668061826, "loss": 4.7172, "step": 612650 }, { "epoch": 3.6944802884673362, "grad_norm": 1.4716897010803223, "learning_rate": 0.00013749908851541455, "loss": 4.654, "step": 612700 }, { "epoch": 3.6947817802486704, "grad_norm": 2.1146609783172607, "learning_rate": 0.00013747881045470792, "loss": 5.0384, "step": 612750 }, { "epoch": 3.695083272030004, "grad_norm": 3.0663115978240967, "learning_rate": 0.00013745853262443587, "loss": 4.8403, "step": 612800 }, { "epoch": 3.6953847638113384, "grad_norm": 1.7922205924987793, "learning_rate": 0.0001374382550249717, "loss": 5.0558, "step": 612850 }, { "epoch": 3.6956862555926726, "grad_norm": 2.043738842010498, "learning_rate": 0.00013741797765668846, "loss": 4.7859, "step": 612900 }, { "epoch": 3.695987747374007, "grad_norm": 2.1419689655303955, "learning_rate": 0.00013739770051995938, "loss": 4.9455, "step": 612950 }, { "epoch": 3.6962892391553406, "grad_norm": 2.8107752799987793, "learning_rate": 0.0001373774236151577, "loss": 5.3716, "step": 613000 }, { "epoch": 3.696590730936675, "grad_norm": 2.4006764888763428, "learning_rate": 0.00013735714694265645, "loss": 4.6894, "step": 613050 }, { "epoch": 3.6968922227180085, "grad_norm": 2.1181821823120117, "learning_rate": 0.00013733687050282892, "loss": 4.9067, "step": 613100 }, { "epoch": 3.6971937144993428, "grad_norm": 1.6508216857910156, "learning_rate": 0.00013731659429604823, "loss": 4.5186, "step": 613150 }, { "epoch": 3.697495206280677, "grad_norm": 3.425781488418579, "learning_rate": 0.00013729631832268746, "loss": 4.8303, "step": 613200 }, { "epoch": 3.6977966980620107, "grad_norm": 2.0986478328704834, "learning_rate": 0.00013727604258311982, "loss": 4.818, "step": 613250 }, { "epoch": 3.698098189843345, "grad_norm": 1.5456129312515259, "learning_rate": 0.00013725576707771854, "loss": 4.8204, "step": 613300 }, { "epoch": 3.6983996816246787, "grad_norm": 3.092226505279541, "learning_rate": 0.00013723549180685664, "loss": 4.8327, "step": 613350 }, { "epoch": 3.698701173406013, "grad_norm": 1.7990245819091797, "learning_rate": 0.00013721521677090726, "loss": 4.8761, "step": 613400 }, { "epoch": 3.699002665187347, "grad_norm": 2.718445062637329, "learning_rate": 0.00013719494197024363, "loss": 4.593, "step": 613450 }, { "epoch": 3.699304156968681, "grad_norm": 2.2847118377685547, "learning_rate": 0.0001371746674052388, "loss": 5.1199, "step": 613500 }, { "epoch": 3.699605648750015, "grad_norm": 1.3419592380523682, "learning_rate": 0.00013715439307626585, "loss": 4.7048, "step": 613550 }, { "epoch": 3.6999071405313493, "grad_norm": 2.169404983520508, "learning_rate": 0.00013713411898369808, "loss": 4.7381, "step": 613600 }, { "epoch": 3.700208632312683, "grad_norm": 1.2222317457199097, "learning_rate": 0.0001371138451279084, "loss": 4.7665, "step": 613650 }, { "epoch": 3.7005101240940173, "grad_norm": 1.2143924236297607, "learning_rate": 0.00013709357150927003, "loss": 4.6438, "step": 613700 }, { "epoch": 3.7008116158753515, "grad_norm": 2.590919256210327, "learning_rate": 0.00013707329812815612, "loss": 5.2151, "step": 613750 }, { "epoch": 3.701113107656685, "grad_norm": 3.8643949031829834, "learning_rate": 0.00013705302498493962, "loss": 4.9996, "step": 613800 }, { "epoch": 3.7014145994380194, "grad_norm": 1.7328286170959473, "learning_rate": 0.00013703275207999373, "loss": 4.5977, "step": 613850 }, { "epoch": 3.701716091219353, "grad_norm": 3.1914124488830566, "learning_rate": 0.00013701247941369157, "loss": 4.8367, "step": 613900 }, { "epoch": 3.7020175830006874, "grad_norm": 1.9492037296295166, "learning_rate": 0.0001369922069864062, "loss": 4.9753, "step": 613950 }, { "epoch": 3.7023190747820216, "grad_norm": 3.1129260063171387, "learning_rate": 0.00013697193479851065, "loss": 4.7833, "step": 614000 }, { "epoch": 3.7026205665633554, "grad_norm": 2.236359119415283, "learning_rate": 0.0001369516628503781, "loss": 4.8165, "step": 614050 }, { "epoch": 3.7029220583446896, "grad_norm": 2.429589033126831, "learning_rate": 0.00013693139114238155, "loss": 4.996, "step": 614100 }, { "epoch": 3.7032235501260233, "grad_norm": 3.7092840671539307, "learning_rate": 0.00013691111967489408, "loss": 4.509, "step": 614150 }, { "epoch": 3.7035250419073575, "grad_norm": 2.2198588848114014, "learning_rate": 0.00013689084844828884, "loss": 4.9461, "step": 614200 }, { "epoch": 3.7038265336886917, "grad_norm": 2.886615753173828, "learning_rate": 0.00013687057746293872, "loss": 4.8102, "step": 614250 }, { "epoch": 3.704128025470026, "grad_norm": 2.0317342281341553, "learning_rate": 0.00013685030671921694, "loss": 4.6121, "step": 614300 }, { "epoch": 3.7044295172513597, "grad_norm": 2.437106132507324, "learning_rate": 0.00013683003621749648, "loss": 4.9146, "step": 614350 }, { "epoch": 3.704731009032694, "grad_norm": 2.367359161376953, "learning_rate": 0.00013680976595815047, "loss": 4.9062, "step": 614400 }, { "epoch": 3.7050325008140277, "grad_norm": 2.046644449234009, "learning_rate": 0.0001367894959415518, "loss": 4.8044, "step": 614450 }, { "epoch": 3.705333992595362, "grad_norm": 1.920454978942871, "learning_rate": 0.00013676922616807368, "loss": 4.8632, "step": 614500 }, { "epoch": 3.705635484376696, "grad_norm": 1.9560751914978027, "learning_rate": 0.00013674895663808907, "loss": 4.6877, "step": 614550 }, { "epoch": 3.70593697615803, "grad_norm": 2.4462318420410156, "learning_rate": 0.00013672868735197096, "loss": 4.8359, "step": 614600 }, { "epoch": 3.706238467939364, "grad_norm": 2.343668222427368, "learning_rate": 0.00013670841831009246, "loss": 5.2881, "step": 614650 }, { "epoch": 3.706539959720698, "grad_norm": 3.0302023887634277, "learning_rate": 0.00013668814951282653, "loss": 4.7508, "step": 614700 }, { "epoch": 3.706841451502032, "grad_norm": 1.7487784624099731, "learning_rate": 0.00013666788096054622, "loss": 4.8403, "step": 614750 }, { "epoch": 3.7071429432833662, "grad_norm": 0.9453732371330261, "learning_rate": 0.0001366476126536245, "loss": 4.6753, "step": 614800 }, { "epoch": 3.7074444350647, "grad_norm": 2.2047626972198486, "learning_rate": 0.00013662734459243448, "loss": 4.7899, "step": 614850 }, { "epoch": 3.707745926846034, "grad_norm": 2.2246828079223633, "learning_rate": 0.00013660707677734906, "loss": 5.0235, "step": 614900 }, { "epoch": 3.708047418627368, "grad_norm": 1.8205296993255615, "learning_rate": 0.00013658680920874126, "loss": 4.628, "step": 614950 }, { "epoch": 3.708348910408702, "grad_norm": 2.507685422897339, "learning_rate": 0.00013656654188698417, "loss": 4.9505, "step": 615000 }, { "epoch": 3.708348910408702, "eval_loss": 5.272672176361084, "eval_runtime": 39.0802, "eval_samples_per_second": 13.101, "eval_steps_per_second": 6.551, "eval_tts_loss": 7.547465633941684, "step": 615000 }, { "epoch": 3.7086504021900364, "grad_norm": 1.8978418111801147, "learning_rate": 0.0001365462748124506, "loss": 4.6229, "step": 615050 }, { "epoch": 3.7089518939713706, "grad_norm": 2.685906410217285, "learning_rate": 0.0001365260079855137, "loss": 4.8215, "step": 615100 }, { "epoch": 3.7092533857527044, "grad_norm": 6.227535724639893, "learning_rate": 0.00013650574140654643, "loss": 4.9455, "step": 615150 }, { "epoch": 3.7095548775340386, "grad_norm": 1.9605213403701782, "learning_rate": 0.00013648547507592168, "loss": 5.0493, "step": 615200 }, { "epoch": 3.7098563693153723, "grad_norm": 1.5023387670516968, "learning_rate": 0.0001364652089940125, "loss": 4.616, "step": 615250 }, { "epoch": 3.7101578610967065, "grad_norm": 2.4404685497283936, "learning_rate": 0.0001364449431611918, "loss": 4.8976, "step": 615300 }, { "epoch": 3.7104593528780407, "grad_norm": 0.9154278039932251, "learning_rate": 0.0001364246775778326, "loss": 5.0327, "step": 615350 }, { "epoch": 3.7107608446593745, "grad_norm": 2.089820384979248, "learning_rate": 0.00013640441224430783, "loss": 4.7671, "step": 615400 }, { "epoch": 3.7110623364407087, "grad_norm": 3.0638930797576904, "learning_rate": 0.00013638414716099048, "loss": 4.7039, "step": 615450 }, { "epoch": 3.7113638282220425, "grad_norm": 2.392252206802368, "learning_rate": 0.00013636388232825345, "loss": 5.1522, "step": 615500 }, { "epoch": 3.7116653200033767, "grad_norm": 0.9834599494934082, "learning_rate": 0.00013634361774646968, "loss": 4.6479, "step": 615550 }, { "epoch": 3.711966811784711, "grad_norm": 2.0469515323638916, "learning_rate": 0.00013632335341601218, "loss": 4.6064, "step": 615600 }, { "epoch": 3.7122683035660446, "grad_norm": 1.9109275341033936, "learning_rate": 0.0001363030893372538, "loss": 5.1291, "step": 615650 }, { "epoch": 3.712569795347379, "grad_norm": 2.8161087036132812, "learning_rate": 0.00013628282551056753, "loss": 5.0118, "step": 615700 }, { "epoch": 3.7128712871287126, "grad_norm": 2.599735736846924, "learning_rate": 0.00013626256193632633, "loss": 4.9219, "step": 615750 }, { "epoch": 3.713172778910047, "grad_norm": 1.1288630962371826, "learning_rate": 0.000136242298614903, "loss": 4.661, "step": 615800 }, { "epoch": 3.713474270691381, "grad_norm": 1.2560670375823975, "learning_rate": 0.00013622203554667053, "loss": 4.8442, "step": 615850 }, { "epoch": 3.7137757624727152, "grad_norm": 2.5047547817230225, "learning_rate": 0.0001362017727320019, "loss": 5.0683, "step": 615900 }, { "epoch": 3.714077254254049, "grad_norm": 3.3414855003356934, "learning_rate": 0.00013618151017126992, "loss": 5.0059, "step": 615950 }, { "epoch": 3.714378746035383, "grad_norm": 2.4697983264923096, "learning_rate": 0.0001361612478648475, "loss": 4.78, "step": 616000 }, { "epoch": 3.714680237816717, "grad_norm": 4.439869403839111, "learning_rate": 0.0001361409858131076, "loss": 4.9128, "step": 616050 }, { "epoch": 3.714981729598051, "grad_norm": 3.459240198135376, "learning_rate": 0.00013612072401642309, "loss": 4.6706, "step": 616100 }, { "epoch": 3.7152832213793854, "grad_norm": 2.6849849224090576, "learning_rate": 0.00013610046247516678, "loss": 5.0912, "step": 616150 }, { "epoch": 3.715584713160719, "grad_norm": 2.683558940887451, "learning_rate": 0.0001360802011897117, "loss": 5.0825, "step": 616200 }, { "epoch": 3.7158862049420534, "grad_norm": 1.629970669746399, "learning_rate": 0.00013605994016043058, "loss": 4.9921, "step": 616250 }, { "epoch": 3.716187696723387, "grad_norm": 3.553046703338623, "learning_rate": 0.00013603967938769643, "loss": 4.8373, "step": 616300 }, { "epoch": 3.7164891885047213, "grad_norm": 2.34061598777771, "learning_rate": 0.00013601941887188206, "loss": 5.19, "step": 616350 }, { "epoch": 3.7167906802860555, "grad_norm": 3.924569606781006, "learning_rate": 0.00013599915861336028, "loss": 4.7605, "step": 616400 }, { "epoch": 3.7170921720673893, "grad_norm": 3.2241010665893555, "learning_rate": 0.000135978898612504, "loss": 4.9996, "step": 616450 }, { "epoch": 3.7173936638487235, "grad_norm": 1.9068453311920166, "learning_rate": 0.00013595863886968615, "loss": 4.6687, "step": 616500 }, { "epoch": 3.7176951556300573, "grad_norm": 3.065568685531616, "learning_rate": 0.00013593837938527948, "loss": 5.0491, "step": 616550 }, { "epoch": 3.7179966474113915, "grad_norm": 1.8970005512237549, "learning_rate": 0.00013591812015965684, "loss": 5.1162, "step": 616600 }, { "epoch": 3.7182981391927257, "grad_norm": 1.8581348657608032, "learning_rate": 0.00013589786119319117, "loss": 4.9609, "step": 616650 }, { "epoch": 3.71859963097406, "grad_norm": 5.008573055267334, "learning_rate": 0.00013587760248625518, "loss": 4.7346, "step": 616700 }, { "epoch": 3.7189011227553936, "grad_norm": 4.056567668914795, "learning_rate": 0.00013585734403922178, "loss": 4.5369, "step": 616750 }, { "epoch": 3.719202614536728, "grad_norm": 1.8071247339248657, "learning_rate": 0.0001358370858524638, "loss": 5.0778, "step": 616800 }, { "epoch": 3.7195041063180616, "grad_norm": 2.2398810386657715, "learning_rate": 0.000135816827926354, "loss": 4.8392, "step": 616850 }, { "epoch": 3.719805598099396, "grad_norm": 2.3536911010742188, "learning_rate": 0.00013579657026126528, "loss": 5.1212, "step": 616900 }, { "epoch": 3.72010708988073, "grad_norm": 1.8110429048538208, "learning_rate": 0.00013577631285757044, "loss": 4.8608, "step": 616950 }, { "epoch": 3.720408581662064, "grad_norm": 2.5204076766967773, "learning_rate": 0.0001357560557156422, "loss": 4.804, "step": 617000 }, { "epoch": 3.720710073443398, "grad_norm": 2.7447099685668945, "learning_rate": 0.00013573579883585342, "loss": 4.7471, "step": 617050 }, { "epoch": 3.7210115652247318, "grad_norm": 2.1889195442199707, "learning_rate": 0.00013571554221857696, "loss": 5.27, "step": 617100 }, { "epoch": 3.721313057006066, "grad_norm": 2.764200448989868, "learning_rate": 0.00013569528586418556, "loss": 4.724, "step": 617150 }, { "epoch": 3.7216145487874, "grad_norm": 2.2611441612243652, "learning_rate": 0.00013567502977305195, "loss": 4.9656, "step": 617200 }, { "epoch": 3.721916040568734, "grad_norm": 2.3356897830963135, "learning_rate": 0.00013565477394554906, "loss": 4.6858, "step": 617250 }, { "epoch": 3.722217532350068, "grad_norm": 2.016627073287964, "learning_rate": 0.00013563451838204956, "loss": 5.2099, "step": 617300 }, { "epoch": 3.7225190241314023, "grad_norm": 2.37690806388855, "learning_rate": 0.00013561426308292618, "loss": 4.7861, "step": 617350 }, { "epoch": 3.722820515912736, "grad_norm": 2.8036704063415527, "learning_rate": 0.00013559400804855185, "loss": 5.3545, "step": 617400 }, { "epoch": 3.7231220076940703, "grad_norm": 0.6122918128967285, "learning_rate": 0.00013557375327929918, "loss": 5.1237, "step": 617450 }, { "epoch": 3.7234234994754045, "grad_norm": 2.0559468269348145, "learning_rate": 0.000135553498775541, "loss": 4.9865, "step": 617500 }, { "epoch": 3.7237249912567383, "grad_norm": 2.543254852294922, "learning_rate": 0.0001355332445376501, "loss": 5.1301, "step": 617550 }, { "epoch": 3.7240264830380725, "grad_norm": 1.9010919332504272, "learning_rate": 0.00013551299056599914, "loss": 4.6885, "step": 617600 }, { "epoch": 3.7243279748194063, "grad_norm": 2.253547191619873, "learning_rate": 0.0001354927368609609, "loss": 5.0166, "step": 617650 }, { "epoch": 3.7246294666007405, "grad_norm": 2.53474760055542, "learning_rate": 0.00013547248342290818, "loss": 4.5955, "step": 617700 }, { "epoch": 3.7249309583820747, "grad_norm": 2.3533244132995605, "learning_rate": 0.0001354522302522137, "loss": 4.6496, "step": 617750 }, { "epoch": 3.7252324501634084, "grad_norm": 2.6662542819976807, "learning_rate": 0.00013543197734925007, "loss": 4.4903, "step": 617800 }, { "epoch": 3.7255339419447426, "grad_norm": 1.8339308500289917, "learning_rate": 0.00013541172471439016, "loss": 4.6385, "step": 617850 }, { "epoch": 3.7258354337260764, "grad_norm": 2.1841204166412354, "learning_rate": 0.0001353914723480067, "loss": 4.9191, "step": 617900 }, { "epoch": 3.7261369255074106, "grad_norm": 1.8745498657226562, "learning_rate": 0.00013537122025047223, "loss": 4.7856, "step": 617950 }, { "epoch": 3.726438417288745, "grad_norm": 2.1730449199676514, "learning_rate": 0.0001353509684221596, "loss": 4.8414, "step": 618000 }, { "epoch": 3.726438417288745, "eval_loss": 5.274357795715332, "eval_runtime": 39.1265, "eval_samples_per_second": 13.086, "eval_steps_per_second": 6.543, "eval_tts_loss": 7.520687283664353, "step": 618000 }, { "epoch": 3.726739909070079, "grad_norm": 2.0332834720611572, "learning_rate": 0.00013533071686344157, "loss": 4.8871, "step": 618050 }, { "epoch": 3.7270414008514128, "grad_norm": 2.863314151763916, "learning_rate": 0.0001353104655746907, "loss": 4.5223, "step": 618100 }, { "epoch": 3.727342892632747, "grad_norm": 2.7418606281280518, "learning_rate": 0.0001352902145562797, "loss": 4.6773, "step": 618150 }, { "epoch": 3.7276443844140807, "grad_norm": 2.6978108882904053, "learning_rate": 0.00013526996380858143, "loss": 4.8152, "step": 618200 }, { "epoch": 3.727945876195415, "grad_norm": 2.589557409286499, "learning_rate": 0.00013524971333196836, "loss": 5.2468, "step": 618250 }, { "epoch": 3.728247367976749, "grad_norm": 1.8368421792984009, "learning_rate": 0.0001352294631268133, "loss": 4.9894, "step": 618300 }, { "epoch": 3.728548859758083, "grad_norm": 2.014169454574585, "learning_rate": 0.00013520921319348892, "loss": 4.7663, "step": 618350 }, { "epoch": 3.728850351539417, "grad_norm": 2.4656126499176025, "learning_rate": 0.0001351889635323678, "loss": 5.1436, "step": 618400 }, { "epoch": 3.729151843320751, "grad_norm": 2.824708938598633, "learning_rate": 0.0001351687141438227, "loss": 4.5618, "step": 618450 }, { "epoch": 3.729453335102085, "grad_norm": 2.57798433303833, "learning_rate": 0.0001351484650282263, "loss": 5.1211, "step": 618500 }, { "epoch": 3.7297548268834193, "grad_norm": 2.0485429763793945, "learning_rate": 0.00013512821618595115, "loss": 4.8173, "step": 618550 }, { "epoch": 3.730056318664753, "grad_norm": 1.7457071542739868, "learning_rate": 0.00013510796761736994, "loss": 4.524, "step": 618600 }, { "epoch": 3.7303578104460873, "grad_norm": 1.9505881071090698, "learning_rate": 0.0001350877193228554, "loss": 4.9415, "step": 618650 }, { "epoch": 3.730659302227421, "grad_norm": 1.827350378036499, "learning_rate": 0.00013506747130278006, "loss": 4.951, "step": 618700 }, { "epoch": 3.7309607940087552, "grad_norm": 4.253168106079102, "learning_rate": 0.00013504722355751658, "loss": 4.7297, "step": 618750 }, { "epoch": 3.7312622857900895, "grad_norm": 1.852886438369751, "learning_rate": 0.00013502697608743766, "loss": 4.9802, "step": 618800 }, { "epoch": 3.7315637775714237, "grad_norm": 2.0213980674743652, "learning_rate": 0.00013500672889291585, "loss": 4.9768, "step": 618850 }, { "epoch": 3.7318652693527574, "grad_norm": 3.4206578731536865, "learning_rate": 0.0001349864819743238, "loss": 5.0659, "step": 618900 }, { "epoch": 3.7321667611340916, "grad_norm": 2.3591973781585693, "learning_rate": 0.00013496623533203417, "loss": 4.7702, "step": 618950 }, { "epoch": 3.7324682529154254, "grad_norm": 2.403069496154785, "learning_rate": 0.00013494598896641945, "loss": 4.7932, "step": 619000 }, { "epoch": 3.7327697446967596, "grad_norm": 3.3113772869110107, "learning_rate": 0.00013492574287785237, "loss": 4.786, "step": 619050 }, { "epoch": 3.733071236478094, "grad_norm": 1.6597809791564941, "learning_rate": 0.0001349054970667055, "loss": 5.1288, "step": 619100 }, { "epoch": 3.7333727282594276, "grad_norm": 2.6253321170806885, "learning_rate": 0.00013488525153335134, "loss": 5.0218, "step": 619150 }, { "epoch": 3.7336742200407618, "grad_norm": 1.754319190979004, "learning_rate": 0.00013486500627816255, "loss": 5.0138, "step": 619200 }, { "epoch": 3.7339757118220955, "grad_norm": 1.6047636270523071, "learning_rate": 0.0001348447613015118, "loss": 4.4455, "step": 619250 }, { "epoch": 3.7342772036034297, "grad_norm": 0.9894838929176331, "learning_rate": 0.00013482451660377158, "loss": 4.7844, "step": 619300 }, { "epoch": 3.734578695384764, "grad_norm": 3.1596157550811768, "learning_rate": 0.00013480427218531444, "loss": 4.9687, "step": 619350 }, { "epoch": 3.7348801871660977, "grad_norm": 2.205878734588623, "learning_rate": 0.00013478402804651302, "loss": 5.1644, "step": 619400 }, { "epoch": 3.735181678947432, "grad_norm": 2.62137508392334, "learning_rate": 0.0001347637841877398, "loss": 5.0101, "step": 619450 }, { "epoch": 3.7354831707287657, "grad_norm": 2.9296257495880127, "learning_rate": 0.0001347435406093674, "loss": 4.9768, "step": 619500 }, { "epoch": 3.7357846625101, "grad_norm": 2.4086716175079346, "learning_rate": 0.00013472329731176844, "loss": 5.0919, "step": 619550 }, { "epoch": 3.736086154291434, "grad_norm": 1.9480268955230713, "learning_rate": 0.0001347030542953153, "loss": 4.8112, "step": 619600 }, { "epoch": 3.7363876460727683, "grad_norm": 5.876729488372803, "learning_rate": 0.00013468281156038065, "loss": 4.7812, "step": 619650 }, { "epoch": 3.736689137854102, "grad_norm": 1.505815863609314, "learning_rate": 0.000134662569107337, "loss": 4.7167, "step": 619700 }, { "epoch": 3.7369906296354363, "grad_norm": 4.556541919708252, "learning_rate": 0.00013464232693655685, "loss": 4.8995, "step": 619750 }, { "epoch": 3.73729212141677, "grad_norm": 3.891157627105713, "learning_rate": 0.00013462208504841274, "loss": 4.715, "step": 619800 }, { "epoch": 3.7375936131981042, "grad_norm": 2.3017361164093018, "learning_rate": 0.00013460184344327728, "loss": 4.8147, "step": 619850 }, { "epoch": 3.7378951049794384, "grad_norm": 1.1486026048660278, "learning_rate": 0.00013458160212152288, "loss": 4.7507, "step": 619900 }, { "epoch": 3.738196596760772, "grad_norm": 1.1426860094070435, "learning_rate": 0.00013456136108352205, "loss": 4.6275, "step": 619950 }, { "epoch": 3.7384980885421064, "grad_norm": 2.6230361461639404, "learning_rate": 0.00013454112032964744, "loss": 4.8555, "step": 620000 }, { "epoch": 3.73879958032344, "grad_norm": 2.133863687515259, "learning_rate": 0.00013452087986027134, "loss": 4.9228, "step": 620050 }, { "epoch": 3.7391010721047744, "grad_norm": 2.302809715270996, "learning_rate": 0.00013450063967576642, "loss": 4.5467, "step": 620100 }, { "epoch": 3.7394025638861086, "grad_norm": 3.086359739303589, "learning_rate": 0.0001344803997765051, "loss": 4.3264, "step": 620150 }, { "epoch": 3.7397040556674424, "grad_norm": 1.6009048223495483, "learning_rate": 0.00013446016016285985, "loss": 5.0025, "step": 620200 }, { "epoch": 3.7400055474487766, "grad_norm": 1.0492205619812012, "learning_rate": 0.0001344399208352032, "loss": 4.8886, "step": 620250 }, { "epoch": 3.7403070392301103, "grad_norm": 3.095560073852539, "learning_rate": 0.0001344196817939076, "loss": 5.2827, "step": 620300 }, { "epoch": 3.7406085310114445, "grad_norm": 1.8643646240234375, "learning_rate": 0.00013439944303934552, "loss": 4.824, "step": 620350 }, { "epoch": 3.7409100227927787, "grad_norm": 2.7438242435455322, "learning_rate": 0.00013437920457188939, "loss": 4.8018, "step": 620400 }, { "epoch": 3.741211514574113, "grad_norm": 2.403240442276001, "learning_rate": 0.00013435896639191178, "loss": 4.7924, "step": 620450 }, { "epoch": 3.7415130063554467, "grad_norm": 2.442481279373169, "learning_rate": 0.00013433872849978503, "loss": 4.9842, "step": 620500 }, { "epoch": 3.741814498136781, "grad_norm": 1.8536934852600098, "learning_rate": 0.0001343184908958816, "loss": 4.8912, "step": 620550 }, { "epoch": 3.7421159899181147, "grad_norm": 2.310462474822998, "learning_rate": 0.00013429825358057405, "loss": 4.8181, "step": 620600 }, { "epoch": 3.742417481699449, "grad_norm": 1.888171911239624, "learning_rate": 0.00013427801655423464, "loss": 4.6863, "step": 620650 }, { "epoch": 3.742718973480783, "grad_norm": 1.6017988920211792, "learning_rate": 0.00013425777981723594, "loss": 4.9126, "step": 620700 }, { "epoch": 3.743020465262117, "grad_norm": 2.1352434158325195, "learning_rate": 0.0001342375433699504, "loss": 4.4109, "step": 620750 }, { "epoch": 3.743321957043451, "grad_norm": 1.9274137020111084, "learning_rate": 0.0001342173072127503, "loss": 4.6308, "step": 620800 }, { "epoch": 3.743623448824785, "grad_norm": 1.8273987770080566, "learning_rate": 0.00013419707134600814, "loss": 4.5382, "step": 620850 }, { "epoch": 3.743924940606119, "grad_norm": 2.36376953125, "learning_rate": 0.00013417683577009637, "loss": 4.972, "step": 620900 }, { "epoch": 3.7442264323874532, "grad_norm": 1.7731367349624634, "learning_rate": 0.00013415660048538735, "loss": 4.8832, "step": 620950 }, { "epoch": 3.7445279241687874, "grad_norm": 2.714452028274536, "learning_rate": 0.00013413636549225344, "loss": 4.7913, "step": 621000 }, { "epoch": 3.7445279241687874, "eval_loss": 5.282707214355469, "eval_runtime": 39.2681, "eval_samples_per_second": 13.039, "eval_steps_per_second": 6.519, "eval_tts_loss": 7.493367973018023, "step": 621000 }, { "epoch": 3.744829415950121, "grad_norm": 2.5874006748199463, "learning_rate": 0.00013411613079106715, "loss": 4.82, "step": 621050 }, { "epoch": 3.7451309077314554, "grad_norm": 1.3939225673675537, "learning_rate": 0.00013409589638220076, "loss": 4.5366, "step": 621100 }, { "epoch": 3.745432399512789, "grad_norm": 2.369868516921997, "learning_rate": 0.0001340756622660267, "loss": 4.9791, "step": 621150 }, { "epoch": 3.7457338912941234, "grad_norm": 2.669572591781616, "learning_rate": 0.00013405542844291735, "loss": 4.597, "step": 621200 }, { "epoch": 3.7460353830754576, "grad_norm": 4.605545997619629, "learning_rate": 0.00013403519491324512, "loss": 4.9128, "step": 621250 }, { "epoch": 3.7463368748567913, "grad_norm": 2.128680467605591, "learning_rate": 0.00013401496167738235, "loss": 4.5058, "step": 621300 }, { "epoch": 3.7466383666381256, "grad_norm": 1.2979190349578857, "learning_rate": 0.00013399472873570133, "loss": 4.5092, "step": 621350 }, { "epoch": 3.7469398584194593, "grad_norm": 2.253652572631836, "learning_rate": 0.00013397449608857457, "loss": 4.8883, "step": 621400 }, { "epoch": 3.7472413502007935, "grad_norm": 2.3304333686828613, "learning_rate": 0.00013395426373637428, "loss": 4.7794, "step": 621450 }, { "epoch": 3.7475428419821277, "grad_norm": 2.3342201709747314, "learning_rate": 0.0001339340316794729, "loss": 4.7077, "step": 621500 }, { "epoch": 3.7478443337634615, "grad_norm": 2.1103508472442627, "learning_rate": 0.00013391379991824277, "loss": 4.8584, "step": 621550 }, { "epoch": 3.7481458255447957, "grad_norm": 2.3885278701782227, "learning_rate": 0.0001338935684530561, "loss": 4.9378, "step": 621600 }, { "epoch": 3.7484473173261295, "grad_norm": 2.523428440093994, "learning_rate": 0.0001338733372842854, "loss": 5.277, "step": 621650 }, { "epoch": 3.7487488091074637, "grad_norm": 2.5961685180664062, "learning_rate": 0.0001338531064123029, "loss": 4.9272, "step": 621700 }, { "epoch": 3.749050300888798, "grad_norm": 2.7718045711517334, "learning_rate": 0.0001338328758374809, "loss": 4.7402, "step": 621750 }, { "epoch": 3.749351792670132, "grad_norm": 3.0580239295959473, "learning_rate": 0.00013381264556019173, "loss": 4.9175, "step": 621800 }, { "epoch": 3.749653284451466, "grad_norm": 1.9800785779953003, "learning_rate": 0.0001337924155808078, "loss": 5.1521, "step": 621850 }, { "epoch": 3.7499547762328, "grad_norm": 3.4763095378875732, "learning_rate": 0.00013377218589970128, "loss": 4.8737, "step": 621900 }, { "epoch": 3.750256268014134, "grad_norm": 2.812737464904785, "learning_rate": 0.00013375195651724452, "loss": 4.5895, "step": 621950 }, { "epoch": 3.750557759795468, "grad_norm": 0.8898810148239136, "learning_rate": 0.00013373172743380985, "loss": 4.6635, "step": 622000 }, { "epoch": 3.7508592515768022, "grad_norm": 2.9258217811584473, "learning_rate": 0.00013371149864976946, "loss": 4.5361, "step": 622050 }, { "epoch": 3.751160743358136, "grad_norm": 2.250490665435791, "learning_rate": 0.0001336912701654957, "loss": 4.9922, "step": 622100 }, { "epoch": 3.75146223513947, "grad_norm": 1.7761558294296265, "learning_rate": 0.0001336710419813609, "loss": 4.9697, "step": 622150 }, { "epoch": 3.751763726920804, "grad_norm": 2.0731682777404785, "learning_rate": 0.00013365081409773722, "loss": 4.8928, "step": 622200 }, { "epoch": 3.752065218702138, "grad_norm": 2.5402982234954834, "learning_rate": 0.000133630586514997, "loss": 4.6824, "step": 622250 }, { "epoch": 3.7523667104834724, "grad_norm": 2.724374771118164, "learning_rate": 0.0001336103592335125, "loss": 5.2191, "step": 622300 }, { "epoch": 3.752668202264806, "grad_norm": 3.032935857772827, "learning_rate": 0.00013359013225365586, "loss": 4.9201, "step": 622350 }, { "epoch": 3.7529696940461403, "grad_norm": 0.5675613284111023, "learning_rate": 0.00013356990557579946, "loss": 4.7974, "step": 622400 }, { "epoch": 3.753271185827474, "grad_norm": 1.8301621675491333, "learning_rate": 0.00013354967920031554, "loss": 4.8785, "step": 622450 }, { "epoch": 3.7535726776088083, "grad_norm": 2.73586106300354, "learning_rate": 0.00013352945312757627, "loss": 4.6068, "step": 622500 }, { "epoch": 3.7538741693901425, "grad_norm": 2.8795998096466064, "learning_rate": 0.0001335092273579539, "loss": 4.8008, "step": 622550 }, { "epoch": 3.7541756611714767, "grad_norm": 1.9367811679840088, "learning_rate": 0.0001334890018918207, "loss": 5.0041, "step": 622600 }, { "epoch": 3.7544771529528105, "grad_norm": 1.5165984630584717, "learning_rate": 0.0001334687767295488, "loss": 4.8625, "step": 622650 }, { "epoch": 3.7547786447341447, "grad_norm": 2.7579681873321533, "learning_rate": 0.0001334485518715105, "loss": 4.9422, "step": 622700 }, { "epoch": 3.7550801365154785, "grad_norm": 2.388282060623169, "learning_rate": 0.00013342832731807802, "loss": 5.0145, "step": 622750 }, { "epoch": 3.7553816282968127, "grad_norm": 2.894110918045044, "learning_rate": 0.00013340810306962348, "loss": 5.0273, "step": 622800 }, { "epoch": 3.755683120078147, "grad_norm": 2.7004172801971436, "learning_rate": 0.00013338787912651914, "loss": 4.9404, "step": 622850 }, { "epoch": 3.7559846118594806, "grad_norm": 2.0923118591308594, "learning_rate": 0.0001333676554891372, "loss": 5.101, "step": 622900 }, { "epoch": 3.756286103640815, "grad_norm": 1.413084626197815, "learning_rate": 0.00013334743215784978, "loss": 4.8216, "step": 622950 }, { "epoch": 3.7565875954221486, "grad_norm": 2.61320161819458, "learning_rate": 0.0001333272091330291, "loss": 4.9173, "step": 623000 }, { "epoch": 3.756889087203483, "grad_norm": 2.684671640396118, "learning_rate": 0.0001333069864150474, "loss": 5.0527, "step": 623050 }, { "epoch": 3.757190578984817, "grad_norm": 1.857029914855957, "learning_rate": 0.0001332867640042768, "loss": 4.9667, "step": 623100 }, { "epoch": 3.7574920707661508, "grad_norm": 1.0038987398147583, "learning_rate": 0.00013326654190108937, "loss": 4.3433, "step": 623150 }, { "epoch": 3.757793562547485, "grad_norm": 1.485772967338562, "learning_rate": 0.00013324632010585746, "loss": 4.7309, "step": 623200 }, { "epoch": 3.7580950543288187, "grad_norm": 1.9762756824493408, "learning_rate": 0.00013322609861895303, "loss": 4.7399, "step": 623250 }, { "epoch": 3.758396546110153, "grad_norm": 3.3228282928466797, "learning_rate": 0.00013320587744074838, "loss": 4.9012, "step": 623300 }, { "epoch": 3.758698037891487, "grad_norm": 2.5335378646850586, "learning_rate": 0.0001331856565716156, "loss": 4.8569, "step": 623350 }, { "epoch": 3.7589995296728214, "grad_norm": 0.8376734256744385, "learning_rate": 0.00013316543601192673, "loss": 4.7796, "step": 623400 }, { "epoch": 3.759301021454155, "grad_norm": 3.4723055362701416, "learning_rate": 0.00013314521576205404, "loss": 4.8454, "step": 623450 }, { "epoch": 3.7596025132354893, "grad_norm": 3.5350208282470703, "learning_rate": 0.00013312499582236964, "loss": 4.5091, "step": 623500 }, { "epoch": 3.759904005016823, "grad_norm": 2.6589674949645996, "learning_rate": 0.00013310477619324555, "loss": 4.7948, "step": 623550 }, { "epoch": 3.7602054967981573, "grad_norm": 3.038330316543579, "learning_rate": 0.0001330845568750539, "loss": 5.28, "step": 623600 }, { "epoch": 3.7605069885794915, "grad_norm": 2.001819133758545, "learning_rate": 0.00013306433786816697, "loss": 4.8877, "step": 623650 }, { "epoch": 3.7608084803608253, "grad_norm": 1.864219307899475, "learning_rate": 0.00013304411917295665, "loss": 5.1207, "step": 623700 }, { "epoch": 3.7611099721421595, "grad_norm": 2.3154659271240234, "learning_rate": 0.00013302390078979512, "loss": 4.8027, "step": 623750 }, { "epoch": 3.7614114639234932, "grad_norm": 2.014853000640869, "learning_rate": 0.0001330036827190545, "loss": 4.9327, "step": 623800 }, { "epoch": 3.7617129557048274, "grad_norm": 1.5696920156478882, "learning_rate": 0.00013298346496110682, "loss": 4.9294, "step": 623850 }, { "epoch": 3.7620144474861617, "grad_norm": 2.473266124725342, "learning_rate": 0.0001329632475163242, "loss": 5.2851, "step": 623900 }, { "epoch": 3.7623159392674954, "grad_norm": 2.85719633102417, "learning_rate": 0.0001329430303850787, "loss": 4.9003, "step": 623950 }, { "epoch": 3.7626174310488296, "grad_norm": 2.611326217651367, "learning_rate": 0.00013292281356774236, "loss": 4.7599, "step": 624000 }, { "epoch": 3.7626174310488296, "eval_loss": 5.250543594360352, "eval_runtime": 39.1171, "eval_samples_per_second": 13.089, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.5401596817852, "step": 624000 }, { "epoch": 3.7629189228301634, "grad_norm": 3.1498334407806396, "learning_rate": 0.00013290259706468728, "loss": 4.9403, "step": 624050 }, { "epoch": 3.7632204146114976, "grad_norm": 1.952913761138916, "learning_rate": 0.00013288238087628553, "loss": 4.6407, "step": 624100 }, { "epoch": 3.763521906392832, "grad_norm": 2.561537504196167, "learning_rate": 0.00013286216500290907, "loss": 4.6148, "step": 624150 }, { "epoch": 3.763823398174166, "grad_norm": 1.854651689529419, "learning_rate": 0.00013284194944493, "loss": 4.9987, "step": 624200 }, { "epoch": 3.7641248899554998, "grad_norm": 2.4200751781463623, "learning_rate": 0.00013282173420272047, "loss": 5.0369, "step": 624250 }, { "epoch": 3.764426381736834, "grad_norm": 2.3923280239105225, "learning_rate": 0.00013280151927665232, "loss": 4.701, "step": 624300 }, { "epoch": 3.7647278735181677, "grad_norm": 0.9900527596473694, "learning_rate": 0.00013278130466709765, "loss": 5.2636, "step": 624350 }, { "epoch": 3.765029365299502, "grad_norm": 2.4005985260009766, "learning_rate": 0.00013276109037442857, "loss": 4.7048, "step": 624400 }, { "epoch": 3.765330857080836, "grad_norm": 2.899778366088867, "learning_rate": 0.00013274087639901698, "loss": 5.0965, "step": 624450 }, { "epoch": 3.76563234886217, "grad_norm": 0.8854833841323853, "learning_rate": 0.00013272066274123493, "loss": 4.9029, "step": 624500 }, { "epoch": 3.765933840643504, "grad_norm": 1.838301181793213, "learning_rate": 0.0001327004494014544, "loss": 4.8402, "step": 624550 }, { "epoch": 3.766235332424838, "grad_norm": 1.3437390327453613, "learning_rate": 0.00013268023638004744, "loss": 4.92, "step": 624600 }, { "epoch": 3.766536824206172, "grad_norm": 2.0205078125, "learning_rate": 0.00013266002367738603, "loss": 4.6603, "step": 624650 }, { "epoch": 3.7668383159875063, "grad_norm": 1.9397178888320923, "learning_rate": 0.00013263981129384207, "loss": 4.848, "step": 624700 }, { "epoch": 3.7671398077688405, "grad_norm": 2.729694366455078, "learning_rate": 0.0001326195992297877, "loss": 4.827, "step": 624750 }, { "epoch": 3.7674412995501743, "grad_norm": 2.5904390811920166, "learning_rate": 0.00013259938748559473, "loss": 4.7739, "step": 624800 }, { "epoch": 3.7677427913315085, "grad_norm": 3.0089046955108643, "learning_rate": 0.00013257917606163523, "loss": 4.8138, "step": 624850 }, { "epoch": 3.7680442831128422, "grad_norm": 2.396883964538574, "learning_rate": 0.00013255896495828118, "loss": 4.8955, "step": 624900 }, { "epoch": 3.7683457748941764, "grad_norm": 2.2391350269317627, "learning_rate": 0.0001325387541759044, "loss": 4.65, "step": 624950 }, { "epoch": 3.7686472666755106, "grad_norm": 2.320666790008545, "learning_rate": 0.00013251854371487695, "loss": 4.8009, "step": 625000 }, { "epoch": 3.7689487584568444, "grad_norm": 2.48933482170105, "learning_rate": 0.00013249833357557083, "loss": 5.0309, "step": 625050 }, { "epoch": 3.7692502502381786, "grad_norm": 2.1376705169677734, "learning_rate": 0.00013247812375835784, "loss": 4.8422, "step": 625100 }, { "epoch": 3.7695517420195124, "grad_norm": 2.4142632484436035, "learning_rate": 0.00013245791426360996, "loss": 5.2003, "step": 625150 }, { "epoch": 3.7698532338008466, "grad_norm": 2.238002300262451, "learning_rate": 0.0001324377050916992, "loss": 4.8153, "step": 625200 }, { "epoch": 3.770154725582181, "grad_norm": 1.868399739265442, "learning_rate": 0.0001324174962429974, "loss": 4.6806, "step": 625250 }, { "epoch": 3.7704562173635146, "grad_norm": 1.6063666343688965, "learning_rate": 0.00013239728771787643, "loss": 4.517, "step": 625300 }, { "epoch": 3.7707577091448488, "grad_norm": 3.8139138221740723, "learning_rate": 0.00013237707951670833, "loss": 4.8844, "step": 625350 }, { "epoch": 3.7710592009261825, "grad_norm": 2.1913113594055176, "learning_rate": 0.00013235687163986486, "loss": 4.838, "step": 625400 }, { "epoch": 3.7713606927075167, "grad_norm": 2.1929891109466553, "learning_rate": 0.00013233666408771805, "loss": 4.3828, "step": 625450 }, { "epoch": 3.771662184488851, "grad_norm": 0.9661460518836975, "learning_rate": 0.00013231645686063976, "loss": 4.6617, "step": 625500 }, { "epoch": 3.771963676270185, "grad_norm": 2.087026834487915, "learning_rate": 0.00013229624995900175, "loss": 4.6615, "step": 625550 }, { "epoch": 3.772265168051519, "grad_norm": 3.4444081783294678, "learning_rate": 0.000132276043383176, "loss": 4.9857, "step": 625600 }, { "epoch": 3.772566659832853, "grad_norm": 2.105607509613037, "learning_rate": 0.00013225583713353446, "loss": 5.1752, "step": 625650 }, { "epoch": 3.772868151614187, "grad_norm": 2.2552828788757324, "learning_rate": 0.0001322356312104489, "loss": 5.0373, "step": 625700 }, { "epoch": 3.773169643395521, "grad_norm": 1.6577812433242798, "learning_rate": 0.00013221542561429113, "loss": 4.9648, "step": 625750 }, { "epoch": 3.7734711351768553, "grad_norm": 1.9436736106872559, "learning_rate": 0.00013219522034543316, "loss": 4.8551, "step": 625800 }, { "epoch": 3.773772626958189, "grad_norm": 1.8608633279800415, "learning_rate": 0.0001321750154042467, "loss": 4.8333, "step": 625850 }, { "epoch": 3.7740741187395233, "grad_norm": 2.038548469543457, "learning_rate": 0.00013215481079110363, "loss": 4.4944, "step": 625900 }, { "epoch": 3.774375610520857, "grad_norm": 2.5364654064178467, "learning_rate": 0.00013213460650637584, "loss": 4.8332, "step": 625950 }, { "epoch": 3.7746771023021912, "grad_norm": 2.070326328277588, "learning_rate": 0.00013211440255043506, "loss": 5.037, "step": 626000 }, { "epoch": 3.7749785940835254, "grad_norm": 1.548756718635559, "learning_rate": 0.00013209419892365324, "loss": 4.8913, "step": 626050 }, { "epoch": 3.775280085864859, "grad_norm": 1.976391077041626, "learning_rate": 0.00013207399562640214, "loss": 4.4473, "step": 626100 }, { "epoch": 3.7755815776461934, "grad_norm": 1.8909624814987183, "learning_rate": 0.00013205379265905352, "loss": 4.8815, "step": 626150 }, { "epoch": 3.775883069427527, "grad_norm": 1.991448998451233, "learning_rate": 0.00013203359002197923, "loss": 5.0524, "step": 626200 }, { "epoch": 3.7761845612088614, "grad_norm": 2.177290201187134, "learning_rate": 0.00013201338771555115, "loss": 4.6766, "step": 626250 }, { "epoch": 3.7764860529901956, "grad_norm": 2.661940813064575, "learning_rate": 0.00013199318574014094, "loss": 5.0739, "step": 626300 }, { "epoch": 3.77678754477153, "grad_norm": 2.8825762271881104, "learning_rate": 0.00013197298409612046, "loss": 4.9563, "step": 626350 }, { "epoch": 3.7770890365528635, "grad_norm": 3.602350950241089, "learning_rate": 0.0001319527827838615, "loss": 4.9002, "step": 626400 }, { "epoch": 3.7773905283341978, "grad_norm": 2.5372400283813477, "learning_rate": 0.00013193258180373577, "loss": 4.9219, "step": 626450 }, { "epoch": 3.7776920201155315, "grad_norm": 1.8459528684616089, "learning_rate": 0.00013191238115611514, "loss": 4.7321, "step": 626500 }, { "epoch": 3.7779935118968657, "grad_norm": 2.5624477863311768, "learning_rate": 0.00013189218084137135, "loss": 4.7147, "step": 626550 }, { "epoch": 3.7782950036782, "grad_norm": 1.7700536251068115, "learning_rate": 0.00013187198085987605, "loss": 5.0201, "step": 626600 }, { "epoch": 3.7785964954595337, "grad_norm": 0.898906946182251, "learning_rate": 0.0001318517812120011, "loss": 5.149, "step": 626650 }, { "epoch": 3.778897987240868, "grad_norm": 2.5587778091430664, "learning_rate": 0.00013183158189811826, "loss": 5.0298, "step": 626700 }, { "epoch": 3.7791994790222017, "grad_norm": 1.8426259756088257, "learning_rate": 0.00013181138291859916, "loss": 5.3096, "step": 626750 }, { "epoch": 3.779500970803536, "grad_norm": 2.648020029067993, "learning_rate": 0.0001317911842738156, "loss": 5.0478, "step": 626800 }, { "epoch": 3.77980246258487, "grad_norm": 1.9497305154800415, "learning_rate": 0.0001317709859641394, "loss": 4.6564, "step": 626850 }, { "epoch": 3.780103954366204, "grad_norm": 2.974196434020996, "learning_rate": 0.00013175078798994213, "loss": 4.6112, "step": 626900 }, { "epoch": 3.780405446147538, "grad_norm": 2.012526273727417, "learning_rate": 0.00013173059035159553, "loss": 4.7381, "step": 626950 }, { "epoch": 3.780706937928872, "grad_norm": 1.2859305143356323, "learning_rate": 0.0001317103930494714, "loss": 4.2269, "step": 627000 }, { "epoch": 3.780706937928872, "eval_loss": 5.267887115478516, "eval_runtime": 38.9266, "eval_samples_per_second": 13.153, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.524424001991077, "step": 627000 }, { "epoch": 3.781008429710206, "grad_norm": 2.8158977031707764, "learning_rate": 0.00013169019608394136, "loss": 5.0852, "step": 627050 }, { "epoch": 3.78130992149154, "grad_norm": 1.953418254852295, "learning_rate": 0.00013166999945537713, "loss": 5.1298, "step": 627100 }, { "epoch": 3.7816114132728744, "grad_norm": 2.559349536895752, "learning_rate": 0.00013164980316415043, "loss": 4.7323, "step": 627150 }, { "epoch": 3.781912905054208, "grad_norm": 2.4624483585357666, "learning_rate": 0.00013162960721063287, "loss": 4.7825, "step": 627200 }, { "epoch": 3.7822143968355424, "grad_norm": 2.661773204803467, "learning_rate": 0.0001316094115951962, "loss": 4.6973, "step": 627250 }, { "epoch": 3.782515888616876, "grad_norm": 2.7948803901672363, "learning_rate": 0.00013158921631821208, "loss": 5.0992, "step": 627300 }, { "epoch": 3.7828173803982104, "grad_norm": 2.192838668823242, "learning_rate": 0.00013156902138005213, "loss": 5.1074, "step": 627350 }, { "epoch": 3.7831188721795446, "grad_norm": 1.9786823987960815, "learning_rate": 0.000131548826781088, "loss": 4.6785, "step": 627400 }, { "epoch": 3.7834203639608783, "grad_norm": 2.6435177326202393, "learning_rate": 0.0001315286325216915, "loss": 4.9228, "step": 627450 }, { "epoch": 3.7837218557422125, "grad_norm": 2.8137102127075195, "learning_rate": 0.00013150843860223408, "loss": 5.2237, "step": 627500 }, { "epoch": 3.7840233475235463, "grad_norm": 2.5535085201263428, "learning_rate": 0.00013148824502308743, "loss": 5.0104, "step": 627550 }, { "epoch": 3.7843248393048805, "grad_norm": 2.130669355392456, "learning_rate": 0.00013146805178462328, "loss": 4.9823, "step": 627600 }, { "epoch": 3.7846263310862147, "grad_norm": 3.3886983394622803, "learning_rate": 0.00013144785888721315, "loss": 4.7375, "step": 627650 }, { "epoch": 3.7849278228675485, "grad_norm": 4.255053520202637, "learning_rate": 0.0001314276663312287, "loss": 4.8348, "step": 627700 }, { "epoch": 3.7852293146488827, "grad_norm": 2.2134897708892822, "learning_rate": 0.0001314074741170416, "loss": 4.5011, "step": 627750 }, { "epoch": 3.7855308064302164, "grad_norm": 1.8626762628555298, "learning_rate": 0.00013138728224502334, "loss": 4.6468, "step": 627800 }, { "epoch": 3.7858322982115507, "grad_norm": 2.6402628421783447, "learning_rate": 0.0001313670907155456, "loss": 4.5174, "step": 627850 }, { "epoch": 3.786133789992885, "grad_norm": 1.8727480173110962, "learning_rate": 0.00013134689952897996, "loss": 4.5393, "step": 627900 }, { "epoch": 3.786435281774219, "grad_norm": 2.5641443729400635, "learning_rate": 0.00013132670868569807, "loss": 4.9524, "step": 627950 }, { "epoch": 3.786736773555553, "grad_norm": 2.5416910648345947, "learning_rate": 0.0001313065181860714, "loss": 4.5439, "step": 628000 }, { "epoch": 3.787038265336887, "grad_norm": 3.39707088470459, "learning_rate": 0.00013128632803047162, "loss": 4.6214, "step": 628050 }, { "epoch": 3.787339757118221, "grad_norm": 1.9889335632324219, "learning_rate": 0.00013126613821927028, "loss": 5.1446, "step": 628100 }, { "epoch": 3.787641248899555, "grad_norm": 1.7889268398284912, "learning_rate": 0.00013124594875283887, "loss": 4.7614, "step": 628150 }, { "epoch": 3.787942740680889, "grad_norm": 0.9162688255310059, "learning_rate": 0.00013122575963154902, "loss": 4.4523, "step": 628200 }, { "epoch": 3.788244232462223, "grad_norm": 2.898714780807495, "learning_rate": 0.00013120557085577234, "loss": 4.999, "step": 628250 }, { "epoch": 3.788545724243557, "grad_norm": 2.152024269104004, "learning_rate": 0.00013118538242588027, "loss": 4.8731, "step": 628300 }, { "epoch": 3.788847216024891, "grad_norm": 2.384878158569336, "learning_rate": 0.00013116519434224437, "loss": 5.042, "step": 628350 }, { "epoch": 3.789148707806225, "grad_norm": 2.176058292388916, "learning_rate": 0.00013114500660523623, "loss": 4.3688, "step": 628400 }, { "epoch": 3.7894501995875594, "grad_norm": 2.556892156600952, "learning_rate": 0.0001311248192152273, "loss": 4.5588, "step": 628450 }, { "epoch": 3.7897516913688936, "grad_norm": 1.9190810918807983, "learning_rate": 0.0001311046321725891, "loss": 4.7464, "step": 628500 }, { "epoch": 3.7900531831502273, "grad_norm": 1.808806300163269, "learning_rate": 0.00013108444547769326, "loss": 5.3153, "step": 628550 }, { "epoch": 3.7903546749315615, "grad_norm": 3.9105594158172607, "learning_rate": 0.00013106425913091114, "loss": 5.0506, "step": 628600 }, { "epoch": 3.7906561667128953, "grad_norm": 2.6113505363464355, "learning_rate": 0.00013104407313261436, "loss": 5.0014, "step": 628650 }, { "epoch": 3.7909576584942295, "grad_norm": 1.7454777956008911, "learning_rate": 0.00013102388748317437, "loss": 4.9354, "step": 628700 }, { "epoch": 3.7912591502755637, "grad_norm": 1.762858271598816, "learning_rate": 0.0001310037021829626, "loss": 4.6101, "step": 628750 }, { "epoch": 3.7915606420568975, "grad_norm": 2.678649663925171, "learning_rate": 0.00013098351723235055, "loss": 4.8105, "step": 628800 }, { "epoch": 3.7918621338382317, "grad_norm": 1.6184555292129517, "learning_rate": 0.00013096333263170983, "loss": 4.9124, "step": 628850 }, { "epoch": 3.7921636256195654, "grad_norm": 2.4851410388946533, "learning_rate": 0.00013094314838141177, "loss": 4.9213, "step": 628900 }, { "epoch": 3.7924651174008996, "grad_norm": 3.0606601238250732, "learning_rate": 0.0001309229644818278, "loss": 4.9331, "step": 628950 }, { "epoch": 3.792766609182234, "grad_norm": 1.8655649423599243, "learning_rate": 0.00013090278093332955, "loss": 4.8303, "step": 629000 }, { "epoch": 3.7930681009635676, "grad_norm": 2.264491558074951, "learning_rate": 0.0001308825977362883, "loss": 5.0236, "step": 629050 }, { "epoch": 3.793369592744902, "grad_norm": 2.498098611831665, "learning_rate": 0.00013086241489107555, "loss": 4.9577, "step": 629100 }, { "epoch": 3.7936710845262356, "grad_norm": 3.293841600418091, "learning_rate": 0.0001308422323980628, "loss": 4.5781, "step": 629150 }, { "epoch": 3.79397257630757, "grad_norm": 2.5371882915496826, "learning_rate": 0.00013082205025762136, "loss": 5.0168, "step": 629200 }, { "epoch": 3.794274068088904, "grad_norm": 1.6503299474716187, "learning_rate": 0.00013080186847012274, "loss": 4.8538, "step": 629250 }, { "epoch": 3.794575559870238, "grad_norm": 3.0197980403900146, "learning_rate": 0.00013078168703593837, "loss": 4.9336, "step": 629300 }, { "epoch": 3.794877051651572, "grad_norm": 1.9903016090393066, "learning_rate": 0.00013076150595543956, "loss": 4.9064, "step": 629350 }, { "epoch": 3.795178543432906, "grad_norm": 2.820178985595703, "learning_rate": 0.00013074132522899779, "loss": 4.7784, "step": 629400 }, { "epoch": 3.79548003521424, "grad_norm": 2.015460729598999, "learning_rate": 0.0001307211448569845, "loss": 4.8642, "step": 629450 }, { "epoch": 3.795781526995574, "grad_norm": 2.85121488571167, "learning_rate": 0.000130700964839771, "loss": 4.3825, "step": 629500 }, { "epoch": 3.7960830187769083, "grad_norm": 1.9574040174484253, "learning_rate": 0.00013068078517772865, "loss": 4.8634, "step": 629550 }, { "epoch": 3.796384510558242, "grad_norm": 1.3781110048294067, "learning_rate": 0.00013066060587122895, "loss": 4.97, "step": 629600 }, { "epoch": 3.7966860023395763, "grad_norm": 2.3600351810455322, "learning_rate": 0.0001306404269206432, "loss": 4.8782, "step": 629650 }, { "epoch": 3.79698749412091, "grad_norm": 2.5108118057250977, "learning_rate": 0.0001306202483263427, "loss": 4.5432, "step": 629700 }, { "epoch": 3.7972889859022443, "grad_norm": 2.8098106384277344, "learning_rate": 0.00013060007008869893, "loss": 4.7132, "step": 629750 }, { "epoch": 3.7975904776835785, "grad_norm": 2.01519513130188, "learning_rate": 0.00013057989220808313, "loss": 4.7308, "step": 629800 }, { "epoch": 3.7978919694649123, "grad_norm": 2.2397356033325195, "learning_rate": 0.00013055971468486674, "loss": 4.9719, "step": 629850 }, { "epoch": 3.7981934612462465, "grad_norm": 1.815693736076355, "learning_rate": 0.00013053953751942108, "loss": 5.0039, "step": 629900 }, { "epoch": 3.7984949530275802, "grad_norm": 3.286348581314087, "learning_rate": 0.0001305193607121174, "loss": 4.8798, "step": 629950 }, { "epoch": 3.7987964448089144, "grad_norm": 1.5436028242111206, "learning_rate": 0.0001304991842633271, "loss": 4.8618, "step": 630000 }, { "epoch": 3.7987964448089144, "eval_loss": 5.250492095947266, "eval_runtime": 39.0433, "eval_samples_per_second": 13.114, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.578568697142557, "step": 630000 }, { "epoch": 3.7990979365902486, "grad_norm": 1.672052025794983, "learning_rate": 0.00013047900817342154, "loss": 4.9293, "step": 630050 }, { "epoch": 3.799399428371583, "grad_norm": 2.2855145931243896, "learning_rate": 0.00013045883244277193, "loss": 4.936, "step": 630100 }, { "epoch": 3.7997009201529166, "grad_norm": 2.526249647140503, "learning_rate": 0.0001304386570717496, "loss": 4.7548, "step": 630150 }, { "epoch": 3.800002411934251, "grad_norm": 2.609872579574585, "learning_rate": 0.00013041848206072593, "loss": 5.2844, "step": 630200 }, { "epoch": 3.8003039037155846, "grad_norm": 2.505378484725952, "learning_rate": 0.00013039830741007214, "loss": 5.0084, "step": 630250 }, { "epoch": 3.800605395496919, "grad_norm": 2.0385239124298096, "learning_rate": 0.00013037813312015949, "loss": 4.7172, "step": 630300 }, { "epoch": 3.800906887278253, "grad_norm": 2.399343729019165, "learning_rate": 0.00013035795919135937, "loss": 4.6459, "step": 630350 }, { "epoch": 3.8012083790595868, "grad_norm": 1.523425817489624, "learning_rate": 0.00013033778562404288, "loss": 4.7229, "step": 630400 }, { "epoch": 3.801509870840921, "grad_norm": 2.6340627670288086, "learning_rate": 0.00013031761241858144, "loss": 4.9655, "step": 630450 }, { "epoch": 3.8018113626222547, "grad_norm": 2.2199835777282715, "learning_rate": 0.00013029743957534629, "loss": 4.7604, "step": 630500 }, { "epoch": 3.802112854403589, "grad_norm": 2.589951753616333, "learning_rate": 0.0001302772670947086, "loss": 4.9623, "step": 630550 }, { "epoch": 3.802414346184923, "grad_norm": 2.100292921066284, "learning_rate": 0.00013025709497703966, "loss": 4.9548, "step": 630600 }, { "epoch": 3.802715837966257, "grad_norm": 1.4751245975494385, "learning_rate": 0.00013023692322271073, "loss": 5.0474, "step": 630650 }, { "epoch": 3.803017329747591, "grad_norm": 2.1460554599761963, "learning_rate": 0.00013021675183209303, "loss": 4.6839, "step": 630700 }, { "epoch": 3.803318821528925, "grad_norm": 2.421703815460205, "learning_rate": 0.00013019658080555774, "loss": 4.8676, "step": 630750 }, { "epoch": 3.803620313310259, "grad_norm": 2.291257858276367, "learning_rate": 0.0001301764101434762, "loss": 5.1162, "step": 630800 }, { "epoch": 3.8039218050915933, "grad_norm": 2.297767400741577, "learning_rate": 0.0001301562398462195, "loss": 4.6491, "step": 630850 }, { "epoch": 3.8042232968729275, "grad_norm": 2.290817975997925, "learning_rate": 0.00013013606991415884, "loss": 5.0117, "step": 630900 }, { "epoch": 3.8045247886542612, "grad_norm": 2.515425682067871, "learning_rate": 0.00013011590034766555, "loss": 5.0726, "step": 630950 }, { "epoch": 3.8048262804355955, "grad_norm": 3.391002655029297, "learning_rate": 0.00013009573114711065, "loss": 4.62, "step": 631000 }, { "epoch": 3.805127772216929, "grad_norm": 1.7795202732086182, "learning_rate": 0.00013007556231286546, "loss": 5.0031, "step": 631050 }, { "epoch": 3.8054292639982634, "grad_norm": 3.4551844596862793, "learning_rate": 0.00013005539384530117, "loss": 4.6878, "step": 631100 }, { "epoch": 3.8057307557795976, "grad_norm": 2.3148622512817383, "learning_rate": 0.00013003522574478877, "loss": 4.9039, "step": 631150 }, { "epoch": 3.8060322475609314, "grad_norm": 2.7637553215026855, "learning_rate": 0.0001300150580116996, "loss": 5.0304, "step": 631200 }, { "epoch": 3.8063337393422656, "grad_norm": 0.7976521253585815, "learning_rate": 0.00012999489064640477, "loss": 4.9332, "step": 631250 }, { "epoch": 3.8066352311235994, "grad_norm": 2.0323638916015625, "learning_rate": 0.0001299747236492755, "loss": 4.6024, "step": 631300 }, { "epoch": 3.8069367229049336, "grad_norm": 2.1832566261291504, "learning_rate": 0.00012995455702068275, "loss": 4.9016, "step": 631350 }, { "epoch": 3.8072382146862678, "grad_norm": 3.5971906185150146, "learning_rate": 0.00012993439076099784, "loss": 4.627, "step": 631400 }, { "epoch": 3.8075397064676015, "grad_norm": 3.831465482711792, "learning_rate": 0.0001299142248705918, "loss": 4.6947, "step": 631450 }, { "epoch": 3.8078411982489357, "grad_norm": 2.5282034873962402, "learning_rate": 0.0001298940593498358, "loss": 4.7568, "step": 631500 }, { "epoch": 3.8081426900302695, "grad_norm": 3.155250072479248, "learning_rate": 0.00012987389419910094, "loss": 5.0491, "step": 631550 }, { "epoch": 3.8084441818116037, "grad_norm": 3.6540379524230957, "learning_rate": 0.00012985372941875836, "loss": 4.7296, "step": 631600 }, { "epoch": 3.808745673592938, "grad_norm": 2.1506168842315674, "learning_rate": 0.00012983356500917913, "loss": 4.774, "step": 631650 }, { "epoch": 3.809047165374272, "grad_norm": 0.5621451735496521, "learning_rate": 0.00012981340097073434, "loss": 4.9981, "step": 631700 }, { "epoch": 3.809348657155606, "grad_norm": 2.012061595916748, "learning_rate": 0.00012979323730379514, "loss": 4.5538, "step": 631750 }, { "epoch": 3.80965014893694, "grad_norm": 2.9038591384887695, "learning_rate": 0.0001297730740087325, "loss": 4.7548, "step": 631800 }, { "epoch": 3.809951640718274, "grad_norm": 2.135606050491333, "learning_rate": 0.0001297529110859176, "loss": 4.7082, "step": 631850 }, { "epoch": 3.810253132499608, "grad_norm": 1.950793743133545, "learning_rate": 0.00012973274853572155, "loss": 4.8563, "step": 631900 }, { "epoch": 3.8105546242809423, "grad_norm": 3.6204144954681396, "learning_rate": 0.00012971258635851524, "loss": 4.801, "step": 631950 }, { "epoch": 3.810856116062276, "grad_norm": 2.4494247436523438, "learning_rate": 0.00012969242455466987, "loss": 4.989, "step": 632000 }, { "epoch": 3.8111576078436102, "grad_norm": 2.158677577972412, "learning_rate": 0.00012967226312455643, "loss": 5.0584, "step": 632050 }, { "epoch": 3.811459099624944, "grad_norm": 2.2840328216552734, "learning_rate": 0.00012965210206854598, "loss": 4.9195, "step": 632100 }, { "epoch": 3.811760591406278, "grad_norm": 3.205134391784668, "learning_rate": 0.00012963194138700953, "loss": 5.0266, "step": 632150 }, { "epoch": 3.8120620831876124, "grad_norm": 2.0926520824432373, "learning_rate": 0.0001296117810803182, "loss": 4.8366, "step": 632200 }, { "epoch": 3.8123635749689466, "grad_norm": 2.0869996547698975, "learning_rate": 0.00012959162114884292, "loss": 4.7815, "step": 632250 }, { "epoch": 3.8126650667502804, "grad_norm": 4.459035873413086, "learning_rate": 0.00012957146159295465, "loss": 4.7576, "step": 632300 }, { "epoch": 3.8129665585316146, "grad_norm": 1.7861626148223877, "learning_rate": 0.0001295513024130246, "loss": 4.7913, "step": 632350 }, { "epoch": 3.8132680503129484, "grad_norm": 2.346254348754883, "learning_rate": 0.00012953114360942354, "loss": 4.7316, "step": 632400 }, { "epoch": 3.8135695420942826, "grad_norm": 1.897430419921875, "learning_rate": 0.0001295109851825226, "loss": 4.5253, "step": 632450 }, { "epoch": 3.8138710338756168, "grad_norm": 2.49164080619812, "learning_rate": 0.0001294908271326928, "loss": 4.7912, "step": 632500 }, { "epoch": 3.8141725256569505, "grad_norm": 2.5871517658233643, "learning_rate": 0.000129470669460305, "loss": 4.7729, "step": 632550 }, { "epoch": 3.8144740174382847, "grad_norm": 2.0567493438720703, "learning_rate": 0.00012945051216573024, "loss": 5.1205, "step": 632600 }, { "epoch": 3.8147755092196185, "grad_norm": 2.2146899700164795, "learning_rate": 0.00012943035524933948, "loss": 5.0342, "step": 632650 }, { "epoch": 3.8150770010009527, "grad_norm": 3.3594346046447754, "learning_rate": 0.00012941019871150368, "loss": 4.6102, "step": 632700 }, { "epoch": 3.815378492782287, "grad_norm": 1.293870210647583, "learning_rate": 0.00012939004255259378, "loss": 4.6789, "step": 632750 }, { "epoch": 3.8156799845636207, "grad_norm": 2.6213672161102295, "learning_rate": 0.00012936988677298076, "loss": 4.9751, "step": 632800 }, { "epoch": 3.815981476344955, "grad_norm": 4.028884410858154, "learning_rate": 0.00012934973137303552, "loss": 4.8917, "step": 632850 }, { "epoch": 3.8162829681262886, "grad_norm": 1.6822195053100586, "learning_rate": 0.00012932957635312898, "loss": 4.1302, "step": 632900 }, { "epoch": 3.816584459907623, "grad_norm": 2.1143834590911865, "learning_rate": 0.00012930942171363215, "loss": 5.3852, "step": 632950 }, { "epoch": 3.816885951688957, "grad_norm": 1.2489492893218994, "learning_rate": 0.00012928926745491581, "loss": 4.7189, "step": 633000 }, { "epoch": 3.816885951688957, "eval_loss": 5.24620246887207, "eval_runtime": 39.3077, "eval_samples_per_second": 13.025, "eval_steps_per_second": 6.513, "eval_tts_loss": 7.507322951741989, "step": 633000 }, { "epoch": 3.8171874434702913, "grad_norm": 0.6561437845230103, "learning_rate": 0.000129269113577351, "loss": 4.8248, "step": 633050 }, { "epoch": 3.817488935251625, "grad_norm": 2.4691596031188965, "learning_rate": 0.00012924896008130858, "loss": 4.8132, "step": 633100 }, { "epoch": 3.8177904270329592, "grad_norm": 1.8857345581054688, "learning_rate": 0.00012922880696715939, "loss": 4.9744, "step": 633150 }, { "epoch": 3.818091918814293, "grad_norm": 2.699448347091675, "learning_rate": 0.00012920865423527437, "loss": 4.7862, "step": 633200 }, { "epoch": 3.818393410595627, "grad_norm": 2.339949607849121, "learning_rate": 0.00012918850188602444, "loss": 4.591, "step": 633250 }, { "epoch": 3.8186949023769614, "grad_norm": 3.107065439224243, "learning_rate": 0.0001291683499197804, "loss": 4.9304, "step": 633300 }, { "epoch": 3.818996394158295, "grad_norm": 0.9438219666481018, "learning_rate": 0.00012914819833691313, "loss": 4.4113, "step": 633350 }, { "epoch": 3.8192978859396294, "grad_norm": 2.7694640159606934, "learning_rate": 0.0001291280471377936, "loss": 4.9127, "step": 633400 }, { "epoch": 3.819599377720963, "grad_norm": 3.247095823287964, "learning_rate": 0.00012910789632279249, "loss": 4.9213, "step": 633450 }, { "epoch": 3.8199008695022973, "grad_norm": 3.520328998565674, "learning_rate": 0.00012908774589228073, "loss": 4.6978, "step": 633500 }, { "epoch": 3.8202023612836316, "grad_norm": 1.93007493019104, "learning_rate": 0.00012906759584662924, "loss": 4.7111, "step": 633550 }, { "epoch": 3.8205038530649653, "grad_norm": 2.1213998794555664, "learning_rate": 0.0001290474461862087, "loss": 5.0586, "step": 633600 }, { "epoch": 3.8208053448462995, "grad_norm": 0.9244591593742371, "learning_rate": 0.00012902729691139004, "loss": 4.7348, "step": 633650 }, { "epoch": 3.8211068366276333, "grad_norm": 2.687955617904663, "learning_rate": 0.00012900714802254408, "loss": 5.2112, "step": 633700 }, { "epoch": 3.8214083284089675, "grad_norm": 1.9424418210983276, "learning_rate": 0.00012898699952004154, "loss": 4.8546, "step": 633750 }, { "epoch": 3.8217098201903017, "grad_norm": 1.6455832719802856, "learning_rate": 0.0001289668514042533, "loss": 4.869, "step": 633800 }, { "epoch": 3.822011311971636, "grad_norm": 1.7534284591674805, "learning_rate": 0.0001289467036755502, "loss": 4.4559, "step": 633850 }, { "epoch": 3.8223128037529697, "grad_norm": 2.7736525535583496, "learning_rate": 0.00012892655633430294, "loss": 5.2345, "step": 633900 }, { "epoch": 3.822614295534304, "grad_norm": 3.3160529136657715, "learning_rate": 0.0001289064093808823, "loss": 4.6468, "step": 633950 }, { "epoch": 3.8229157873156376, "grad_norm": 1.9955949783325195, "learning_rate": 0.0001288862628156592, "loss": 4.8478, "step": 634000 }, { "epoch": 3.823217279096972, "grad_norm": 2.9309728145599365, "learning_rate": 0.00012886611663900422, "loss": 4.7906, "step": 634050 }, { "epoch": 3.823518770878306, "grad_norm": 1.953321933746338, "learning_rate": 0.0001288459708512882, "loss": 4.8076, "step": 634100 }, { "epoch": 3.82382026265964, "grad_norm": 2.0360608100891113, "learning_rate": 0.00012882582545288194, "loss": 5.0201, "step": 634150 }, { "epoch": 3.824121754440974, "grad_norm": 1.8916923999786377, "learning_rate": 0.00012880568044415608, "loss": 5.3162, "step": 634200 }, { "epoch": 3.824423246222308, "grad_norm": 2.4881606101989746, "learning_rate": 0.00012878553582548147, "loss": 4.7953, "step": 634250 }, { "epoch": 3.824724738003642, "grad_norm": 1.22182297706604, "learning_rate": 0.00012876539159722885, "loss": 4.6211, "step": 634300 }, { "epoch": 3.825026229784976, "grad_norm": 1.8730387687683105, "learning_rate": 0.0001287452477597688, "loss": 4.795, "step": 634350 }, { "epoch": 3.82532772156631, "grad_norm": 2.8326804637908936, "learning_rate": 0.00012872510431347214, "loss": 4.9501, "step": 634400 }, { "epoch": 3.825629213347644, "grad_norm": 2.3489789962768555, "learning_rate": 0.00012870496125870964, "loss": 4.8991, "step": 634450 }, { "epoch": 3.825930705128978, "grad_norm": 2.1872808933258057, "learning_rate": 0.00012868481859585191, "loss": 4.6876, "step": 634500 }, { "epoch": 3.826232196910312, "grad_norm": 2.0890016555786133, "learning_rate": 0.00012866467632526965, "loss": 5.0338, "step": 634550 }, { "epoch": 3.8265336886916463, "grad_norm": 2.3230817317962646, "learning_rate": 0.00012864453444733358, "loss": 4.7282, "step": 634600 }, { "epoch": 3.8268351804729805, "grad_norm": 2.499063014984131, "learning_rate": 0.00012862439296241443, "loss": 4.7404, "step": 634650 }, { "epoch": 3.8271366722543143, "grad_norm": 1.828985571861267, "learning_rate": 0.00012860425187088276, "loss": 4.4467, "step": 634700 }, { "epoch": 3.8274381640356485, "grad_norm": 1.082631230354309, "learning_rate": 0.0001285841111731093, "loss": 4.7563, "step": 634750 }, { "epoch": 3.8277396558169823, "grad_norm": 1.9984934329986572, "learning_rate": 0.00012856397086946476, "loss": 4.8591, "step": 634800 }, { "epoch": 3.8280411475983165, "grad_norm": 2.5732295513153076, "learning_rate": 0.00012854383096031974, "loss": 5.0901, "step": 634850 }, { "epoch": 3.8283426393796507, "grad_norm": 2.356389284133911, "learning_rate": 0.00012852369144604483, "loss": 4.9587, "step": 634900 }, { "epoch": 3.8286441311609845, "grad_norm": 2.4406583309173584, "learning_rate": 0.00012850355232701081, "loss": 4.6388, "step": 634950 }, { "epoch": 3.8289456229423187, "grad_norm": 2.907986879348755, "learning_rate": 0.00012848341360358818, "loss": 4.5746, "step": 635000 }, { "epoch": 3.8292471147236524, "grad_norm": 3.1504690647125244, "learning_rate": 0.00012846327527614763, "loss": 5.0591, "step": 635050 }, { "epoch": 3.8295486065049866, "grad_norm": 1.47507643699646, "learning_rate": 0.0001284431373450598, "loss": 4.8771, "step": 635100 }, { "epoch": 3.829850098286321, "grad_norm": 3.0585877895355225, "learning_rate": 0.00012842299981069522, "loss": 4.8617, "step": 635150 }, { "epoch": 3.8301515900676546, "grad_norm": 0.9618071913719177, "learning_rate": 0.00012840286267342454, "loss": 4.2316, "step": 635200 }, { "epoch": 3.830453081848989, "grad_norm": 2.912992000579834, "learning_rate": 0.0001283827259336184, "loss": 5.0589, "step": 635250 }, { "epoch": 3.8307545736303226, "grad_norm": 2.276047706604004, "learning_rate": 0.00012836258959164726, "loss": 5.2778, "step": 635300 }, { "epoch": 3.8310560654116568, "grad_norm": 0.9686090350151062, "learning_rate": 0.00012834245364788178, "loss": 5.1345, "step": 635350 }, { "epoch": 3.831357557192991, "grad_norm": 1.1223020553588867, "learning_rate": 0.00012832231810269262, "loss": 4.896, "step": 635400 }, { "epoch": 3.831659048974325, "grad_norm": 1.469330072402954, "learning_rate": 0.0001283021829564502, "loss": 4.775, "step": 635450 }, { "epoch": 3.831960540755659, "grad_norm": 2.008079767227173, "learning_rate": 0.00012828204820952515, "loss": 4.8967, "step": 635500 }, { "epoch": 3.832262032536993, "grad_norm": 1.1369224786758423, "learning_rate": 0.00012826191386228805, "loss": 5.0163, "step": 635550 }, { "epoch": 3.832563524318327, "grad_norm": 3.008671998977661, "learning_rate": 0.00012824177991510932, "loss": 4.9079, "step": 635600 }, { "epoch": 3.832865016099661, "grad_norm": 1.7945634126663208, "learning_rate": 0.00012822164636835964, "loss": 4.7351, "step": 635650 }, { "epoch": 3.8331665078809953, "grad_norm": 1.1341701745986938, "learning_rate": 0.0001282015132224095, "loss": 4.8792, "step": 635700 }, { "epoch": 3.833467999662329, "grad_norm": 2.8239550590515137, "learning_rate": 0.00012818138047762933, "loss": 4.7512, "step": 635750 }, { "epoch": 3.8337694914436633, "grad_norm": 3.2436888217926025, "learning_rate": 0.00012816124813438973, "loss": 4.9744, "step": 635800 }, { "epoch": 3.834070983224997, "grad_norm": 2.534968614578247, "learning_rate": 0.00012814111619306127, "loss": 4.9603, "step": 635850 }, { "epoch": 3.8343724750063313, "grad_norm": 1.8585333824157715, "learning_rate": 0.00012812098465401426, "loss": 5.1401, "step": 635900 }, { "epoch": 3.8346739667876655, "grad_norm": 1.8799443244934082, "learning_rate": 0.00012810085351761934, "loss": 4.4885, "step": 635950 }, { "epoch": 3.8349754585689997, "grad_norm": 2.6019749641418457, "learning_rate": 0.00012808072278424704, "loss": 4.8925, "step": 636000 }, { "epoch": 3.8349754585689997, "eval_loss": 5.250950813293457, "eval_runtime": 39.0338, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.5692196324107615, "step": 636000 }, { "epoch": 3.8352769503503334, "grad_norm": 2.751906156539917, "learning_rate": 0.00012806059245426768, "loss": 5.0367, "step": 636050 }, { "epoch": 3.8355784421316677, "grad_norm": 2.118561267852783, "learning_rate": 0.0001280404625280518, "loss": 4.7171, "step": 636100 }, { "epoch": 3.8358799339130014, "grad_norm": 1.6872751712799072, "learning_rate": 0.00012802033300596992, "loss": 4.8936, "step": 636150 }, { "epoch": 3.8361814256943356, "grad_norm": 2.3931734561920166, "learning_rate": 0.00012800020388839238, "loss": 4.7009, "step": 636200 }, { "epoch": 3.83648291747567, "grad_norm": 2.3261959552764893, "learning_rate": 0.00012798007517568973, "loss": 5.1802, "step": 636250 }, { "epoch": 3.8367844092570036, "grad_norm": 1.7914118766784668, "learning_rate": 0.00012795994686823242, "loss": 5.1727, "step": 636300 }, { "epoch": 3.837085901038338, "grad_norm": 2.0680127143859863, "learning_rate": 0.00012793981896639075, "loss": 5.0078, "step": 636350 }, { "epoch": 3.8373873928196716, "grad_norm": 4.648085117340088, "learning_rate": 0.00012791969147053526, "loss": 4.8904, "step": 636400 }, { "epoch": 3.8376888846010058, "grad_norm": 2.903743267059326, "learning_rate": 0.0001278995643810364, "loss": 4.8438, "step": 636450 }, { "epoch": 3.83799037638234, "grad_norm": 2.01041579246521, "learning_rate": 0.00012787943769826442, "loss": 4.6968, "step": 636500 }, { "epoch": 3.8382918681636737, "grad_norm": 2.1421310901641846, "learning_rate": 0.00012785931142258983, "loss": 4.6398, "step": 636550 }, { "epoch": 3.838593359945008, "grad_norm": 2.2918202877044678, "learning_rate": 0.0001278391855543831, "loss": 5.0175, "step": 636600 }, { "epoch": 3.8388948517263417, "grad_norm": 2.8717334270477295, "learning_rate": 0.00012781906009401447, "loss": 4.6485, "step": 636650 }, { "epoch": 3.839196343507676, "grad_norm": 2.035233974456787, "learning_rate": 0.00012779893504185437, "loss": 5.1107, "step": 636700 }, { "epoch": 3.83949783528901, "grad_norm": 2.8241302967071533, "learning_rate": 0.00012777881039827325, "loss": 4.6031, "step": 636750 }, { "epoch": 3.8397993270703443, "grad_norm": 1.8759255409240723, "learning_rate": 0.00012775868616364137, "loss": 4.844, "step": 636800 }, { "epoch": 3.840100818851678, "grad_norm": 0.6169860363006592, "learning_rate": 0.00012773856233832913, "loss": 4.5771, "step": 636850 }, { "epoch": 3.8404023106330123, "grad_norm": 2.979933738708496, "learning_rate": 0.00012771843892270694, "loss": 4.8856, "step": 636900 }, { "epoch": 3.840703802414346, "grad_norm": 2.3151912689208984, "learning_rate": 0.000127698315917145, "loss": 4.6591, "step": 636950 }, { "epoch": 3.8410052941956803, "grad_norm": 1.7065914869308472, "learning_rate": 0.00012767819332201377, "loss": 4.9548, "step": 637000 }, { "epoch": 3.8413067859770145, "grad_norm": 2.432382345199585, "learning_rate": 0.00012765807113768353, "loss": 4.9845, "step": 637050 }, { "epoch": 3.8416082777583482, "grad_norm": 0.3963763415813446, "learning_rate": 0.00012763794936452463, "loss": 4.2879, "step": 637100 }, { "epoch": 3.8419097695396824, "grad_norm": 1.9732556343078613, "learning_rate": 0.0001276178280029073, "loss": 4.7045, "step": 637150 }, { "epoch": 3.842211261321016, "grad_norm": 2.118062734603882, "learning_rate": 0.000127597707053202, "loss": 5.008, "step": 637200 }, { "epoch": 3.8425127531023504, "grad_norm": 1.7212629318237305, "learning_rate": 0.00012757758651577884, "loss": 4.9021, "step": 637250 }, { "epoch": 3.8428142448836846, "grad_norm": 1.8580915927886963, "learning_rate": 0.0001275574663910082, "loss": 4.8666, "step": 637300 }, { "epoch": 3.8431157366650184, "grad_norm": 1.7168664932250977, "learning_rate": 0.0001275373466792605, "loss": 4.4541, "step": 637350 }, { "epoch": 3.8434172284463526, "grad_norm": 2.0705180168151855, "learning_rate": 0.00012751722738090572, "loss": 4.5379, "step": 637400 }, { "epoch": 3.8437187202276863, "grad_norm": 2.43215274810791, "learning_rate": 0.00012749710849631435, "loss": 5.1396, "step": 637450 }, { "epoch": 3.8440202120090206, "grad_norm": 3.048320770263672, "learning_rate": 0.0001274769900258566, "loss": 4.9864, "step": 637500 }, { "epoch": 3.8443217037903548, "grad_norm": 2.2693893909454346, "learning_rate": 0.0001274568719699027, "loss": 4.6799, "step": 637550 }, { "epoch": 3.844623195571689, "grad_norm": 1.9166858196258545, "learning_rate": 0.00012743675432882288, "loss": 4.7199, "step": 637600 }, { "epoch": 3.8449246873530227, "grad_norm": 2.2451562881469727, "learning_rate": 0.00012741663710298738, "loss": 4.8988, "step": 637650 }, { "epoch": 3.845226179134357, "grad_norm": 2.4121272563934326, "learning_rate": 0.0001273965202927665, "loss": 5.1207, "step": 637700 }, { "epoch": 3.8455276709156907, "grad_norm": 3.350637197494507, "learning_rate": 0.00012737640389853035, "loss": 4.998, "step": 637750 }, { "epoch": 3.845829162697025, "grad_norm": 1.5540728569030762, "learning_rate": 0.00012735628792064926, "loss": 4.8277, "step": 637800 }, { "epoch": 3.846130654478359, "grad_norm": 2.755255699157715, "learning_rate": 0.00012733617235949332, "loss": 4.7097, "step": 637850 }, { "epoch": 3.846432146259693, "grad_norm": 1.8238437175750732, "learning_rate": 0.00012731605721543282, "loss": 5.1155, "step": 637900 }, { "epoch": 3.846733638041027, "grad_norm": 0.6531262397766113, "learning_rate": 0.00012729594248883793, "loss": 4.6008, "step": 637950 }, { "epoch": 3.847035129822361, "grad_norm": 2.167314052581787, "learning_rate": 0.00012727582818007876, "loss": 5.1322, "step": 638000 }, { "epoch": 3.847336621603695, "grad_norm": 1.7300999164581299, "learning_rate": 0.00012725571428952559, "loss": 4.7181, "step": 638050 }, { "epoch": 3.8476381133850293, "grad_norm": 2.8807790279388428, "learning_rate": 0.00012723560081754847, "loss": 4.9081, "step": 638100 }, { "epoch": 3.847939605166363, "grad_norm": 2.155560255050659, "learning_rate": 0.00012721548776451773, "loss": 4.7891, "step": 638150 }, { "epoch": 3.8482410969476972, "grad_norm": 0.8551265597343445, "learning_rate": 0.00012719537513080335, "loss": 5.0005, "step": 638200 }, { "epoch": 3.848542588729031, "grad_norm": 1.4932992458343506, "learning_rate": 0.00012717526291677558, "loss": 4.8394, "step": 638250 }, { "epoch": 3.848844080510365, "grad_norm": 1.8423899412155151, "learning_rate": 0.00012715515112280456, "loss": 4.7291, "step": 638300 }, { "epoch": 3.8491455722916994, "grad_norm": 1.808298110961914, "learning_rate": 0.00012713503974926032, "loss": 4.9556, "step": 638350 }, { "epoch": 3.8494470640730336, "grad_norm": 1.0977258682250977, "learning_rate": 0.00012711492879651306, "loss": 5.0117, "step": 638400 }, { "epoch": 3.8497485558543674, "grad_norm": 2.1821346282958984, "learning_rate": 0.00012709481826493293, "loss": 4.9189, "step": 638450 }, { "epoch": 3.8500500476357016, "grad_norm": 2.0975844860076904, "learning_rate": 0.0001270747081548899, "loss": 5.2942, "step": 638500 }, { "epoch": 3.8503515394170353, "grad_norm": 3.3907740116119385, "learning_rate": 0.00012705459846675416, "loss": 5.0644, "step": 638550 }, { "epoch": 3.8506530311983695, "grad_norm": 3.0605289936065674, "learning_rate": 0.00012703448920089585, "loss": 4.7958, "step": 638600 }, { "epoch": 3.8509545229797038, "grad_norm": 3.6168034076690674, "learning_rate": 0.00012701438035768496, "loss": 4.9271, "step": 638650 }, { "epoch": 3.8512560147610375, "grad_norm": 2.07696270942688, "learning_rate": 0.00012699427193749158, "loss": 4.626, "step": 638700 }, { "epoch": 3.8515575065423717, "grad_norm": 1.5543042421340942, "learning_rate": 0.00012697416394068584, "loss": 5.2218, "step": 638750 }, { "epoch": 3.8518589983237055, "grad_norm": 2.3712897300720215, "learning_rate": 0.00012695405636763767, "loss": 5.0323, "step": 638800 }, { "epoch": 3.8521604901050397, "grad_norm": 2.0754683017730713, "learning_rate": 0.00012693394921871727, "loss": 4.7845, "step": 638850 }, { "epoch": 3.852461981886374, "grad_norm": 1.735985517501831, "learning_rate": 0.00012691384249429466, "loss": 5.2329, "step": 638900 }, { "epoch": 3.8527634736677077, "grad_norm": 1.42947256565094, "learning_rate": 0.00012689373619473974, "loss": 4.6057, "step": 638950 }, { "epoch": 3.853064965449042, "grad_norm": 2.176825761795044, "learning_rate": 0.00012687363032042268, "loss": 4.7356, "step": 639000 }, { "epoch": 3.853064965449042, "eval_loss": 5.25038480758667, "eval_runtime": 39.2864, "eval_samples_per_second": 13.033, "eval_steps_per_second": 6.516, "eval_tts_loss": 7.600818277620221, "step": 639000 }, { "epoch": 3.8533664572303756, "grad_norm": 5.0047197341918945, "learning_rate": 0.00012685352487171346, "loss": 5.0589, "step": 639050 }, { "epoch": 3.85366794901171, "grad_norm": 3.254485845565796, "learning_rate": 0.00012683341984898205, "loss": 5.0517, "step": 639100 }, { "epoch": 3.853969440793044, "grad_norm": 3.2942991256713867, "learning_rate": 0.00012681331525259845, "loss": 4.8464, "step": 639150 }, { "epoch": 3.8542709325743783, "grad_norm": 3.8250069618225098, "learning_rate": 0.00012679321108293278, "loss": 4.7421, "step": 639200 }, { "epoch": 3.854572424355712, "grad_norm": 2.47505259513855, "learning_rate": 0.0001267731073403549, "loss": 4.6238, "step": 639250 }, { "epoch": 3.854873916137046, "grad_norm": 3.243575096130371, "learning_rate": 0.00012675300402523483, "loss": 4.6968, "step": 639300 }, { "epoch": 3.85517540791838, "grad_norm": 2.194218158721924, "learning_rate": 0.0001267329011379426, "loss": 4.6133, "step": 639350 }, { "epoch": 3.855476899699714, "grad_norm": 2.2134177684783936, "learning_rate": 0.00012671279867884804, "loss": 5.0595, "step": 639400 }, { "epoch": 3.8557783914810484, "grad_norm": 3.321337938308716, "learning_rate": 0.0001266926966483212, "loss": 4.8511, "step": 639450 }, { "epoch": 3.856079883262382, "grad_norm": 2.084031105041504, "learning_rate": 0.00012667259504673209, "loss": 4.7203, "step": 639500 }, { "epoch": 3.8563813750437164, "grad_norm": 2.4856667518615723, "learning_rate": 0.0001266524938744505, "loss": 4.659, "step": 639550 }, { "epoch": 3.85668286682505, "grad_norm": 2.438516139984131, "learning_rate": 0.00012663239313184647, "loss": 4.6729, "step": 639600 }, { "epoch": 3.8569843586063843, "grad_norm": 2.790337085723877, "learning_rate": 0.00012661229281928993, "loss": 4.8855, "step": 639650 }, { "epoch": 3.8572858503877185, "grad_norm": 2.450300455093384, "learning_rate": 0.0001265921929371507, "loss": 4.9882, "step": 639700 }, { "epoch": 3.8575873421690527, "grad_norm": 3.0377535820007324, "learning_rate": 0.00012657209348579875, "loss": 5.2131, "step": 639750 }, { "epoch": 3.8578888339503865, "grad_norm": 3.048952579498291, "learning_rate": 0.00012655199446560405, "loss": 4.8464, "step": 639800 }, { "epoch": 3.8581903257317207, "grad_norm": 1.8064162731170654, "learning_rate": 0.00012653189587693642, "loss": 5.0115, "step": 639850 }, { "epoch": 3.8584918175130545, "grad_norm": 1.20871102809906, "learning_rate": 0.00012651179772016572, "loss": 4.6474, "step": 639900 }, { "epoch": 3.8587933092943887, "grad_norm": 1.9317299127578735, "learning_rate": 0.0001264916999956619, "loss": 5.052, "step": 639950 }, { "epoch": 3.859094801075723, "grad_norm": 1.8656619787216187, "learning_rate": 0.00012647160270379474, "loss": 4.9466, "step": 640000 }, { "epoch": 3.8593962928570567, "grad_norm": 2.4008688926696777, "learning_rate": 0.0001264515058449342, "loss": 5.0161, "step": 640050 }, { "epoch": 3.859697784638391, "grad_norm": 3.2552759647369385, "learning_rate": 0.0001264314094194501, "loss": 4.9879, "step": 640100 }, { "epoch": 3.8599992764197246, "grad_norm": 3.9022486209869385, "learning_rate": 0.00012641131342771222, "loss": 4.839, "step": 640150 }, { "epoch": 3.860300768201059, "grad_norm": 0.9991656541824341, "learning_rate": 0.0001263912178700905, "loss": 4.2037, "step": 640200 }, { "epoch": 3.860602259982393, "grad_norm": 1.8507070541381836, "learning_rate": 0.00012637112274695476, "loss": 4.8942, "step": 640250 }, { "epoch": 3.860903751763727, "grad_norm": 2.130153179168701, "learning_rate": 0.00012635102805867468, "loss": 4.3297, "step": 640300 }, { "epoch": 3.861205243545061, "grad_norm": 1.2519978284835815, "learning_rate": 0.00012633093380562022, "loss": 4.8504, "step": 640350 }, { "epoch": 3.8615067353263948, "grad_norm": 1.9436910152435303, "learning_rate": 0.0001263108399881612, "loss": 4.9277, "step": 640400 }, { "epoch": 3.861808227107729, "grad_norm": 2.0385499000549316, "learning_rate": 0.00012629074660666732, "loss": 4.9111, "step": 640450 }, { "epoch": 3.862109718889063, "grad_norm": 3.052764892578125, "learning_rate": 0.00012627065366150838, "loss": 4.7327, "step": 640500 }, { "epoch": 3.8624112106703974, "grad_norm": 2.4109275341033936, "learning_rate": 0.0001262505611530543, "loss": 4.7077, "step": 640550 }, { "epoch": 3.862712702451731, "grad_norm": 0.7137277126312256, "learning_rate": 0.00012623046908167464, "loss": 5.3059, "step": 640600 }, { "epoch": 3.8630141942330654, "grad_norm": 1.9651873111724854, "learning_rate": 0.00012621037744773932, "loss": 4.7573, "step": 640650 }, { "epoch": 3.863315686014399, "grad_norm": 1.9265347719192505, "learning_rate": 0.0001261902862516181, "loss": 4.7041, "step": 640700 }, { "epoch": 3.8636171777957333, "grad_norm": 2.585101842880249, "learning_rate": 0.0001261701954936806, "loss": 4.8032, "step": 640750 }, { "epoch": 3.8639186695770675, "grad_norm": 3.2770605087280273, "learning_rate": 0.00012615010517429668, "loss": 5.2564, "step": 640800 }, { "epoch": 3.8642201613584013, "grad_norm": 6.61129093170166, "learning_rate": 0.0001261300152938361, "loss": 4.8533, "step": 640850 }, { "epoch": 3.8645216531397355, "grad_norm": 2.189499855041504, "learning_rate": 0.0001261099258526684, "loss": 5.0872, "step": 640900 }, { "epoch": 3.8648231449210693, "grad_norm": 2.3100709915161133, "learning_rate": 0.00012608983685116347, "loss": 4.5454, "step": 640950 }, { "epoch": 3.8651246367024035, "grad_norm": 2.424757242202759, "learning_rate": 0.00012606974828969104, "loss": 5.0303, "step": 641000 }, { "epoch": 3.8654261284837377, "grad_norm": 2.1690328121185303, "learning_rate": 0.00012604966016862068, "loss": 5.124, "step": 641050 }, { "epoch": 3.8657276202650714, "grad_norm": 1.9213650226593018, "learning_rate": 0.00012602957248832212, "loss": 4.902, "step": 641100 }, { "epoch": 3.8660291120464056, "grad_norm": 0.7937310934066772, "learning_rate": 0.00012600948524916516, "loss": 4.6305, "step": 641150 }, { "epoch": 3.8663306038277394, "grad_norm": 0.6769958734512329, "learning_rate": 0.0001259893984515193, "loss": 4.7519, "step": 641200 }, { "epoch": 3.8666320956090736, "grad_norm": 1.9970099925994873, "learning_rate": 0.00012596931209575432, "loss": 4.9692, "step": 641250 }, { "epoch": 3.866933587390408, "grad_norm": 3.0925137996673584, "learning_rate": 0.0001259492261822399, "loss": 4.5806, "step": 641300 }, { "epoch": 3.867235079171742, "grad_norm": 1.9503955841064453, "learning_rate": 0.0001259291407113456, "loss": 4.8341, "step": 641350 }, { "epoch": 3.867536570953076, "grad_norm": 1.7302323579788208, "learning_rate": 0.00012590905568344112, "loss": 4.5061, "step": 641400 }, { "epoch": 3.86783806273441, "grad_norm": 1.948608636856079, "learning_rate": 0.0001258889710988961, "loss": 5.1363, "step": 641450 }, { "epoch": 3.8681395545157438, "grad_norm": 1.1030155420303345, "learning_rate": 0.00012586888695808018, "loss": 4.6805, "step": 641500 }, { "epoch": 3.868441046297078, "grad_norm": 2.152648448944092, "learning_rate": 0.0001258488032613629, "loss": 5.0003, "step": 641550 }, { "epoch": 3.868742538078412, "grad_norm": 3.5930233001708984, "learning_rate": 0.000125828720009114, "loss": 4.7202, "step": 641600 }, { "epoch": 3.869044029859746, "grad_norm": 2.3534393310546875, "learning_rate": 0.000125808637201703, "loss": 4.8653, "step": 641650 }, { "epoch": 3.86934552164108, "grad_norm": 2.4189441204071045, "learning_rate": 0.00012578855483949948, "loss": 5.0841, "step": 641700 }, { "epoch": 3.869647013422414, "grad_norm": 2.5171186923980713, "learning_rate": 0.00012576847292287302, "loss": 4.5224, "step": 641750 }, { "epoch": 3.869948505203748, "grad_norm": 1.206006646156311, "learning_rate": 0.00012574839145219332, "loss": 4.8829, "step": 641800 }, { "epoch": 3.8702499969850823, "grad_norm": 3.095435619354248, "learning_rate": 0.00012572831042782984, "loss": 4.6935, "step": 641850 }, { "epoch": 3.870551488766416, "grad_norm": 2.7400448322296143, "learning_rate": 0.00012570822985015213, "loss": 4.6276, "step": 641900 }, { "epoch": 3.8708529805477503, "grad_norm": 1.7959998846054077, "learning_rate": 0.00012568814971952987, "loss": 5.0611, "step": 641950 }, { "epoch": 3.871154472329084, "grad_norm": 3.102180004119873, "learning_rate": 0.00012566807003633245, "loss": 4.5506, "step": 642000 }, { "epoch": 3.871154472329084, "eval_loss": 5.233304023742676, "eval_runtime": 38.8926, "eval_samples_per_second": 13.164, "eval_steps_per_second": 6.582, "eval_tts_loss": 7.532123929952027, "step": 642000 }, { "epoch": 3.8714559641104183, "grad_norm": 2.032839775085449, "learning_rate": 0.00012564799080092947, "loss": 4.844, "step": 642050 }, { "epoch": 3.8717574558917525, "grad_norm": 2.275331497192383, "learning_rate": 0.00012562791201369052, "loss": 4.9651, "step": 642100 }, { "epoch": 3.8720589476730867, "grad_norm": 2.3601768016815186, "learning_rate": 0.000125607833674985, "loss": 4.6847, "step": 642150 }, { "epoch": 3.8723604394544204, "grad_norm": 2.6221089363098145, "learning_rate": 0.00012558775578518254, "loss": 4.8988, "step": 642200 }, { "epoch": 3.8726619312357546, "grad_norm": 2.2336645126342773, "learning_rate": 0.00012556767834465262, "loss": 4.6603, "step": 642250 }, { "epoch": 3.8729634230170884, "grad_norm": 1.9325793981552124, "learning_rate": 0.00012554760135376466, "loss": 4.6965, "step": 642300 }, { "epoch": 3.8732649147984226, "grad_norm": 1.6493563652038574, "learning_rate": 0.00012552752481288817, "loss": 4.446, "step": 642350 }, { "epoch": 3.873566406579757, "grad_norm": 1.9149665832519531, "learning_rate": 0.00012550744872239276, "loss": 5.0087, "step": 642400 }, { "epoch": 3.8738678983610906, "grad_norm": 2.79054594039917, "learning_rate": 0.00012548737308264776, "loss": 5.013, "step": 642450 }, { "epoch": 3.874169390142425, "grad_norm": 2.8104424476623535, "learning_rate": 0.00012546729789402264, "loss": 4.9558, "step": 642500 }, { "epoch": 3.8744708819237585, "grad_norm": 1.4393383264541626, "learning_rate": 0.00012544722315688695, "loss": 4.9166, "step": 642550 }, { "epoch": 3.8747723737050928, "grad_norm": 2.268521785736084, "learning_rate": 0.00012542714887161006, "loss": 4.5278, "step": 642600 }, { "epoch": 3.875073865486427, "grad_norm": 2.5552237033843994, "learning_rate": 0.0001254070750385614, "loss": 4.6891, "step": 642650 }, { "epoch": 3.8753753572677607, "grad_norm": 2.2692737579345703, "learning_rate": 0.00012538700165811047, "loss": 4.8656, "step": 642700 }, { "epoch": 3.875676849049095, "grad_norm": 2.5942885875701904, "learning_rate": 0.0001253669287306266, "loss": 4.7659, "step": 642750 }, { "epoch": 3.8759783408304287, "grad_norm": 0.969201922416687, "learning_rate": 0.0001253468562564793, "loss": 4.8643, "step": 642800 }, { "epoch": 3.876279832611763, "grad_norm": 2.460116386413574, "learning_rate": 0.00012532678423603794, "loss": 4.6564, "step": 642850 }, { "epoch": 3.876581324393097, "grad_norm": 1.8662664890289307, "learning_rate": 0.00012530671266967186, "loss": 4.681, "step": 642900 }, { "epoch": 3.8768828161744313, "grad_norm": 2.570566415786743, "learning_rate": 0.00012528664155775048, "loss": 4.7519, "step": 642950 }, { "epoch": 3.877184307955765, "grad_norm": 2.1612749099731445, "learning_rate": 0.00012526657090064324, "loss": 4.5323, "step": 643000 }, { "epoch": 3.8774857997370993, "grad_norm": 2.618812084197998, "learning_rate": 0.00012524650069871946, "loss": 5.1892, "step": 643050 }, { "epoch": 3.877787291518433, "grad_norm": 3.2082810401916504, "learning_rate": 0.00012522643095234847, "loss": 4.675, "step": 643100 }, { "epoch": 3.8780887832997673, "grad_norm": 1.334609866142273, "learning_rate": 0.00012520636166189975, "loss": 4.4719, "step": 643150 }, { "epoch": 3.8783902750811015, "grad_norm": 3.24055552482605, "learning_rate": 0.00012518629282774247, "loss": 4.9133, "step": 643200 }, { "epoch": 3.878691766862435, "grad_norm": 2.3176021575927734, "learning_rate": 0.0001251662244502461, "loss": 4.8737, "step": 643250 }, { "epoch": 3.8789932586437694, "grad_norm": 3.146193504333496, "learning_rate": 0.00012514615652977997, "loss": 4.7649, "step": 643300 }, { "epoch": 3.879294750425103, "grad_norm": 1.6216212511062622, "learning_rate": 0.00012512608906671327, "loss": 4.1798, "step": 643350 }, { "epoch": 3.8795962422064374, "grad_norm": 2.949078321456909, "learning_rate": 0.00012510602206141546, "loss": 4.6977, "step": 643400 }, { "epoch": 3.8798977339877716, "grad_norm": 1.8773982524871826, "learning_rate": 0.0001250859555142558, "loss": 5.3222, "step": 643450 }, { "epoch": 3.880199225769106, "grad_norm": 2.040201187133789, "learning_rate": 0.00012506588942560353, "loss": 4.6328, "step": 643500 }, { "epoch": 3.8805007175504396, "grad_norm": 2.105705976486206, "learning_rate": 0.000125045823795828, "loss": 4.6846, "step": 643550 }, { "epoch": 3.8808022093317738, "grad_norm": 2.7009694576263428, "learning_rate": 0.0001250257586252985, "loss": 4.4604, "step": 643600 }, { "epoch": 3.8811037011131075, "grad_norm": 1.568795084953308, "learning_rate": 0.00012500569391438426, "loss": 4.784, "step": 643650 }, { "epoch": 3.8814051928944417, "grad_norm": 2.8137905597686768, "learning_rate": 0.0001249856296634545, "loss": 4.9842, "step": 643700 }, { "epoch": 3.881706684675776, "grad_norm": 2.261312246322632, "learning_rate": 0.0001249655658728786, "loss": 4.6394, "step": 643750 }, { "epoch": 3.8820081764571097, "grad_norm": 1.728981852531433, "learning_rate": 0.00012494550254302567, "loss": 5.0326, "step": 643800 }, { "epoch": 3.882309668238444, "grad_norm": 1.9883421659469604, "learning_rate": 0.00012492543967426505, "loss": 5.016, "step": 643850 }, { "epoch": 3.8826111600197777, "grad_norm": 2.212949514389038, "learning_rate": 0.00012490537726696598, "loss": 4.7541, "step": 643900 }, { "epoch": 3.882912651801112, "grad_norm": 3.319941759109497, "learning_rate": 0.00012488531532149754, "loss": 4.5001, "step": 643950 }, { "epoch": 3.883214143582446, "grad_norm": 3.6020305156707764, "learning_rate": 0.00012486525383822907, "loss": 5.1117, "step": 644000 }, { "epoch": 3.88351563536378, "grad_norm": 4.930557727813721, "learning_rate": 0.00012484519281752977, "loss": 4.1277, "step": 644050 }, { "epoch": 3.883817127145114, "grad_norm": 1.2933346033096313, "learning_rate": 0.00012482513225976874, "loss": 4.7761, "step": 644100 }, { "epoch": 3.884118618926448, "grad_norm": 2.484726905822754, "learning_rate": 0.0001248050721653152, "loss": 5.005, "step": 644150 }, { "epoch": 3.884420110707782, "grad_norm": 2.015361785888672, "learning_rate": 0.00012478501253453844, "loss": 4.88, "step": 644200 }, { "epoch": 3.8847216024891162, "grad_norm": 1.9608436822891235, "learning_rate": 0.0001247649533678075, "loss": 4.8812, "step": 644250 }, { "epoch": 3.8850230942704504, "grad_norm": 2.877453327178955, "learning_rate": 0.00012474489466549155, "loss": 4.418, "step": 644300 }, { "epoch": 3.885324586051784, "grad_norm": 2.6437766551971436, "learning_rate": 0.00012472483642795982, "loss": 4.6361, "step": 644350 }, { "epoch": 3.8856260778331184, "grad_norm": 2.98966908454895, "learning_rate": 0.00012470477865558137, "loss": 5.1349, "step": 644400 }, { "epoch": 3.885927569614452, "grad_norm": 3.592290163040161, "learning_rate": 0.0001246847213487254, "loss": 4.7432, "step": 644450 }, { "epoch": 3.8862290613957864, "grad_norm": 1.7408990859985352, "learning_rate": 0.00012466466450776102, "loss": 4.6487, "step": 644500 }, { "epoch": 3.8865305531771206, "grad_norm": 1.6940321922302246, "learning_rate": 0.00012464460813305727, "loss": 4.5207, "step": 644550 }, { "epoch": 3.8868320449584544, "grad_norm": 2.0466086864471436, "learning_rate": 0.00012462455222498337, "loss": 4.783, "step": 644600 }, { "epoch": 3.8871335367397886, "grad_norm": 2.233959197998047, "learning_rate": 0.0001246044967839084, "loss": 5.1566, "step": 644650 }, { "epoch": 3.8874350285211223, "grad_norm": 2.3124895095825195, "learning_rate": 0.00012458444181020136, "loss": 5.08, "step": 644700 }, { "epoch": 3.8877365203024565, "grad_norm": 3.1839473247528076, "learning_rate": 0.00012456438730423138, "loss": 4.374, "step": 644750 }, { "epoch": 3.8880380120837907, "grad_norm": 1.996436357498169, "learning_rate": 0.00012454433326636763, "loss": 4.5623, "step": 644800 }, { "epoch": 3.8883395038651245, "grad_norm": 1.7867941856384277, "learning_rate": 0.0001245242796969791, "loss": 4.6276, "step": 644850 }, { "epoch": 3.8886409956464587, "grad_norm": 2.0339245796203613, "learning_rate": 0.00012450422659643478, "loss": 4.6336, "step": 644900 }, { "epoch": 3.8889424874277925, "grad_norm": 1.7806341648101807, "learning_rate": 0.00012448417396510378, "loss": 5.2472, "step": 644950 }, { "epoch": 3.8892439792091267, "grad_norm": 3.766789436340332, "learning_rate": 0.00012446412180335524, "loss": 5.2232, "step": 645000 }, { "epoch": 3.8892439792091267, "eval_loss": 5.2232770919799805, "eval_runtime": 39.1101, "eval_samples_per_second": 13.091, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.560395046533402, "step": 645000 }, { "epoch": 3.889545470990461, "grad_norm": 2.866828680038452, "learning_rate": 0.00012444407011155803, "loss": 4.4343, "step": 645050 }, { "epoch": 3.889846962771795, "grad_norm": 3.1099796295166016, "learning_rate": 0.00012442401889008124, "loss": 4.6764, "step": 645100 }, { "epoch": 3.890148454553129, "grad_norm": 1.987142562866211, "learning_rate": 0.00012440396813929392, "loss": 4.6848, "step": 645150 }, { "epoch": 3.890449946334463, "grad_norm": 2.904245376586914, "learning_rate": 0.000124383917859565, "loss": 4.5317, "step": 645200 }, { "epoch": 3.890751438115797, "grad_norm": 1.1459583044052124, "learning_rate": 0.00012436386805126352, "loss": 4.7173, "step": 645250 }, { "epoch": 3.891052929897131, "grad_norm": 1.8023797273635864, "learning_rate": 0.00012434381871475852, "loss": 4.7245, "step": 645300 }, { "epoch": 3.8913544216784652, "grad_norm": 5.454824447631836, "learning_rate": 0.00012432376985041884, "loss": 4.4907, "step": 645350 }, { "epoch": 3.891655913459799, "grad_norm": 1.7862787246704102, "learning_rate": 0.0001243037214586136, "loss": 4.5099, "step": 645400 }, { "epoch": 3.891957405241133, "grad_norm": 1.8607256412506104, "learning_rate": 0.00012428367353971168, "loss": 4.5678, "step": 645450 }, { "epoch": 3.892258897022467, "grad_norm": 2.449169397354126, "learning_rate": 0.00012426362609408201, "loss": 4.7557, "step": 645500 }, { "epoch": 3.892560388803801, "grad_norm": 2.8215935230255127, "learning_rate": 0.00012424357912209355, "loss": 4.8434, "step": 645550 }, { "epoch": 3.8928618805851354, "grad_norm": 1.5983986854553223, "learning_rate": 0.00012422353262411535, "loss": 4.6026, "step": 645600 }, { "epoch": 3.893163372366469, "grad_norm": 3.2575602531433105, "learning_rate": 0.00012420348660051623, "loss": 4.5146, "step": 645650 }, { "epoch": 3.8934648641478034, "grad_norm": 0.42919790744781494, "learning_rate": 0.00012418344105166507, "loss": 5.0555, "step": 645700 }, { "epoch": 3.893766355929137, "grad_norm": 1.9892983436584473, "learning_rate": 0.0001241633959779309, "loss": 5.1154, "step": 645750 }, { "epoch": 3.8940678477104713, "grad_norm": 1.8408808708190918, "learning_rate": 0.00012414335137968252, "loss": 4.6306, "step": 645800 }, { "epoch": 3.8943693394918055, "grad_norm": 2.4230599403381348, "learning_rate": 0.00012412330725728883, "loss": 4.6314, "step": 645850 }, { "epoch": 3.8946708312731397, "grad_norm": 1.8537274599075317, "learning_rate": 0.00012410326361111883, "loss": 4.4591, "step": 645900 }, { "epoch": 3.8949723230544735, "grad_norm": 2.611065149307251, "learning_rate": 0.00012408322044154123, "loss": 4.7287, "step": 645950 }, { "epoch": 3.8952738148358077, "grad_norm": 3.0941224098205566, "learning_rate": 0.00012406317774892498, "loss": 4.6363, "step": 646000 }, { "epoch": 3.8955753066171415, "grad_norm": 2.919196367263794, "learning_rate": 0.00012404313553363898, "loss": 4.8175, "step": 646050 }, { "epoch": 3.8958767983984757, "grad_norm": 3.1373074054718018, "learning_rate": 0.00012402309379605197, "loss": 4.6408, "step": 646100 }, { "epoch": 3.89617829017981, "grad_norm": 0.5183100700378418, "learning_rate": 0.00012400305253653282, "loss": 4.6913, "step": 646150 }, { "epoch": 3.8964797819611436, "grad_norm": 2.981232166290283, "learning_rate": 0.0001239830117554505, "loss": 4.9016, "step": 646200 }, { "epoch": 3.896781273742478, "grad_norm": 4.033341884613037, "learning_rate": 0.00012396297145317363, "loss": 4.8021, "step": 646250 }, { "epoch": 3.8970827655238116, "grad_norm": 2.6603505611419678, "learning_rate": 0.00012394293163007112, "loss": 4.9384, "step": 646300 }, { "epoch": 3.897384257305146, "grad_norm": 2.0814497470855713, "learning_rate": 0.00012392289228651185, "loss": 4.4883, "step": 646350 }, { "epoch": 3.89768574908648, "grad_norm": 2.422419309616089, "learning_rate": 0.00012390285342286445, "loss": 4.285, "step": 646400 }, { "epoch": 3.897987240867814, "grad_norm": 3.140907049179077, "learning_rate": 0.0001238828150394978, "loss": 5.4308, "step": 646450 }, { "epoch": 3.898288732649148, "grad_norm": 2.138411283493042, "learning_rate": 0.0001238627771367807, "loss": 4.7976, "step": 646500 }, { "epoch": 3.8985902244304818, "grad_norm": 2.37943959236145, "learning_rate": 0.00012384273971508187, "loss": 4.6199, "step": 646550 }, { "epoch": 3.898891716211816, "grad_norm": 2.101653575897217, "learning_rate": 0.00012382270277477007, "loss": 4.559, "step": 646600 }, { "epoch": 3.89919320799315, "grad_norm": 1.6780197620391846, "learning_rate": 0.00012380266631621412, "loss": 5.0724, "step": 646650 }, { "epoch": 3.8994946997744844, "grad_norm": 2.243587017059326, "learning_rate": 0.00012378263033978264, "loss": 4.9325, "step": 646700 }, { "epoch": 3.899796191555818, "grad_norm": 2.36246395111084, "learning_rate": 0.00012376259484584446, "loss": 4.9688, "step": 646750 }, { "epoch": 3.9000976833371523, "grad_norm": 1.1483830213546753, "learning_rate": 0.00012374255983476831, "loss": 4.9057, "step": 646800 }, { "epoch": 3.900399175118486, "grad_norm": 1.3286726474761963, "learning_rate": 0.00012372252530692287, "loss": 5.0174, "step": 646850 }, { "epoch": 3.9007006668998203, "grad_norm": 1.7030445337295532, "learning_rate": 0.00012370249126267678, "loss": 4.8438, "step": 646900 }, { "epoch": 3.9010021586811545, "grad_norm": 1.77719247341156, "learning_rate": 0.00012368245770239892, "loss": 4.5707, "step": 646950 }, { "epoch": 3.9013036504624883, "grad_norm": 1.980286717414856, "learning_rate": 0.0001236624246264578, "loss": 4.9956, "step": 647000 }, { "epoch": 3.9016051422438225, "grad_norm": 0.7382556796073914, "learning_rate": 0.00012364239203522216, "loss": 4.7368, "step": 647050 }, { "epoch": 3.9019066340251563, "grad_norm": 1.7388767004013062, "learning_rate": 0.00012362235992906072, "loss": 4.6369, "step": 647100 }, { "epoch": 3.9022081258064905, "grad_norm": 3.2747929096221924, "learning_rate": 0.00012360232830834207, "loss": 4.9664, "step": 647150 }, { "epoch": 3.9025096175878247, "grad_norm": 2.660562038421631, "learning_rate": 0.00012358229717343493, "loss": 4.6136, "step": 647200 }, { "epoch": 3.902811109369159, "grad_norm": 2.8303463459014893, "learning_rate": 0.00012356226652470792, "loss": 4.5429, "step": 647250 }, { "epoch": 3.9031126011504926, "grad_norm": 11.449273109436035, "learning_rate": 0.0001235422363625296, "loss": 4.8613, "step": 647300 }, { "epoch": 3.903414092931827, "grad_norm": 2.1323060989379883, "learning_rate": 0.0001235222066872687, "loss": 4.4592, "step": 647350 }, { "epoch": 3.9037155847131606, "grad_norm": 2.0338950157165527, "learning_rate": 0.00012350217749929383, "loss": 4.5375, "step": 647400 }, { "epoch": 3.904017076494495, "grad_norm": 1.3197314739227295, "learning_rate": 0.00012348214879897353, "loss": 4.6527, "step": 647450 }, { "epoch": 3.904318568275829, "grad_norm": 2.1328208446502686, "learning_rate": 0.0001234621205866764, "loss": 4.9604, "step": 647500 }, { "epoch": 3.904620060057163, "grad_norm": 2.791888952255249, "learning_rate": 0.00012344209286277116, "loss": 4.9757, "step": 647550 }, { "epoch": 3.904921551838497, "grad_norm": 2.227489471435547, "learning_rate": 0.00012342206562762626, "loss": 4.62, "step": 647600 }, { "epoch": 3.9052230436198307, "grad_norm": 1.4221121072769165, "learning_rate": 0.00012340203888161027, "loss": 4.6279, "step": 647650 }, { "epoch": 3.905524535401165, "grad_norm": 3.302381753921509, "learning_rate": 0.0001233820126250919, "loss": 5.0448, "step": 647700 }, { "epoch": 3.905826027182499, "grad_norm": 2.649160623550415, "learning_rate": 0.0001233619868584395, "loss": 4.8315, "step": 647750 }, { "epoch": 3.906127518963833, "grad_norm": 3.3120203018188477, "learning_rate": 0.00012334196158202175, "loss": 4.5815, "step": 647800 }, { "epoch": 3.906429010745167, "grad_norm": 3.0637447834014893, "learning_rate": 0.0001233219367962072, "loss": 4.3626, "step": 647850 }, { "epoch": 3.906730502526501, "grad_norm": 1.980254054069519, "learning_rate": 0.00012330191250136427, "loss": 4.3812, "step": 647900 }, { "epoch": 3.907031994307835, "grad_norm": 2.4033470153808594, "learning_rate": 0.00012328188869786154, "loss": 4.7585, "step": 647950 }, { "epoch": 3.9073334860891693, "grad_norm": 1.97444486618042, "learning_rate": 0.00012326186538606757, "loss": 4.9597, "step": 648000 }, { "epoch": 3.9073334860891693, "eval_loss": 5.233810901641846, "eval_runtime": 39.0259, "eval_samples_per_second": 13.119, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.509046443940481, "step": 648000 }, { "epoch": 3.9076349778705035, "grad_norm": 2.7944188117980957, "learning_rate": 0.0001232418425663508, "loss": 4.5552, "step": 648050 }, { "epoch": 3.9079364696518373, "grad_norm": 2.2301788330078125, "learning_rate": 0.00012322182023907967, "loss": 4.769, "step": 648100 }, { "epoch": 3.9082379614331715, "grad_norm": 2.759199857711792, "learning_rate": 0.00012320179840462277, "loss": 4.7137, "step": 648150 }, { "epoch": 3.9085394532145052, "grad_norm": 4.012668132781982, "learning_rate": 0.00012318177706334855, "loss": 4.6781, "step": 648200 }, { "epoch": 3.9088409449958395, "grad_norm": 2.501258373260498, "learning_rate": 0.0001231617562156254, "loss": 4.8867, "step": 648250 }, { "epoch": 3.9091424367771737, "grad_norm": 4.174848556518555, "learning_rate": 0.00012314173586182184, "loss": 4.9543, "step": 648300 }, { "epoch": 3.9094439285585074, "grad_norm": 2.352637529373169, "learning_rate": 0.00012312171600230634, "loss": 4.937, "step": 648350 }, { "epoch": 3.9097454203398416, "grad_norm": 2.1202640533447266, "learning_rate": 0.0001231016966374473, "loss": 4.6595, "step": 648400 }, { "epoch": 3.9100469121211754, "grad_norm": 2.9700253009796143, "learning_rate": 0.00012308167776761307, "loss": 4.7369, "step": 648450 }, { "epoch": 3.9103484039025096, "grad_norm": 2.127164363861084, "learning_rate": 0.00012306165939317225, "loss": 4.5727, "step": 648500 }, { "epoch": 3.910649895683844, "grad_norm": 1.7668246030807495, "learning_rate": 0.00012304164151449306, "loss": 5.0942, "step": 648550 }, { "epoch": 3.9109513874651776, "grad_norm": 1.3587913513183594, "learning_rate": 0.00012302162413194403, "loss": 4.7345, "step": 648600 }, { "epoch": 3.9112528792465118, "grad_norm": 1.2027093172073364, "learning_rate": 0.0001230016072458935, "loss": 4.5941, "step": 648650 }, { "epoch": 3.9115543710278455, "grad_norm": 2.667231798171997, "learning_rate": 0.00012298159085670987, "loss": 4.733, "step": 648700 }, { "epoch": 3.9118558628091797, "grad_norm": 2.648810863494873, "learning_rate": 0.00012296157496476147, "loss": 5.1459, "step": 648750 }, { "epoch": 3.912157354590514, "grad_norm": 2.183361053466797, "learning_rate": 0.00012294155957041672, "loss": 4.491, "step": 648800 }, { "epoch": 3.912458846371848, "grad_norm": 2.8745436668395996, "learning_rate": 0.00012292154467404397, "loss": 4.5406, "step": 648850 }, { "epoch": 3.912760338153182, "grad_norm": 2.233678102493286, "learning_rate": 0.0001229015302760115, "loss": 4.8479, "step": 648900 }, { "epoch": 3.913061829934516, "grad_norm": 2.5449023246765137, "learning_rate": 0.00012288151637668775, "loss": 4.9422, "step": 648950 }, { "epoch": 3.91336332171585, "grad_norm": 1.591692566871643, "learning_rate": 0.00012286150297644098, "loss": 4.5772, "step": 649000 }, { "epoch": 3.913664813497184, "grad_norm": 4.287527561187744, "learning_rate": 0.00012284149007563946, "loss": 4.6158, "step": 649050 }, { "epoch": 3.9139663052785183, "grad_norm": 2.6464791297912598, "learning_rate": 0.00012282147767465163, "loss": 4.6592, "step": 649100 }, { "epoch": 3.914267797059852, "grad_norm": 3.1492719650268555, "learning_rate": 0.00012280146577384565, "loss": 4.5597, "step": 649150 }, { "epoch": 3.9145692888411863, "grad_norm": 2.096107006072998, "learning_rate": 0.00012278145437358994, "loss": 4.5961, "step": 649200 }, { "epoch": 3.91487078062252, "grad_norm": 2.347492218017578, "learning_rate": 0.00012276144347425272, "loss": 4.9405, "step": 649250 }, { "epoch": 3.9151722724038542, "grad_norm": 2.752650737762451, "learning_rate": 0.0001227414330762022, "loss": 5.1087, "step": 649300 }, { "epoch": 3.9154737641851884, "grad_norm": 1.2258375883102417, "learning_rate": 0.00012272142317980674, "loss": 4.6847, "step": 649350 }, { "epoch": 3.915775255966522, "grad_norm": 3.335845470428467, "learning_rate": 0.00012270141378543456, "loss": 4.4776, "step": 649400 }, { "epoch": 3.9160767477478564, "grad_norm": 4.0950493812561035, "learning_rate": 0.00012268140489345392, "loss": 4.9373, "step": 649450 }, { "epoch": 3.91637823952919, "grad_norm": 2.0512757301330566, "learning_rate": 0.00012266139650423297, "loss": 4.7174, "step": 649500 }, { "epoch": 3.9166797313105244, "grad_norm": 1.7857260704040527, "learning_rate": 0.0001226413886181401, "loss": 4.6282, "step": 649550 }, { "epoch": 3.9169812230918586, "grad_norm": 3.749481439590454, "learning_rate": 0.00012262138123554338, "loss": 4.7276, "step": 649600 }, { "epoch": 3.917282714873193, "grad_norm": 2.5953867435455322, "learning_rate": 0.00012260137435681103, "loss": 4.8744, "step": 649650 }, { "epoch": 3.9175842066545266, "grad_norm": 2.463442087173462, "learning_rate": 0.00012258136798231137, "loss": 4.9148, "step": 649700 }, { "epoch": 3.9178856984358608, "grad_norm": 2.715920925140381, "learning_rate": 0.0001225613621124124, "loss": 4.815, "step": 649750 }, { "epoch": 3.9181871902171945, "grad_norm": 2.303316831588745, "learning_rate": 0.00012254135674748248, "loss": 4.9673, "step": 649800 }, { "epoch": 3.9184886819985287, "grad_norm": 2.4119832515716553, "learning_rate": 0.0001225213518878897, "loss": 5.0027, "step": 649850 }, { "epoch": 3.918790173779863, "grad_norm": 3.0772571563720703, "learning_rate": 0.00012250134753400218, "loss": 5.0767, "step": 649900 }, { "epoch": 3.9190916655611967, "grad_norm": 1.0068573951721191, "learning_rate": 0.00012248134368618812, "loss": 4.8521, "step": 649950 }, { "epoch": 3.919393157342531, "grad_norm": 2.044389009475708, "learning_rate": 0.0001224613403448157, "loss": 4.5298, "step": 650000 }, { "epoch": 3.9196946491238647, "grad_norm": 1.9975858926773071, "learning_rate": 0.00012244133751025302, "loss": 4.7014, "step": 650050 }, { "epoch": 3.919996140905199, "grad_norm": 2.47711181640625, "learning_rate": 0.00012242133518286814, "loss": 4.8108, "step": 650100 }, { "epoch": 3.920297632686533, "grad_norm": 2.147707462310791, "learning_rate": 0.00012240133336302928, "loss": 4.8337, "step": 650150 }, { "epoch": 3.920599124467867, "grad_norm": 1.9898719787597656, "learning_rate": 0.00012238133205110446, "loss": 4.6943, "step": 650200 }, { "epoch": 3.920900616249201, "grad_norm": 3.8926258087158203, "learning_rate": 0.0001223613312474618, "loss": 4.7763, "step": 650250 }, { "epoch": 3.921202108030535, "grad_norm": 1.8026559352874756, "learning_rate": 0.00012234133095246946, "loss": 5.0283, "step": 650300 }, { "epoch": 3.921503599811869, "grad_norm": 2.2591516971588135, "learning_rate": 0.00012232133116649537, "loss": 4.3866, "step": 650350 }, { "epoch": 3.9218050915932032, "grad_norm": 1.0197503566741943, "learning_rate": 0.00012230133188990773, "loss": 4.6436, "step": 650400 }, { "epoch": 3.9221065833745374, "grad_norm": 5.206390857696533, "learning_rate": 0.00012228133312307455, "loss": 4.7692, "step": 650450 }, { "epoch": 3.922408075155871, "grad_norm": 1.9981634616851807, "learning_rate": 0.00012226133486636383, "loss": 4.5271, "step": 650500 }, { "epoch": 3.9227095669372054, "grad_norm": 3.0328028202056885, "learning_rate": 0.00012224133712014366, "loss": 5.0194, "step": 650550 }, { "epoch": 3.923011058718539, "grad_norm": 2.2488749027252197, "learning_rate": 0.00012222133988478215, "loss": 4.7702, "step": 650600 }, { "epoch": 3.9233125504998734, "grad_norm": 3.008007287979126, "learning_rate": 0.0001222013431606472, "loss": 4.8589, "step": 650650 }, { "epoch": 3.9236140422812076, "grad_norm": 2.5331521034240723, "learning_rate": 0.00012218134694810681, "loss": 4.5173, "step": 650700 }, { "epoch": 3.9239155340625413, "grad_norm": 3.0564348697662354, "learning_rate": 0.00012216135124752908, "loss": 5.0044, "step": 650750 }, { "epoch": 3.9242170258438756, "grad_norm": 2.3144423961639404, "learning_rate": 0.00012214135605928195, "loss": 4.6999, "step": 650800 }, { "epoch": 3.9245185176252093, "grad_norm": 3.3294317722320557, "learning_rate": 0.00012212136138373337, "loss": 4.8884, "step": 650850 }, { "epoch": 3.9248200094065435, "grad_norm": 2.707672357559204, "learning_rate": 0.0001221013672212514, "loss": 4.286, "step": 650900 }, { "epoch": 3.9251215011878777, "grad_norm": 2.5662405490875244, "learning_rate": 0.0001220813735722039, "loss": 5.1577, "step": 650950 }, { "epoch": 3.925422992969212, "grad_norm": 2.467970132827759, "learning_rate": 0.00012206138043695888, "loss": 4.8553, "step": 651000 }, { "epoch": 3.925422992969212, "eval_loss": 5.233432769775391, "eval_runtime": 39.0177, "eval_samples_per_second": 13.122, "eval_steps_per_second": 6.561, "eval_tts_loss": 7.518275088409602, "step": 651000 }, { "epoch": 3.9257244847505457, "grad_norm": 2.1564157009124756, "learning_rate": 0.00012204138781588435, "loss": 4.9568, "step": 651050 }, { "epoch": 3.92602597653188, "grad_norm": 3.249526023864746, "learning_rate": 0.0001220213957093481, "loss": 5.1023, "step": 651100 }, { "epoch": 3.9263274683132137, "grad_norm": 2.7595715522766113, "learning_rate": 0.00012200140411771814, "loss": 4.8159, "step": 651150 }, { "epoch": 3.926628960094548, "grad_norm": 3.0238730907440186, "learning_rate": 0.00012198141304136243, "loss": 4.7317, "step": 651200 }, { "epoch": 3.926930451875882, "grad_norm": 2.250758409500122, "learning_rate": 0.0001219614224806488, "loss": 4.9872, "step": 651250 }, { "epoch": 3.927231943657216, "grad_norm": 4.270317077636719, "learning_rate": 0.00012194143243594516, "loss": 4.6347, "step": 651300 }, { "epoch": 3.92753343543855, "grad_norm": 2.2193830013275146, "learning_rate": 0.00012192144290761948, "loss": 4.7388, "step": 651350 }, { "epoch": 3.927834927219884, "grad_norm": 3.4606106281280518, "learning_rate": 0.00012190145389603951, "loss": 4.6542, "step": 651400 }, { "epoch": 3.928136419001218, "grad_norm": 2.8218541145324707, "learning_rate": 0.00012188146540157318, "loss": 4.9214, "step": 651450 }, { "epoch": 3.9284379107825522, "grad_norm": 1.8981239795684814, "learning_rate": 0.00012186147742458833, "loss": 4.7407, "step": 651500 }, { "epoch": 3.928739402563886, "grad_norm": 2.5391643047332764, "learning_rate": 0.00012184148996545291, "loss": 4.5648, "step": 651550 }, { "epoch": 3.92904089434522, "grad_norm": 1.8551557064056396, "learning_rate": 0.00012182150302453466, "loss": 4.9066, "step": 651600 }, { "epoch": 3.929342386126554, "grad_norm": 3.1348018646240234, "learning_rate": 0.00012180151660220139, "loss": 5.0759, "step": 651650 }, { "epoch": 3.929643877907888, "grad_norm": 3.690760850906372, "learning_rate": 0.00012178153069882103, "loss": 4.9304, "step": 651700 }, { "epoch": 3.9299453696892224, "grad_norm": 2.0151174068450928, "learning_rate": 0.00012176154531476126, "loss": 4.9199, "step": 651750 }, { "epoch": 3.9302468614705566, "grad_norm": 2.0074336528778076, "learning_rate": 0.00012174156045039, "loss": 4.5344, "step": 651800 }, { "epoch": 3.9305483532518903, "grad_norm": 2.616410255432129, "learning_rate": 0.00012172157610607499, "loss": 4.5423, "step": 651850 }, { "epoch": 3.9308498450332245, "grad_norm": 1.2825759649276733, "learning_rate": 0.00012170159228218398, "loss": 4.984, "step": 651900 }, { "epoch": 3.9311513368145583, "grad_norm": 1.7482155561447144, "learning_rate": 0.00012168160897908481, "loss": 5.039, "step": 651950 }, { "epoch": 3.9314528285958925, "grad_norm": 3.1367650032043457, "learning_rate": 0.00012166162619714524, "loss": 4.7968, "step": 652000 }, { "epoch": 3.9317543203772267, "grad_norm": 2.3475348949432373, "learning_rate": 0.00012164164393673294, "loss": 4.7963, "step": 652050 }, { "epoch": 3.9320558121585605, "grad_norm": 2.8558006286621094, "learning_rate": 0.0001216216621982157, "loss": 4.6798, "step": 652100 }, { "epoch": 3.9323573039398947, "grad_norm": 2.9037930965423584, "learning_rate": 0.00012160168098196135, "loss": 4.9464, "step": 652150 }, { "epoch": 3.9326587957212285, "grad_norm": 2.8104629516601562, "learning_rate": 0.00012158170028833749, "loss": 4.3239, "step": 652200 }, { "epoch": 3.9329602875025627, "grad_norm": 1.765039324760437, "learning_rate": 0.00012156172011771186, "loss": 4.664, "step": 652250 }, { "epoch": 3.933261779283897, "grad_norm": 1.9852317571640015, "learning_rate": 0.00012154174047045225, "loss": 4.4563, "step": 652300 }, { "epoch": 3.9335632710652306, "grad_norm": 2.429331064224243, "learning_rate": 0.00012152176134692622, "loss": 4.7428, "step": 652350 }, { "epoch": 3.933864762846565, "grad_norm": 2.093261241912842, "learning_rate": 0.00012150178274750158, "loss": 4.7528, "step": 652400 }, { "epoch": 3.9341662546278986, "grad_norm": 2.3149378299713135, "learning_rate": 0.00012148180467254599, "loss": 4.7755, "step": 652450 }, { "epoch": 3.934467746409233, "grad_norm": 1.7433178424835205, "learning_rate": 0.00012146182712242701, "loss": 5.1888, "step": 652500 }, { "epoch": 3.934769238190567, "grad_norm": 1.3478282690048218, "learning_rate": 0.00012144185009751241, "loss": 4.8427, "step": 652550 }, { "epoch": 3.935070729971901, "grad_norm": 1.9472548961639404, "learning_rate": 0.00012142187359816985, "loss": 4.9667, "step": 652600 }, { "epoch": 3.935372221753235, "grad_norm": 2.0693514347076416, "learning_rate": 0.00012140189762476685, "loss": 4.9797, "step": 652650 }, { "epoch": 3.935673713534569, "grad_norm": 2.8099777698516846, "learning_rate": 0.00012138192217767111, "loss": 4.4976, "step": 652700 }, { "epoch": 3.935975205315903, "grad_norm": 2.3441131114959717, "learning_rate": 0.0001213619472572503, "loss": 4.6851, "step": 652750 }, { "epoch": 3.936276697097237, "grad_norm": 2.786407470703125, "learning_rate": 0.00012134197286387198, "loss": 4.8243, "step": 652800 }, { "epoch": 3.9365781888785714, "grad_norm": 3.357213020324707, "learning_rate": 0.00012132199899790371, "loss": 5.2331, "step": 652850 }, { "epoch": 3.936879680659905, "grad_norm": 2.328517436981201, "learning_rate": 0.00012130202565971319, "loss": 4.755, "step": 652900 }, { "epoch": 3.9371811724412393, "grad_norm": 2.644315242767334, "learning_rate": 0.00012128205284966786, "loss": 4.579, "step": 652950 }, { "epoch": 3.937482664222573, "grad_norm": 1.3664495944976807, "learning_rate": 0.0001212620805681354, "loss": 4.4506, "step": 653000 }, { "epoch": 3.9377841560039073, "grad_norm": 2.0178253650665283, "learning_rate": 0.00012124210881548336, "loss": 4.8938, "step": 653050 }, { "epoch": 3.9380856477852415, "grad_norm": 1.8574753999710083, "learning_rate": 0.00012122213759207922, "loss": 4.669, "step": 653100 }, { "epoch": 3.9383871395665753, "grad_norm": 3.17280912399292, "learning_rate": 0.0001212021668982906, "loss": 4.3436, "step": 653150 }, { "epoch": 3.9386886313479095, "grad_norm": 2.820639133453369, "learning_rate": 0.00012118219673448502, "loss": 5.5525, "step": 653200 }, { "epoch": 3.9389901231292432, "grad_norm": 2.1346933841705322, "learning_rate": 0.00012116222710102993, "loss": 4.3962, "step": 653250 }, { "epoch": 3.9392916149105774, "grad_norm": 3.0590286254882812, "learning_rate": 0.0001211422579982929, "loss": 4.7468, "step": 653300 }, { "epoch": 3.9395931066919117, "grad_norm": 2.296006441116333, "learning_rate": 0.00012112228942664148, "loss": 4.317, "step": 653350 }, { "epoch": 3.939894598473246, "grad_norm": 3.4806976318359375, "learning_rate": 0.0001211023213864431, "loss": 4.7571, "step": 653400 }, { "epoch": 3.9401960902545796, "grad_norm": 2.6143743991851807, "learning_rate": 0.0001210823538780652, "loss": 5.1731, "step": 653450 }, { "epoch": 3.940497582035914, "grad_norm": 2.4270036220550537, "learning_rate": 0.00012106238690187539, "loss": 4.5999, "step": 653500 }, { "epoch": 3.9407990738172476, "grad_norm": 0.8158072233200073, "learning_rate": 0.00012104242045824099, "loss": 4.8912, "step": 653550 }, { "epoch": 3.941100565598582, "grad_norm": 2.720566749572754, "learning_rate": 0.00012102245454752953, "loss": 4.7578, "step": 653600 }, { "epoch": 3.941402057379916, "grad_norm": 2.458484649658203, "learning_rate": 0.00012100248917010849, "loss": 4.5903, "step": 653650 }, { "epoch": 3.9417035491612498, "grad_norm": 1.9275277853012085, "learning_rate": 0.00012098252432634517, "loss": 4.7874, "step": 653700 }, { "epoch": 3.942005040942584, "grad_norm": 1.420935869216919, "learning_rate": 0.00012096256001660714, "loss": 4.3445, "step": 653750 }, { "epoch": 3.9423065327239177, "grad_norm": 2.745032548904419, "learning_rate": 0.00012094259624126173, "loss": 4.4985, "step": 653800 }, { "epoch": 3.942608024505252, "grad_norm": 1.490502119064331, "learning_rate": 0.0001209226330006764, "loss": 4.9093, "step": 653850 }, { "epoch": 3.942909516286586, "grad_norm": 2.7990801334381104, "learning_rate": 0.00012090267029521846, "loss": 4.8558, "step": 653900 }, { "epoch": 3.94321100806792, "grad_norm": 2.9718234539031982, "learning_rate": 0.00012088270812525541, "loss": 4.7822, "step": 653950 }, { "epoch": 3.943512499849254, "grad_norm": 1.2798776626586914, "learning_rate": 0.00012086274649115454, "loss": 4.6217, "step": 654000 }, { "epoch": 3.943512499849254, "eval_loss": 5.244325637817383, "eval_runtime": 38.9644, "eval_samples_per_second": 13.14, "eval_steps_per_second": 6.57, "eval_tts_loss": 7.646233801966998, "step": 654000 }, { "epoch": 3.943813991630588, "grad_norm": 1.8097476959228516, "learning_rate": 0.00012084278539328321, "loss": 4.6012, "step": 654050 }, { "epoch": 3.944115483411922, "grad_norm": 2.123014211654663, "learning_rate": 0.00012082282483200886, "loss": 5.1043, "step": 654100 }, { "epoch": 3.9444169751932563, "grad_norm": 1.9926913976669312, "learning_rate": 0.00012080286480769872, "loss": 5.1596, "step": 654150 }, { "epoch": 3.9447184669745905, "grad_norm": 1.6797269582748413, "learning_rate": 0.00012078290532072023, "loss": 4.775, "step": 654200 }, { "epoch": 3.9450199587559243, "grad_norm": 3.584148645401001, "learning_rate": 0.0001207629463714407, "loss": 5.0682, "step": 654250 }, { "epoch": 3.9453214505372585, "grad_norm": 2.1125385761260986, "learning_rate": 0.00012074298796022736, "loss": 4.6961, "step": 654300 }, { "epoch": 3.9456229423185922, "grad_norm": 2.7033331394195557, "learning_rate": 0.00012072303008744759, "loss": 4.7924, "step": 654350 }, { "epoch": 3.9459244340999264, "grad_norm": 3.2568976879119873, "learning_rate": 0.00012070307275346867, "loss": 4.4357, "step": 654400 }, { "epoch": 3.9462259258812606, "grad_norm": 1.9607586860656738, "learning_rate": 0.0001206831159586579, "loss": 4.5526, "step": 654450 }, { "epoch": 3.9465274176625944, "grad_norm": 3.0413005352020264, "learning_rate": 0.0001206631597033825, "loss": 4.3321, "step": 654500 }, { "epoch": 3.9468289094439286, "grad_norm": 2.3626673221588135, "learning_rate": 0.00012064320398800985, "loss": 4.754, "step": 654550 }, { "epoch": 3.9471304012252624, "grad_norm": 4.8870768547058105, "learning_rate": 0.00012062324881290711, "loss": 4.7929, "step": 654600 }, { "epoch": 3.9474318930065966, "grad_norm": 2.5968830585479736, "learning_rate": 0.0001206032941784415, "loss": 4.6678, "step": 654650 }, { "epoch": 3.947733384787931, "grad_norm": 3.2331836223602295, "learning_rate": 0.0001205833400849804, "loss": 4.7657, "step": 654700 }, { "epoch": 3.948034876569265, "grad_norm": 3.1567747592926025, "learning_rate": 0.00012056338653289086, "loss": 5.0675, "step": 654750 }, { "epoch": 3.9483363683505988, "grad_norm": 0.41724860668182373, "learning_rate": 0.00012054343352254023, "loss": 4.8125, "step": 654800 }, { "epoch": 3.948637860131933, "grad_norm": 1.9606879949569702, "learning_rate": 0.00012052348105429563, "loss": 4.5482, "step": 654850 }, { "epoch": 3.9489393519132667, "grad_norm": 1.8215458393096924, "learning_rate": 0.00012050352912852436, "loss": 4.5736, "step": 654900 }, { "epoch": 3.949240843694601, "grad_norm": 2.528334379196167, "learning_rate": 0.00012048357774559348, "loss": 4.8061, "step": 654950 }, { "epoch": 3.949542335475935, "grad_norm": 3.24778413772583, "learning_rate": 0.00012046362690587026, "loss": 4.5621, "step": 655000 }, { "epoch": 3.949843827257269, "grad_norm": 2.139951705932617, "learning_rate": 0.00012044367660972189, "loss": 4.393, "step": 655050 }, { "epoch": 3.950145319038603, "grad_norm": 1.6550331115722656, "learning_rate": 0.00012042372685751537, "loss": 4.707, "step": 655100 }, { "epoch": 3.950446810819937, "grad_norm": 1.9160622358322144, "learning_rate": 0.000120403777649618, "loss": 4.6132, "step": 655150 }, { "epoch": 3.950748302601271, "grad_norm": 1.174199104309082, "learning_rate": 0.0001203838289863969, "loss": 5.1437, "step": 655200 }, { "epoch": 3.9510497943826053, "grad_norm": 2.0477917194366455, "learning_rate": 0.0001203638808682191, "loss": 4.6426, "step": 655250 }, { "epoch": 3.951351286163939, "grad_norm": 2.115243911743164, "learning_rate": 0.00012034393329545178, "loss": 4.7787, "step": 655300 }, { "epoch": 3.9516527779452733, "grad_norm": 2.112578868865967, "learning_rate": 0.00012032398626846212, "loss": 4.978, "step": 655350 }, { "epoch": 3.951954269726607, "grad_norm": 2.141700506210327, "learning_rate": 0.0001203040397876171, "loss": 4.831, "step": 655400 }, { "epoch": 3.9522557615079412, "grad_norm": 2.712435007095337, "learning_rate": 0.0001202840938532838, "loss": 5.2233, "step": 655450 }, { "epoch": 3.9525572532892754, "grad_norm": 2.625088930130005, "learning_rate": 0.00012026414846582942, "loss": 4.7644, "step": 655500 }, { "epoch": 3.9528587450706096, "grad_norm": 2.6696081161499023, "learning_rate": 0.00012024420362562089, "loss": 4.9715, "step": 655550 }, { "epoch": 3.9531602368519434, "grad_norm": 2.5129287242889404, "learning_rate": 0.00012022425933302537, "loss": 4.8164, "step": 655600 }, { "epoch": 3.9534617286332776, "grad_norm": 1.2775124311447144, "learning_rate": 0.00012020431558840989, "loss": 4.5474, "step": 655650 }, { "epoch": 3.9537632204146114, "grad_norm": 1.8830897808074951, "learning_rate": 0.00012018437239214139, "loss": 4.6717, "step": 655700 }, { "epoch": 3.9540647121959456, "grad_norm": 2.8568859100341797, "learning_rate": 0.00012016442974458703, "loss": 4.7915, "step": 655750 }, { "epoch": 3.95436620397728, "grad_norm": 2.2352817058563232, "learning_rate": 0.00012014448764611376, "loss": 4.8278, "step": 655800 }, { "epoch": 3.9546676957586135, "grad_norm": 1.1351046562194824, "learning_rate": 0.00012012454609708852, "loss": 4.6161, "step": 655850 }, { "epoch": 3.9549691875399478, "grad_norm": 2.1838812828063965, "learning_rate": 0.00012010460509787839, "loss": 4.8822, "step": 655900 }, { "epoch": 3.9552706793212815, "grad_norm": 1.072839379310608, "learning_rate": 0.00012008466464885043, "loss": 4.8358, "step": 655950 }, { "epoch": 3.9555721711026157, "grad_norm": 3.5879478454589844, "learning_rate": 0.00012006472475037146, "loss": 5.1357, "step": 656000 }, { "epoch": 3.95587366288395, "grad_norm": 1.339140772819519, "learning_rate": 0.00012004478540280847, "loss": 4.6829, "step": 656050 }, { "epoch": 3.9561751546652837, "grad_norm": 1.8248900175094604, "learning_rate": 0.00012002484660652855, "loss": 4.6806, "step": 656100 }, { "epoch": 3.956476646446618, "grad_norm": 2.308758020401001, "learning_rate": 0.00012000490836189847, "loss": 4.7261, "step": 656150 }, { "epoch": 3.9567781382279517, "grad_norm": 2.9066011905670166, "learning_rate": 0.00011998497066928529, "loss": 4.8213, "step": 656200 }, { "epoch": 3.957079630009286, "grad_norm": 0.6209771633148193, "learning_rate": 0.00011996503352905591, "loss": 5.1731, "step": 656250 }, { "epoch": 3.95738112179062, "grad_norm": 2.5638644695281982, "learning_rate": 0.00011994509694157717, "loss": 4.7086, "step": 656300 }, { "epoch": 3.9576826135719543, "grad_norm": 2.406057834625244, "learning_rate": 0.00011992516090721606, "loss": 4.5042, "step": 656350 }, { "epoch": 3.957984105353288, "grad_norm": 2.8637359142303467, "learning_rate": 0.00011990522542633949, "loss": 4.7916, "step": 656400 }, { "epoch": 3.9582855971346222, "grad_norm": 2.7770307064056396, "learning_rate": 0.00011988529049931421, "loss": 4.8608, "step": 656450 }, { "epoch": 3.958587088915956, "grad_norm": 2.2298531532287598, "learning_rate": 0.0001198653561265072, "loss": 4.9948, "step": 656500 }, { "epoch": 3.95888858069729, "grad_norm": 2.191202402114868, "learning_rate": 0.00011984542230828535, "loss": 4.379, "step": 656550 }, { "epoch": 3.9591900724786244, "grad_norm": 2.291048765182495, "learning_rate": 0.00011982548904501546, "loss": 5.1599, "step": 656600 }, { "epoch": 3.959491564259958, "grad_norm": 2.6463992595672607, "learning_rate": 0.00011980555633706433, "loss": 4.4701, "step": 656650 }, { "epoch": 3.9597930560412924, "grad_norm": 3.393555164337158, "learning_rate": 0.00011978562418479893, "loss": 4.6442, "step": 656700 }, { "epoch": 3.960094547822626, "grad_norm": 1.7595778703689575, "learning_rate": 0.00011976569258858593, "loss": 4.7694, "step": 656750 }, { "epoch": 3.9603960396039604, "grad_norm": 3.274083137512207, "learning_rate": 0.00011974576154879223, "loss": 5.0691, "step": 656800 }, { "epoch": 3.9606975313852946, "grad_norm": 2.0499155521392822, "learning_rate": 0.00011972583106578467, "loss": 4.4468, "step": 656850 }, { "epoch": 3.9609990231666283, "grad_norm": 2.350745916366577, "learning_rate": 0.00011970590113992991, "loss": 4.4001, "step": 656900 }, { "epoch": 3.9613005149479625, "grad_norm": 2.589953660964966, "learning_rate": 0.00011968597177159482, "loss": 5.0229, "step": 656950 }, { "epoch": 3.9616020067292963, "grad_norm": 2.7288570404052734, "learning_rate": 0.00011966604296114621, "loss": 4.5098, "step": 657000 }, { "epoch": 3.9616020067292963, "eval_loss": 5.228086948394775, "eval_runtime": 39.0399, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.569946945224455, "step": 657000 }, { "epoch": 3.9619034985106305, "grad_norm": 4.034355163574219, "learning_rate": 0.00011964611470895072, "loss": 4.5932, "step": 657050 }, { "epoch": 3.9622049902919647, "grad_norm": 1.892457127571106, "learning_rate": 0.00011962618701537517, "loss": 4.8563, "step": 657100 }, { "epoch": 3.962506482073299, "grad_norm": 1.2949960231781006, "learning_rate": 0.00011960625988078639, "loss": 5.0416, "step": 657150 }, { "epoch": 3.9628079738546327, "grad_norm": 1.3543881177902222, "learning_rate": 0.00011958633330555097, "loss": 4.9661, "step": 657200 }, { "epoch": 3.963109465635967, "grad_norm": 3.725278854370117, "learning_rate": 0.00011956640729003566, "loss": 4.6448, "step": 657250 }, { "epoch": 3.9634109574173007, "grad_norm": 2.1338632106781006, "learning_rate": 0.00011954648183460723, "loss": 4.8121, "step": 657300 }, { "epoch": 3.963712449198635, "grad_norm": 2.046807289123535, "learning_rate": 0.0001195265569396323, "loss": 4.8518, "step": 657350 }, { "epoch": 3.964013940979969, "grad_norm": 2.002542734146118, "learning_rate": 0.00011950663260547764, "loss": 4.6737, "step": 657400 }, { "epoch": 3.964315432761303, "grad_norm": 2.999936819076538, "learning_rate": 0.00011948670883250991, "loss": 5.0864, "step": 657450 }, { "epoch": 3.964616924542637, "grad_norm": 3.0737342834472656, "learning_rate": 0.00011946678562109571, "loss": 4.8029, "step": 657500 }, { "epoch": 3.964918416323971, "grad_norm": 2.8652026653289795, "learning_rate": 0.00011944686297160177, "loss": 5.0645, "step": 657550 }, { "epoch": 3.965219908105305, "grad_norm": 2.1892240047454834, "learning_rate": 0.00011942694088439475, "loss": 4.6968, "step": 657600 }, { "epoch": 3.965521399886639, "grad_norm": 3.218984603881836, "learning_rate": 0.00011940701935984117, "loss": 5.0912, "step": 657650 }, { "epoch": 3.965822891667973, "grad_norm": 1.4200007915496826, "learning_rate": 0.00011938709839830776, "loss": 4.7582, "step": 657700 }, { "epoch": 3.966124383449307, "grad_norm": 2.1162054538726807, "learning_rate": 0.00011936717800016117, "loss": 4.568, "step": 657750 }, { "epoch": 3.966425875230641, "grad_norm": 4.501220226287842, "learning_rate": 0.00011934725816576795, "loss": 4.7285, "step": 657800 }, { "epoch": 3.966727367011975, "grad_norm": 2.3122944831848145, "learning_rate": 0.00011932733889549464, "loss": 5.1446, "step": 657850 }, { "epoch": 3.9670288587933094, "grad_norm": 4.184366226196289, "learning_rate": 0.00011930742018970795, "loss": 4.8786, "step": 657900 }, { "epoch": 3.9673303505746436, "grad_norm": 0.9701599478721619, "learning_rate": 0.00011928750204877432, "loss": 4.6792, "step": 657950 }, { "epoch": 3.9676318423559773, "grad_norm": 3.1303927898406982, "learning_rate": 0.00011926758447306044, "loss": 4.9626, "step": 658000 }, { "epoch": 3.9679333341373115, "grad_norm": 1.1473703384399414, "learning_rate": 0.00011924766746293283, "loss": 4.5669, "step": 658050 }, { "epoch": 3.9682348259186453, "grad_norm": 2.5257277488708496, "learning_rate": 0.00011922775101875794, "loss": 4.9494, "step": 658100 }, { "epoch": 3.9685363176999795, "grad_norm": 2.137667179107666, "learning_rate": 0.00011920783514090244, "loss": 4.8445, "step": 658150 }, { "epoch": 3.9688378094813137, "grad_norm": 0.3884045481681824, "learning_rate": 0.0001191879198297328, "loss": 4.895, "step": 658200 }, { "epoch": 3.9691393012626475, "grad_norm": 2.0183258056640625, "learning_rate": 0.00011916800508561548, "loss": 5.0021, "step": 658250 }, { "epoch": 3.9694407930439817, "grad_norm": 4.38682222366333, "learning_rate": 0.00011914809090891701, "loss": 4.5522, "step": 658300 }, { "epoch": 3.9697422848253154, "grad_norm": 1.7567635774612427, "learning_rate": 0.00011912817730000392, "loss": 4.8786, "step": 658350 }, { "epoch": 3.9700437766066496, "grad_norm": 2.037882089614868, "learning_rate": 0.00011910826425924272, "loss": 4.5882, "step": 658400 }, { "epoch": 3.970345268387984, "grad_norm": 2.899486541748047, "learning_rate": 0.00011908835178699977, "loss": 5.0354, "step": 658450 }, { "epoch": 3.970646760169318, "grad_norm": 2.943600654602051, "learning_rate": 0.00011906843988364156, "loss": 4.739, "step": 658500 }, { "epoch": 3.970948251950652, "grad_norm": 3.4566447734832764, "learning_rate": 0.00011904852854953468, "loss": 4.5355, "step": 658550 }, { "epoch": 3.971249743731986, "grad_norm": 2.0692808628082275, "learning_rate": 0.00011902861778504541, "loss": 4.9618, "step": 658600 }, { "epoch": 3.97155123551332, "grad_norm": 2.70076322555542, "learning_rate": 0.00011900870759054018, "loss": 4.8517, "step": 658650 }, { "epoch": 3.971852727294654, "grad_norm": 1.73414146900177, "learning_rate": 0.00011898879796638556, "loss": 5.0937, "step": 658700 }, { "epoch": 3.972154219075988, "grad_norm": 2.204113006591797, "learning_rate": 0.00011896888891294781, "loss": 5.019, "step": 658750 }, { "epoch": 3.972455710857322, "grad_norm": 2.6822023391723633, "learning_rate": 0.00011894898043059335, "loss": 4.6908, "step": 658800 }, { "epoch": 3.972757202638656, "grad_norm": 2.8316965103149414, "learning_rate": 0.00011892907251968864, "loss": 4.6144, "step": 658850 }, { "epoch": 3.97305869441999, "grad_norm": 2.3023226261138916, "learning_rate": 0.00011890916518059997, "loss": 4.5118, "step": 658900 }, { "epoch": 3.973360186201324, "grad_norm": 2.54236102104187, "learning_rate": 0.00011888925841369376, "loss": 5.3734, "step": 658950 }, { "epoch": 3.9736616779826583, "grad_norm": 2.199521541595459, "learning_rate": 0.00011886935221933638, "loss": 4.8929, "step": 659000 }, { "epoch": 3.973963169763992, "grad_norm": 2.759972095489502, "learning_rate": 0.0001188494465978941, "loss": 4.8126, "step": 659050 }, { "epoch": 3.9742646615453263, "grad_norm": 1.780336618423462, "learning_rate": 0.00011882954154973329, "loss": 5.0557, "step": 659100 }, { "epoch": 3.97456615332666, "grad_norm": 2.1356987953186035, "learning_rate": 0.00011880963707522034, "loss": 5.0158, "step": 659150 }, { "epoch": 3.9748676451079943, "grad_norm": 1.6981489658355713, "learning_rate": 0.00011878973317472149, "loss": 5.091, "step": 659200 }, { "epoch": 3.9751691368893285, "grad_norm": 1.8868076801300049, "learning_rate": 0.00011876982984860303, "loss": 4.9816, "step": 659250 }, { "epoch": 3.9754706286706627, "grad_norm": 3.0803067684173584, "learning_rate": 0.00011874992709723134, "loss": 4.5233, "step": 659300 }, { "epoch": 3.9757721204519965, "grad_norm": 1.053223729133606, "learning_rate": 0.00011873002492097262, "loss": 4.5187, "step": 659350 }, { "epoch": 3.9760736122333307, "grad_norm": 2.108635425567627, "learning_rate": 0.00011871012332019315, "loss": 4.6183, "step": 659400 }, { "epoch": 3.9763751040146644, "grad_norm": 1.4012675285339355, "learning_rate": 0.00011869022229525924, "loss": 4.9474, "step": 659450 }, { "epoch": 3.9766765957959986, "grad_norm": 2.989687919616699, "learning_rate": 0.00011867032184653707, "loss": 5.0056, "step": 659500 }, { "epoch": 3.976978087577333, "grad_norm": 4.518123149871826, "learning_rate": 0.00011865042197439294, "loss": 4.6151, "step": 659550 }, { "epoch": 3.9772795793586666, "grad_norm": 2.10583758354187, "learning_rate": 0.00011863052267919308, "loss": 4.8198, "step": 659600 }, { "epoch": 3.977581071140001, "grad_norm": 2.8740458488464355, "learning_rate": 0.00011861062396130362, "loss": 4.7714, "step": 659650 }, { "epoch": 3.9778825629213346, "grad_norm": 2.313253164291382, "learning_rate": 0.00011859072582109083, "loss": 4.9937, "step": 659700 }, { "epoch": 3.978184054702669, "grad_norm": 2.1537179946899414, "learning_rate": 0.00011857082825892097, "loss": 5.0765, "step": 659750 }, { "epoch": 3.978485546484003, "grad_norm": 2.3788890838623047, "learning_rate": 0.00011855093127516015, "loss": 4.448, "step": 659800 }, { "epoch": 3.9787870382653368, "grad_norm": 2.283637046813965, "learning_rate": 0.00011853103487017452, "loss": 4.5122, "step": 659850 }, { "epoch": 3.979088530046671, "grad_norm": 0.959082305431366, "learning_rate": 0.00011851113904433033, "loss": 4.8801, "step": 659900 }, { "epoch": 3.9793900218280047, "grad_norm": 2.698615312576294, "learning_rate": 0.00011849124379799368, "loss": 4.9237, "step": 659950 }, { "epoch": 3.979691513609339, "grad_norm": 3.840517282485962, "learning_rate": 0.00011847134913153068, "loss": 4.9222, "step": 660000 }, { "epoch": 3.979691513609339, "eval_loss": 5.232255935668945, "eval_runtime": 39.0123, "eval_samples_per_second": 13.124, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.588564002097092, "step": 660000 }, { "epoch": 3.979993005390673, "grad_norm": 2.7839560508728027, "learning_rate": 0.00011845145504530758, "loss": 4.91, "step": 660050 }, { "epoch": 3.9802944971720073, "grad_norm": 1.9031544923782349, "learning_rate": 0.00011843156153969034, "loss": 4.841, "step": 660100 }, { "epoch": 3.980595988953341, "grad_norm": 2.5307581424713135, "learning_rate": 0.00011841166861504521, "loss": 4.9769, "step": 660150 }, { "epoch": 3.9808974807346753, "grad_norm": 2.684514284133911, "learning_rate": 0.00011839177627173827, "loss": 4.9694, "step": 660200 }, { "epoch": 3.981198972516009, "grad_norm": 3.3968160152435303, "learning_rate": 0.00011837188451013551, "loss": 4.6079, "step": 660250 }, { "epoch": 3.9815004642973433, "grad_norm": 2.5922505855560303, "learning_rate": 0.0001183519933306031, "loss": 4.6386, "step": 660300 }, { "epoch": 3.9818019560786775, "grad_norm": 2.0907013416290283, "learning_rate": 0.00011833210273350713, "loss": 4.9012, "step": 660350 }, { "epoch": 3.9821034478600112, "grad_norm": 1.8650211095809937, "learning_rate": 0.00011831221271921358, "loss": 4.6843, "step": 660400 }, { "epoch": 3.9824049396413455, "grad_norm": 4.103187561035156, "learning_rate": 0.00011829232328808852, "loss": 4.9971, "step": 660450 }, { "epoch": 3.982706431422679, "grad_norm": 3.2809154987335205, "learning_rate": 0.00011827243444049806, "loss": 4.7901, "step": 660500 }, { "epoch": 3.9830079232040134, "grad_norm": 1.7750732898712158, "learning_rate": 0.0001182525461768081, "loss": 5.0967, "step": 660550 }, { "epoch": 3.9833094149853476, "grad_norm": 2.5243208408355713, "learning_rate": 0.00011823265849738476, "loss": 4.9068, "step": 660600 }, { "epoch": 3.9836109067666814, "grad_norm": 2.521683692932129, "learning_rate": 0.00011821277140259402, "loss": 4.704, "step": 660650 }, { "epoch": 3.9839123985480156, "grad_norm": 1.816095232963562, "learning_rate": 0.00011819288489280181, "loss": 4.5938, "step": 660700 }, { "epoch": 3.9842138903293494, "grad_norm": 1.6411397457122803, "learning_rate": 0.0001181729989683742, "loss": 4.9064, "step": 660750 }, { "epoch": 3.9845153821106836, "grad_norm": 2.2368855476379395, "learning_rate": 0.00011815311362967714, "loss": 4.7993, "step": 660800 }, { "epoch": 3.9848168738920178, "grad_norm": 2.460092782974243, "learning_rate": 0.00011813322887707651, "loss": 5.2506, "step": 660850 }, { "epoch": 3.985118365673352, "grad_norm": 2.0176007747650146, "learning_rate": 0.00011811334471093835, "loss": 4.8856, "step": 660900 }, { "epoch": 3.9854198574546857, "grad_norm": 2.548255681991577, "learning_rate": 0.00011809346113162862, "loss": 4.6557, "step": 660950 }, { "epoch": 3.98572134923602, "grad_norm": 2.3619565963745117, "learning_rate": 0.0001180735781395132, "loss": 4.3552, "step": 661000 }, { "epoch": 3.9860228410173537, "grad_norm": 3.1150999069213867, "learning_rate": 0.00011805369573495794, "loss": 5.0052, "step": 661050 }, { "epoch": 3.986324332798688, "grad_norm": 4.076954364776611, "learning_rate": 0.0001180338139183289, "loss": 5.149, "step": 661100 }, { "epoch": 3.986625824580022, "grad_norm": 2.5232059955596924, "learning_rate": 0.00011801393268999185, "loss": 4.6255, "step": 661150 }, { "epoch": 3.986927316361356, "grad_norm": 2.70499587059021, "learning_rate": 0.00011799405205031273, "loss": 4.6373, "step": 661200 }, { "epoch": 3.98722880814269, "grad_norm": 4.968796730041504, "learning_rate": 0.00011797417199965747, "loss": 4.5694, "step": 661250 }, { "epoch": 3.987530299924024, "grad_norm": 2.7995407581329346, "learning_rate": 0.00011795429253839177, "loss": 4.7244, "step": 661300 }, { "epoch": 3.987831791705358, "grad_norm": 1.9898135662078857, "learning_rate": 0.00011793441366688163, "loss": 5.0756, "step": 661350 }, { "epoch": 3.9881332834866923, "grad_norm": 4.220852375030518, "learning_rate": 0.0001179145353854929, "loss": 4.7451, "step": 661400 }, { "epoch": 3.988434775268026, "grad_norm": 2.8020904064178467, "learning_rate": 0.00011789465769459126, "loss": 4.7987, "step": 661450 }, { "epoch": 3.9887362670493602, "grad_norm": 2.5314013957977295, "learning_rate": 0.00011787478059454265, "loss": 5.0204, "step": 661500 }, { "epoch": 3.989037758830694, "grad_norm": 2.048368453979492, "learning_rate": 0.00011785490408571292, "loss": 4.627, "step": 661550 }, { "epoch": 3.989339250612028, "grad_norm": 2.322226047515869, "learning_rate": 0.00011783502816846776, "loss": 4.9145, "step": 661600 }, { "epoch": 3.9896407423933624, "grad_norm": 1.8447566032409668, "learning_rate": 0.00011781515284317298, "loss": 5.1841, "step": 661650 }, { "epoch": 3.9899422341746966, "grad_norm": 3.275439739227295, "learning_rate": 0.00011779527811019439, "loss": 4.7939, "step": 661700 }, { "epoch": 3.9902437259560304, "grad_norm": 2.026949405670166, "learning_rate": 0.0001177754039698978, "loss": 4.6212, "step": 661750 }, { "epoch": 3.9905452177373646, "grad_norm": 2.252154588699341, "learning_rate": 0.00011775553042264889, "loss": 4.8226, "step": 661800 }, { "epoch": 3.9908467095186984, "grad_norm": 2.0288846492767334, "learning_rate": 0.00011773565746881341, "loss": 4.8478, "step": 661850 }, { "epoch": 3.9911482013000326, "grad_norm": 2.939067840576172, "learning_rate": 0.00011771578510875716, "loss": 4.9923, "step": 661900 }, { "epoch": 3.9914496930813668, "grad_norm": 2.915107488632202, "learning_rate": 0.00011769591334284578, "loss": 4.8056, "step": 661950 }, { "epoch": 3.9917511848627005, "grad_norm": 2.5324699878692627, "learning_rate": 0.00011767604217144497, "loss": 4.8882, "step": 662000 }, { "epoch": 3.9920526766440347, "grad_norm": 1.5396192073822021, "learning_rate": 0.00011765617159492054, "loss": 4.7243, "step": 662050 }, { "epoch": 3.9923541684253685, "grad_norm": 1.9221806526184082, "learning_rate": 0.00011763630161363806, "loss": 4.7329, "step": 662100 }, { "epoch": 3.9926556602067027, "grad_norm": 3.3324968814849854, "learning_rate": 0.00011761643222796331, "loss": 5.1831, "step": 662150 }, { "epoch": 3.992957151988037, "grad_norm": 3.1112184524536133, "learning_rate": 0.00011759656343826191, "loss": 5.2588, "step": 662200 }, { "epoch": 3.993258643769371, "grad_norm": 2.5330374240875244, "learning_rate": 0.0001175766952448995, "loss": 4.5757, "step": 662250 }, { "epoch": 3.993560135550705, "grad_norm": 2.6115567684173584, "learning_rate": 0.0001175568276482417, "loss": 4.8815, "step": 662300 }, { "epoch": 3.993861627332039, "grad_norm": 2.1747562885284424, "learning_rate": 0.00011753696064865426, "loss": 4.7162, "step": 662350 }, { "epoch": 3.994163119113373, "grad_norm": 2.329925060272217, "learning_rate": 0.00011751709424650271, "loss": 4.7387, "step": 662400 }, { "epoch": 3.994464610894707, "grad_norm": 1.4558910131454468, "learning_rate": 0.00011749722844215266, "loss": 4.3929, "step": 662450 }, { "epoch": 3.9947661026760413, "grad_norm": 2.715933084487915, "learning_rate": 0.00011747736323596979, "loss": 5.0141, "step": 662500 }, { "epoch": 3.995067594457375, "grad_norm": 2.592533588409424, "learning_rate": 0.00011745749862831961, "loss": 4.5757, "step": 662550 }, { "epoch": 3.9953690862387092, "grad_norm": 2.954535484313965, "learning_rate": 0.00011743763461956769, "loss": 4.7508, "step": 662600 }, { "epoch": 3.995670578020043, "grad_norm": 3.9982657432556152, "learning_rate": 0.0001174177712100797, "loss": 4.5597, "step": 662650 }, { "epoch": 3.995972069801377, "grad_norm": 3.4064676761627197, "learning_rate": 0.00011739790840022108, "loss": 4.5732, "step": 662700 }, { "epoch": 3.9962735615827114, "grad_norm": 3.1340808868408203, "learning_rate": 0.00011737804619035745, "loss": 4.8874, "step": 662750 }, { "epoch": 3.996575053364045, "grad_norm": 2.6029696464538574, "learning_rate": 0.00011735818458085435, "loss": 4.8272, "step": 662800 }, { "epoch": 3.9968765451453794, "grad_norm": 2.5532641410827637, "learning_rate": 0.00011733832357207725, "loss": 5.0108, "step": 662850 }, { "epoch": 3.997178036926713, "grad_norm": 4.2623090744018555, "learning_rate": 0.00011731846316439165, "loss": 4.3107, "step": 662900 }, { "epoch": 3.9974795287080473, "grad_norm": 2.132863759994507, "learning_rate": 0.0001172986033581632, "loss": 4.4594, "step": 662950 }, { "epoch": 3.9977810204893816, "grad_norm": 3.437586784362793, "learning_rate": 0.00011727874415375724, "loss": 4.9848, "step": 663000 }, { "epoch": 3.9977810204893816, "eval_loss": 5.231171131134033, "eval_runtime": 39.1083, "eval_samples_per_second": 13.092, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.560638255120608, "step": 663000 }, { "epoch": 3.9980825122707158, "grad_norm": 2.205249786376953, "learning_rate": 0.00011725888555153925, "loss": 4.6639, "step": 663050 }, { "epoch": 3.9983840040520495, "grad_norm": 1.9540221691131592, "learning_rate": 0.00011723902755187484, "loss": 4.7979, "step": 663100 }, { "epoch": 3.9986854958333837, "grad_norm": 2.616316318511963, "learning_rate": 0.00011721917015512932, "loss": 4.8276, "step": 663150 }, { "epoch": 3.9989869876147175, "grad_norm": 1.8416675329208374, "learning_rate": 0.00011719931336166818, "loss": 4.833, "step": 663200 }, { "epoch": 3.9992884793960517, "grad_norm": 3.0960593223571777, "learning_rate": 0.0001171794571718569, "loss": 4.7766, "step": 663250 }, { "epoch": 3.999589971177386, "grad_norm": 3.2923271656036377, "learning_rate": 0.00011715960158606084, "loss": 4.8223, "step": 663300 }, { "epoch": 3.9998914629587197, "grad_norm": 2.0194129943847656, "learning_rate": 0.00011713974660464548, "loss": 4.8488, "step": 663350 }, { "epoch": 4.000192954740053, "grad_norm": 2.269092082977295, "learning_rate": 0.00011711989222797621, "loss": 4.9531, "step": 663400 }, { "epoch": 4.000494446521388, "grad_norm": 2.4420831203460693, "learning_rate": 0.00011710003845641833, "loss": 4.6282, "step": 663450 }, { "epoch": 4.000795938302722, "grad_norm": 1.8105008602142334, "learning_rate": 0.00011708018529033731, "loss": 4.8511, "step": 663500 }, { "epoch": 4.001097430084056, "grad_norm": 1.9632065296173096, "learning_rate": 0.00011706033273009857, "loss": 4.6848, "step": 663550 }, { "epoch": 4.00139892186539, "grad_norm": 1.0649633407592773, "learning_rate": 0.00011704048077606735, "loss": 4.7005, "step": 663600 }, { "epoch": 4.001700413646724, "grad_norm": 1.8486039638519287, "learning_rate": 0.00011702062942860901, "loss": 4.8947, "step": 663650 }, { "epoch": 4.002001905428058, "grad_norm": 2.1374619007110596, "learning_rate": 0.00011700077868808898, "loss": 4.7611, "step": 663700 }, { "epoch": 4.002303397209392, "grad_norm": 2.424412727355957, "learning_rate": 0.0001169809285548725, "loss": 4.9304, "step": 663750 }, { "epoch": 4.002604888990726, "grad_norm": 2.2184081077575684, "learning_rate": 0.0001169610790293249, "loss": 4.8552, "step": 663800 }, { "epoch": 4.00290638077206, "grad_norm": 2.215493679046631, "learning_rate": 0.00011694123011181153, "loss": 4.7598, "step": 663850 }, { "epoch": 4.003207872553395, "grad_norm": 2.968792676925659, "learning_rate": 0.00011692138180269759, "loss": 4.0672, "step": 663900 }, { "epoch": 4.003509364334728, "grad_norm": 1.1745033264160156, "learning_rate": 0.00011690153410234843, "loss": 4.8687, "step": 663950 }, { "epoch": 4.003810856116062, "grad_norm": 4.775969505310059, "learning_rate": 0.00011688168701112935, "loss": 4.751, "step": 664000 }, { "epoch": 4.004112347897396, "grad_norm": 2.636723279953003, "learning_rate": 0.0001168618405294055, "loss": 4.8236, "step": 664050 }, { "epoch": 4.0044138396787305, "grad_norm": 2.988846778869629, "learning_rate": 0.00011684199465754215, "loss": 4.3921, "step": 664100 }, { "epoch": 4.004715331460065, "grad_norm": 2.5032191276550293, "learning_rate": 0.00011682214939590465, "loss": 4.6849, "step": 664150 }, { "epoch": 4.005016823241398, "grad_norm": 5.5310845375061035, "learning_rate": 0.00011680230474485813, "loss": 4.562, "step": 664200 }, { "epoch": 4.005318315022732, "grad_norm": 3.273132801055908, "learning_rate": 0.00011678246070476777, "loss": 4.8592, "step": 664250 }, { "epoch": 4.0056198068040665, "grad_norm": 2.8854057788848877, "learning_rate": 0.00011676261727599889, "loss": 4.6993, "step": 664300 }, { "epoch": 4.005921298585401, "grad_norm": 2.427478551864624, "learning_rate": 0.00011674277445891655, "loss": 4.5486, "step": 664350 }, { "epoch": 4.006222790366735, "grad_norm": 2.4770328998565674, "learning_rate": 0.00011672293225388596, "loss": 4.6651, "step": 664400 }, { "epoch": 4.006524282148068, "grad_norm": 2.389040231704712, "learning_rate": 0.0001167030906612724, "loss": 4.8642, "step": 664450 }, { "epoch": 4.006825773929402, "grad_norm": 1.2853702306747437, "learning_rate": 0.00011668324968144085, "loss": 4.6005, "step": 664500 }, { "epoch": 4.007127265710737, "grad_norm": 1.2925118207931519, "learning_rate": 0.00011666340931475659, "loss": 4.8617, "step": 664550 }, { "epoch": 4.007428757492071, "grad_norm": 0.7960173487663269, "learning_rate": 0.00011664356956158473, "loss": 5.1588, "step": 664600 }, { "epoch": 4.007730249273405, "grad_norm": 1.847598671913147, "learning_rate": 0.00011662373042229032, "loss": 4.8457, "step": 664650 }, { "epoch": 4.008031741054739, "grad_norm": 3.224219799041748, "learning_rate": 0.00011660389189723853, "loss": 4.6578, "step": 664700 }, { "epoch": 4.008333232836073, "grad_norm": 3.2430574893951416, "learning_rate": 0.00011658405398679448, "loss": 4.5603, "step": 664750 }, { "epoch": 4.008634724617407, "grad_norm": 2.4870049953460693, "learning_rate": 0.00011656421669132321, "loss": 4.6177, "step": 664800 }, { "epoch": 4.008936216398741, "grad_norm": 2.7932679653167725, "learning_rate": 0.00011654438001118978, "loss": 5.3483, "step": 664850 }, { "epoch": 4.009237708180075, "grad_norm": 1.8884243965148926, "learning_rate": 0.00011652454394675938, "loss": 5.0676, "step": 664900 }, { "epoch": 4.009539199961409, "grad_norm": 2.348681926727295, "learning_rate": 0.00011650470849839692, "loss": 4.8469, "step": 664950 }, { "epoch": 4.009840691742743, "grad_norm": 2.7322261333465576, "learning_rate": 0.00011648487366646748, "loss": 4.6685, "step": 665000 }, { "epoch": 4.010142183524077, "grad_norm": 1.314914584159851, "learning_rate": 0.00011646503945133612, "loss": 4.8975, "step": 665050 }, { "epoch": 4.010443675305411, "grad_norm": 1.2294942140579224, "learning_rate": 0.00011644520585336791, "loss": 4.7814, "step": 665100 }, { "epoch": 4.010745167086745, "grad_norm": 3.0289254188537598, "learning_rate": 0.00011642537287292776, "loss": 4.8119, "step": 665150 }, { "epoch": 4.0110466588680795, "grad_norm": 2.4123783111572266, "learning_rate": 0.00011640554051038068, "loss": 4.6223, "step": 665200 }, { "epoch": 4.011348150649413, "grad_norm": 2.2578985691070557, "learning_rate": 0.00011638570876609175, "loss": 4.5354, "step": 665250 }, { "epoch": 4.011649642430747, "grad_norm": 2.2239136695861816, "learning_rate": 0.00011636587764042582, "loss": 4.6093, "step": 665300 }, { "epoch": 4.011951134212081, "grad_norm": 3.7361700534820557, "learning_rate": 0.00011634604713374794, "loss": 4.7113, "step": 665350 }, { "epoch": 4.0122526259934155, "grad_norm": 2.675084352493286, "learning_rate": 0.00011632621724642308, "loss": 4.419, "step": 665400 }, { "epoch": 4.01255411777475, "grad_norm": 2.789656639099121, "learning_rate": 0.00011630638797881605, "loss": 4.8756, "step": 665450 }, { "epoch": 4.012855609556084, "grad_norm": 3.499748945236206, "learning_rate": 0.00011628655933129191, "loss": 4.92, "step": 665500 }, { "epoch": 4.013157101337417, "grad_norm": 2.2741897106170654, "learning_rate": 0.00011626673130421554, "loss": 4.7206, "step": 665550 }, { "epoch": 4.013458593118751, "grad_norm": 2.694011926651001, "learning_rate": 0.00011624690389795183, "loss": 4.7005, "step": 665600 }, { "epoch": 4.013760084900086, "grad_norm": 2.395792245864868, "learning_rate": 0.00011622707711286565, "loss": 4.3953, "step": 665650 }, { "epoch": 4.01406157668142, "grad_norm": 4.571759223937988, "learning_rate": 0.000116207250949322, "loss": 4.4586, "step": 665700 }, { "epoch": 4.014363068462754, "grad_norm": 3.1554160118103027, "learning_rate": 0.00011618742540768564, "loss": 4.7006, "step": 665750 }, { "epoch": 4.014664560244087, "grad_norm": 3.051319122314453, "learning_rate": 0.00011616760048832142, "loss": 4.2745, "step": 665800 }, { "epoch": 4.014966052025422, "grad_norm": 3.1408982276916504, "learning_rate": 0.00011614777619159429, "loss": 4.5649, "step": 665850 }, { "epoch": 4.015267543806756, "grad_norm": 1.5745214223861694, "learning_rate": 0.00011612795251786896, "loss": 4.6022, "step": 665900 }, { "epoch": 4.01556903558809, "grad_norm": 3.1111905574798584, "learning_rate": 0.00011610812946751037, "loss": 4.9366, "step": 665950 }, { "epoch": 4.015870527369424, "grad_norm": 2.1069109439849854, "learning_rate": 0.00011608830704088334, "loss": 4.6363, "step": 666000 }, { "epoch": 4.015870527369424, "eval_loss": 5.218989372253418, "eval_runtime": 39.0122, "eval_samples_per_second": 13.124, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.5644050331971044, "step": 666000 }, { "epoch": 4.016172019150758, "grad_norm": 0.8396453261375427, "learning_rate": 0.00011606848523835254, "loss": 4.9236, "step": 666050 }, { "epoch": 4.016473510932092, "grad_norm": 4.822430610656738, "learning_rate": 0.00011604866406028287, "loss": 4.9154, "step": 666100 }, { "epoch": 4.016775002713426, "grad_norm": 3.028163433074951, "learning_rate": 0.0001160288435070391, "loss": 4.5996, "step": 666150 }, { "epoch": 4.01707649449476, "grad_norm": 1.8548996448516846, "learning_rate": 0.00011600902357898599, "loss": 4.3753, "step": 666200 }, { "epoch": 4.017377986276094, "grad_norm": 2.3098666667938232, "learning_rate": 0.00011598920427648825, "loss": 4.9852, "step": 666250 }, { "epoch": 4.0176794780574285, "grad_norm": 2.501500129699707, "learning_rate": 0.00011596938559991074, "loss": 4.7262, "step": 666300 }, { "epoch": 4.017980969838762, "grad_norm": 2.783700704574585, "learning_rate": 0.00011594956754961809, "loss": 4.9778, "step": 666350 }, { "epoch": 4.018282461620096, "grad_norm": 3.6010468006134033, "learning_rate": 0.00011592975012597503, "loss": 5.0199, "step": 666400 }, { "epoch": 4.01858395340143, "grad_norm": 2.9825758934020996, "learning_rate": 0.00011590993332934635, "loss": 4.7474, "step": 666450 }, { "epoch": 4.0188854451827645, "grad_norm": 2.93898868560791, "learning_rate": 0.00011589011716009664, "loss": 5.0203, "step": 666500 }, { "epoch": 4.019186936964099, "grad_norm": 2.188941717147827, "learning_rate": 0.00011587030161859069, "loss": 4.8042, "step": 666550 }, { "epoch": 4.019488428745432, "grad_norm": 1.9023933410644531, "learning_rate": 0.00011585048670519315, "loss": 4.6376, "step": 666600 }, { "epoch": 4.019789920526766, "grad_norm": 2.652491569519043, "learning_rate": 0.0001158306724202686, "loss": 4.3195, "step": 666650 }, { "epoch": 4.0200914123081, "grad_norm": 2.2188563346862793, "learning_rate": 0.00011581085876418181, "loss": 4.9126, "step": 666700 }, { "epoch": 4.020392904089435, "grad_norm": 2.7284741401672363, "learning_rate": 0.00011579104573729736, "loss": 4.2532, "step": 666750 }, { "epoch": 4.020694395870769, "grad_norm": 1.3162935972213745, "learning_rate": 0.0001157712333399799, "loss": 4.3645, "step": 666800 }, { "epoch": 4.020995887652103, "grad_norm": 4.169203281402588, "learning_rate": 0.00011575142157259402, "loss": 5.0545, "step": 666850 }, { "epoch": 4.021297379433436, "grad_norm": 2.3407838344573975, "learning_rate": 0.0001157316104355044, "loss": 4.9533, "step": 666900 }, { "epoch": 4.0215988712147706, "grad_norm": 0.7036099433898926, "learning_rate": 0.00011571179992907555, "loss": 4.7384, "step": 666950 }, { "epoch": 4.021900362996105, "grad_norm": 2.1759088039398193, "learning_rate": 0.00011569199005367209, "loss": 4.9773, "step": 667000 }, { "epoch": 4.022201854777439, "grad_norm": 2.783576011657715, "learning_rate": 0.00011567218080965863, "loss": 4.7589, "step": 667050 }, { "epoch": 4.022503346558773, "grad_norm": 3.2337570190429688, "learning_rate": 0.00011565237219739963, "loss": 5.387, "step": 667100 }, { "epoch": 4.0228048383401065, "grad_norm": 3.606461763381958, "learning_rate": 0.00011563256421725974, "loss": 4.7628, "step": 667150 }, { "epoch": 4.023106330121441, "grad_norm": 2.8526618480682373, "learning_rate": 0.00011561275686960351, "loss": 4.8333, "step": 667200 }, { "epoch": 4.023407821902775, "grad_norm": 3.135093927383423, "learning_rate": 0.00011559295015479532, "loss": 5.0977, "step": 667250 }, { "epoch": 4.023709313684109, "grad_norm": 1.909953236579895, "learning_rate": 0.00011557314407319981, "loss": 4.6078, "step": 667300 }, { "epoch": 4.024010805465443, "grad_norm": 2.300576686859131, "learning_rate": 0.00011555333862518151, "loss": 4.5327, "step": 667350 }, { "epoch": 4.024312297246777, "grad_norm": 2.6485812664031982, "learning_rate": 0.0001155335338111048, "loss": 4.9078, "step": 667400 }, { "epoch": 4.024613789028111, "grad_norm": 2.2924630641937256, "learning_rate": 0.0001155137296313342, "loss": 4.7342, "step": 667450 }, { "epoch": 4.024915280809445, "grad_norm": 2.7916576862335205, "learning_rate": 0.00011549392608623426, "loss": 4.8228, "step": 667500 }, { "epoch": 4.025216772590779, "grad_norm": 1.5393364429473877, "learning_rate": 0.00011547412317616932, "loss": 5.0307, "step": 667550 }, { "epoch": 4.0255182643721135, "grad_norm": 1.9123417139053345, "learning_rate": 0.00011545432090150387, "loss": 5.0388, "step": 667600 }, { "epoch": 4.025819756153448, "grad_norm": 2.632694959640503, "learning_rate": 0.00011543451926260238, "loss": 4.4891, "step": 667650 }, { "epoch": 4.026121247934781, "grad_norm": 3.36049485206604, "learning_rate": 0.00011541471825982918, "loss": 4.8997, "step": 667700 }, { "epoch": 4.026422739716115, "grad_norm": 1.5559769868850708, "learning_rate": 0.00011539491789354877, "loss": 4.4814, "step": 667750 }, { "epoch": 4.026724231497449, "grad_norm": 2.1062521934509277, "learning_rate": 0.00011537511816412552, "loss": 4.6722, "step": 667800 }, { "epoch": 4.027025723278784, "grad_norm": 2.265652894973755, "learning_rate": 0.00011535531907192377, "loss": 4.7516, "step": 667850 }, { "epoch": 4.027327215060118, "grad_norm": 2.769129753112793, "learning_rate": 0.00011533552061730794, "loss": 4.9956, "step": 667900 }, { "epoch": 4.027628706841451, "grad_norm": 1.7599360942840576, "learning_rate": 0.00011531572280064242, "loss": 4.8655, "step": 667950 }, { "epoch": 4.027930198622785, "grad_norm": 2.437662363052368, "learning_rate": 0.00011529592562229151, "loss": 4.913, "step": 668000 }, { "epoch": 4.0282316904041195, "grad_norm": 1.885430932044983, "learning_rate": 0.00011527612908261954, "loss": 4.6655, "step": 668050 }, { "epoch": 4.028533182185454, "grad_norm": 2.385890483856201, "learning_rate": 0.0001152563331819909, "loss": 4.5439, "step": 668100 }, { "epoch": 4.028834673966788, "grad_norm": 3.540738344192505, "learning_rate": 0.00011523653792076985, "loss": 4.5181, "step": 668150 }, { "epoch": 4.029136165748121, "grad_norm": 1.0067287683486938, "learning_rate": 0.00011521674329932067, "loss": 5.2574, "step": 668200 }, { "epoch": 4.0294376575294555, "grad_norm": 1.7165762186050415, "learning_rate": 0.00011519694931800778, "loss": 4.4728, "step": 668250 }, { "epoch": 4.02973914931079, "grad_norm": 2.7494168281555176, "learning_rate": 0.00011517715597719527, "loss": 5.0447, "step": 668300 }, { "epoch": 4.030040641092124, "grad_norm": 1.143684983253479, "learning_rate": 0.00011515736327724757, "loss": 4.7035, "step": 668350 }, { "epoch": 4.030342132873458, "grad_norm": 2.158566474914551, "learning_rate": 0.00011513757121852882, "loss": 4.7742, "step": 668400 }, { "epoch": 4.030643624654792, "grad_norm": 2.226459503173828, "learning_rate": 0.0001151177798014034, "loss": 5.0532, "step": 668450 }, { "epoch": 4.030945116436126, "grad_norm": 2.134448528289795, "learning_rate": 0.0001150979890262354, "loss": 4.9293, "step": 668500 }, { "epoch": 4.03124660821746, "grad_norm": 2.7443830966949463, "learning_rate": 0.00011507819889338913, "loss": 4.6947, "step": 668550 }, { "epoch": 4.031548099998794, "grad_norm": 2.205925703048706, "learning_rate": 0.00011505840940322882, "loss": 4.7616, "step": 668600 }, { "epoch": 4.031849591780128, "grad_norm": 3.977206230163574, "learning_rate": 0.00011503862055611853, "loss": 4.6031, "step": 668650 }, { "epoch": 4.0321510835614625, "grad_norm": 2.391632556915283, "learning_rate": 0.00011501883235242258, "loss": 4.7192, "step": 668700 }, { "epoch": 4.032452575342796, "grad_norm": 4.03931188583374, "learning_rate": 0.00011499904479250514, "loss": 4.936, "step": 668750 }, { "epoch": 4.03275406712413, "grad_norm": 3.5945699214935303, "learning_rate": 0.00011497925787673026, "loss": 5.0741, "step": 668800 }, { "epoch": 4.033055558905464, "grad_norm": 1.788308024406433, "learning_rate": 0.00011495947160546216, "loss": 4.7621, "step": 668850 }, { "epoch": 4.033357050686798, "grad_norm": 2.263110637664795, "learning_rate": 0.00011493968597906505, "loss": 4.5015, "step": 668900 }, { "epoch": 4.033658542468133, "grad_norm": 2.2908129692077637, "learning_rate": 0.00011491990099790296, "loss": 4.3023, "step": 668950 }, { "epoch": 4.033960034249466, "grad_norm": 2.4312682151794434, "learning_rate": 0.00011490011666233999, "loss": 4.2265, "step": 669000 }, { "epoch": 4.033960034249466, "eval_loss": 5.2222795486450195, "eval_runtime": 39.2221, "eval_samples_per_second": 13.054, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.598986679448134, "step": 669000 }, { "epoch": 4.0342615260308, "grad_norm": 2.598074436187744, "learning_rate": 0.00011488033297274032, "loss": 4.59, "step": 669050 }, { "epoch": 4.034563017812134, "grad_norm": 3.476719617843628, "learning_rate": 0.00011486054992946798, "loss": 4.6249, "step": 669100 }, { "epoch": 4.0348645095934685, "grad_norm": 1.584890365600586, "learning_rate": 0.00011484076753288708, "loss": 4.9513, "step": 669150 }, { "epoch": 4.035166001374803, "grad_norm": 2.9909441471099854, "learning_rate": 0.00011482098578336172, "loss": 4.8713, "step": 669200 }, { "epoch": 4.035467493156137, "grad_norm": 1.9557558298110962, "learning_rate": 0.00011480120468125586, "loss": 4.6862, "step": 669250 }, { "epoch": 4.03576898493747, "grad_norm": 2.2475340366363525, "learning_rate": 0.0001147814242269336, "loss": 4.7179, "step": 669300 }, { "epoch": 4.0360704767188045, "grad_norm": 2.480482816696167, "learning_rate": 0.00011476164442075902, "loss": 4.3906, "step": 669350 }, { "epoch": 4.036371968500139, "grad_norm": 0.8698076009750366, "learning_rate": 0.00011474186526309603, "loss": 4.9383, "step": 669400 }, { "epoch": 4.036673460281473, "grad_norm": 2.252702236175537, "learning_rate": 0.00011472208675430866, "loss": 5.018, "step": 669450 }, { "epoch": 4.036974952062807, "grad_norm": 3.2194347381591797, "learning_rate": 0.00011470230889476103, "loss": 4.9935, "step": 669500 }, { "epoch": 4.03727644384414, "grad_norm": 1.7666865587234497, "learning_rate": 0.00011468253168481701, "loss": 4.3119, "step": 669550 }, { "epoch": 4.037577935625475, "grad_norm": 2.8152272701263428, "learning_rate": 0.00011466275512484054, "loss": 4.5999, "step": 669600 }, { "epoch": 4.037879427406809, "grad_norm": 2.7658276557922363, "learning_rate": 0.0001146429792151957, "loss": 4.8016, "step": 669650 }, { "epoch": 4.038180919188143, "grad_norm": 2.826592206954956, "learning_rate": 0.00011462320395624632, "loss": 4.553, "step": 669700 }, { "epoch": 4.038482410969477, "grad_norm": 3.359157085418701, "learning_rate": 0.0001146034293483564, "loss": 4.6004, "step": 669750 }, { "epoch": 4.0387839027508114, "grad_norm": 2.2970690727233887, "learning_rate": 0.00011458365539188989, "loss": 4.789, "step": 669800 }, { "epoch": 4.039085394532145, "grad_norm": 3.460230827331543, "learning_rate": 0.00011456388208721059, "loss": 4.6556, "step": 669850 }, { "epoch": 4.039386886313479, "grad_norm": 2.7176480293273926, "learning_rate": 0.00011454410943468251, "loss": 4.2816, "step": 669900 }, { "epoch": 4.039688378094813, "grad_norm": 2.588285207748413, "learning_rate": 0.00011452433743466954, "loss": 4.4947, "step": 669950 }, { "epoch": 4.039989869876147, "grad_norm": 0.7850621342658997, "learning_rate": 0.00011450456608753542, "loss": 4.6707, "step": 670000 }, { "epoch": 4.040291361657482, "grad_norm": 3.1909568309783936, "learning_rate": 0.00011448479539364413, "loss": 5.084, "step": 670050 }, { "epoch": 4.040592853438815, "grad_norm": 2.447314739227295, "learning_rate": 0.00011446502535335956, "loss": 4.867, "step": 670100 }, { "epoch": 4.040894345220149, "grad_norm": 2.4178459644317627, "learning_rate": 0.00011444525596704546, "loss": 5.2813, "step": 670150 }, { "epoch": 4.041195837001483, "grad_norm": 2.6160647869110107, "learning_rate": 0.00011442548723506564, "loss": 4.6933, "step": 670200 }, { "epoch": 4.0414973287828175, "grad_norm": 3.0894360542297363, "learning_rate": 0.00011440571915778402, "loss": 4.5357, "step": 670250 }, { "epoch": 4.041798820564152, "grad_norm": 0.9381284117698669, "learning_rate": 0.0001143859517355643, "loss": 4.5185, "step": 670300 }, { "epoch": 4.042100312345485, "grad_norm": 3.499829053878784, "learning_rate": 0.00011436618496877033, "loss": 4.9115, "step": 670350 }, { "epoch": 4.042401804126819, "grad_norm": 1.3857004642486572, "learning_rate": 0.0001143464188577659, "loss": 4.9716, "step": 670400 }, { "epoch": 4.0427032959081535, "grad_norm": 2.6911866664886475, "learning_rate": 0.00011432665340291469, "loss": 4.7576, "step": 670450 }, { "epoch": 4.043004787689488, "grad_norm": 2.920773506164551, "learning_rate": 0.00011430688860458054, "loss": 4.6257, "step": 670500 }, { "epoch": 4.043306279470822, "grad_norm": 2.5972402095794678, "learning_rate": 0.00011428712446312722, "loss": 4.6715, "step": 670550 }, { "epoch": 4.043607771252156, "grad_norm": 3.7199788093566895, "learning_rate": 0.00011426736097891834, "loss": 4.5909, "step": 670600 }, { "epoch": 4.043909263033489, "grad_norm": 2.7772035598754883, "learning_rate": 0.00011424759815231768, "loss": 4.6908, "step": 670650 }, { "epoch": 4.044210754814824, "grad_norm": 2.631068229675293, "learning_rate": 0.00011422783598368904, "loss": 4.8045, "step": 670700 }, { "epoch": 4.044512246596158, "grad_norm": 3.0234484672546387, "learning_rate": 0.00011420807447339597, "loss": 4.8, "step": 670750 }, { "epoch": 4.044813738377492, "grad_norm": 2.6371657848358154, "learning_rate": 0.00011418831362180219, "loss": 4.372, "step": 670800 }, { "epoch": 4.045115230158826, "grad_norm": 3.691436529159546, "learning_rate": 0.00011416855342927147, "loss": 4.8692, "step": 670850 }, { "epoch": 4.0454167219401596, "grad_norm": 3.269155740737915, "learning_rate": 0.0001141487938961673, "loss": 4.4694, "step": 670900 }, { "epoch": 4.045718213721494, "grad_norm": 1.909833312034607, "learning_rate": 0.00011412903502285345, "loss": 4.8897, "step": 670950 }, { "epoch": 4.046019705502828, "grad_norm": 2.465102434158325, "learning_rate": 0.00011410927680969358, "loss": 4.7438, "step": 671000 }, { "epoch": 4.046321197284162, "grad_norm": 3.618293046951294, "learning_rate": 0.00011408951925705116, "loss": 5.0799, "step": 671050 }, { "epoch": 4.046622689065496, "grad_norm": 6.65767765045166, "learning_rate": 0.00011406976236528994, "loss": 4.8199, "step": 671100 }, { "epoch": 4.04692418084683, "grad_norm": 2.673252820968628, "learning_rate": 0.00011405000613477346, "loss": 5.0303, "step": 671150 }, { "epoch": 4.047225672628164, "grad_norm": 0.8793462514877319, "learning_rate": 0.00011403025056586532, "loss": 4.51, "step": 671200 }, { "epoch": 4.047527164409498, "grad_norm": 2.9289278984069824, "learning_rate": 0.00011401049565892904, "loss": 5.0442, "step": 671250 }, { "epoch": 4.047828656190832, "grad_norm": 3.1363182067871094, "learning_rate": 0.00011399074141432831, "loss": 4.9312, "step": 671300 }, { "epoch": 4.0481301479721665, "grad_norm": 2.620419502258301, "learning_rate": 0.00011397098783242656, "loss": 5.0276, "step": 671350 }, { "epoch": 4.048431639753501, "grad_norm": 2.7090024948120117, "learning_rate": 0.00011395123491358731, "loss": 4.6474, "step": 671400 }, { "epoch": 4.048733131534834, "grad_norm": 2.5471231937408447, "learning_rate": 0.00011393148265817423, "loss": 5.2054, "step": 671450 }, { "epoch": 4.049034623316168, "grad_norm": 2.363292694091797, "learning_rate": 0.00011391173106655065, "loss": 4.6335, "step": 671500 }, { "epoch": 4.0493361150975025, "grad_norm": 1.8547446727752686, "learning_rate": 0.0001138919801390802, "loss": 4.7747, "step": 671550 }, { "epoch": 4.049637606878837, "grad_norm": 1.9796394109725952, "learning_rate": 0.00011387222987612636, "loss": 4.7269, "step": 671600 }, { "epoch": 4.049939098660171, "grad_norm": 3.216102123260498, "learning_rate": 0.00011385248027805251, "loss": 4.8523, "step": 671650 }, { "epoch": 4.050240590441504, "grad_norm": 4.2912702560424805, "learning_rate": 0.00011383273134522222, "loss": 4.5067, "step": 671700 }, { "epoch": 4.050542082222838, "grad_norm": 1.3866360187530518, "learning_rate": 0.00011381298307799885, "loss": 4.9967, "step": 671750 }, { "epoch": 4.050843574004173, "grad_norm": 3.0432963371276855, "learning_rate": 0.00011379323547674596, "loss": 5.1696, "step": 671800 }, { "epoch": 4.051145065785507, "grad_norm": 2.4304468631744385, "learning_rate": 0.00011377348854182684, "loss": 4.5927, "step": 671850 }, { "epoch": 4.051446557566841, "grad_norm": 1.3911031484603882, "learning_rate": 0.00011375374227360497, "loss": 4.4956, "step": 671900 }, { "epoch": 4.051748049348174, "grad_norm": 2.3379266262054443, "learning_rate": 0.00011373399667244378, "loss": 4.6607, "step": 671950 }, { "epoch": 4.0520495411295085, "grad_norm": 2.465743064880371, "learning_rate": 0.0001137142517387066, "loss": 4.8788, "step": 672000 }, { "epoch": 4.0520495411295085, "eval_loss": 5.208630084991455, "eval_runtime": 39.268, "eval_samples_per_second": 13.039, "eval_steps_per_second": 6.519, "eval_tts_loss": 7.652119280761353, "step": 672000 }, { "epoch": 4.052351032910843, "grad_norm": 2.411440849304199, "learning_rate": 0.0001136945074727568, "loss": 4.5948, "step": 672050 }, { "epoch": 4.052652524692177, "grad_norm": 2.1489765644073486, "learning_rate": 0.00011367476387495786, "loss": 4.66, "step": 672100 }, { "epoch": 4.052954016473511, "grad_norm": 2.3147857189178467, "learning_rate": 0.00011365502094567301, "loss": 5.0598, "step": 672150 }, { "epoch": 4.053255508254845, "grad_norm": 3.2820425033569336, "learning_rate": 0.0001136352786852656, "loss": 4.8276, "step": 672200 }, { "epoch": 4.053557000036179, "grad_norm": 1.4163085222244263, "learning_rate": 0.00011361553709409906, "loss": 4.875, "step": 672250 }, { "epoch": 4.053858491817513, "grad_norm": 0.6281382441520691, "learning_rate": 0.00011359579617253656, "loss": 4.5548, "step": 672300 }, { "epoch": 4.054159983598847, "grad_norm": 2.364466905593872, "learning_rate": 0.00011357605592094154, "loss": 4.7488, "step": 672350 }, { "epoch": 4.054461475380181, "grad_norm": 2.216277599334717, "learning_rate": 0.00011355631633967725, "loss": 4.7183, "step": 672400 }, { "epoch": 4.0547629671615155, "grad_norm": 1.040338158607483, "learning_rate": 0.00011353657742910688, "loss": 4.3682, "step": 672450 }, { "epoch": 4.055064458942849, "grad_norm": 3.1060383319854736, "learning_rate": 0.0001135168391895938, "loss": 4.9664, "step": 672500 }, { "epoch": 4.055365950724183, "grad_norm": 2.0969789028167725, "learning_rate": 0.00011349710162150125, "loss": 4.5964, "step": 672550 }, { "epoch": 4.055667442505517, "grad_norm": 2.4242334365844727, "learning_rate": 0.00011347736472519241, "loss": 4.5001, "step": 672600 }, { "epoch": 4.0559689342868515, "grad_norm": 3.20326828956604, "learning_rate": 0.00011345762850103055, "loss": 4.8569, "step": 672650 }, { "epoch": 4.056270426068186, "grad_norm": 1.4832532405853271, "learning_rate": 0.00011343789294937894, "loss": 4.9794, "step": 672700 }, { "epoch": 4.056571917849519, "grad_norm": 1.9906827211380005, "learning_rate": 0.00011341815807060072, "loss": 4.501, "step": 672750 }, { "epoch": 4.056873409630853, "grad_norm": 0.9555158615112305, "learning_rate": 0.00011339842386505905, "loss": 4.8544, "step": 672800 }, { "epoch": 4.057174901412187, "grad_norm": 2.2403175830841064, "learning_rate": 0.00011337869033311722, "loss": 4.6712, "step": 672850 }, { "epoch": 4.057476393193522, "grad_norm": 3.0624234676361084, "learning_rate": 0.00011335895747513826, "loss": 4.7568, "step": 672900 }, { "epoch": 4.057777884974856, "grad_norm": 3.2889583110809326, "learning_rate": 0.00011333922529148544, "loss": 5.0362, "step": 672950 }, { "epoch": 4.05807937675619, "grad_norm": 2.421083688735962, "learning_rate": 0.0001133194937825219, "loss": 5.0721, "step": 673000 }, { "epoch": 4.058380868537523, "grad_norm": 2.660362482070923, "learning_rate": 0.00011329976294861065, "loss": 4.2381, "step": 673050 }, { "epoch": 4.0586823603188575, "grad_norm": 0.6782130002975464, "learning_rate": 0.00011328003279011493, "loss": 4.9991, "step": 673100 }, { "epoch": 4.058983852100192, "grad_norm": 1.0900537967681885, "learning_rate": 0.00011326030330739784, "loss": 4.7443, "step": 673150 }, { "epoch": 4.059285343881526, "grad_norm": 1.7214066982269287, "learning_rate": 0.00011324057450082237, "loss": 5.1525, "step": 673200 }, { "epoch": 4.05958683566286, "grad_norm": 2.1297624111175537, "learning_rate": 0.00011322084637075166, "loss": 4.7176, "step": 673250 }, { "epoch": 4.0598883274441935, "grad_norm": 3.809281349182129, "learning_rate": 0.00011320111891754887, "loss": 4.6151, "step": 673300 }, { "epoch": 4.060189819225528, "grad_norm": 0.7610174417495728, "learning_rate": 0.00011318139214157691, "loss": 4.985, "step": 673350 }, { "epoch": 4.060491311006862, "grad_norm": 2.28684139251709, "learning_rate": 0.00011316166604319889, "loss": 4.7135, "step": 673400 }, { "epoch": 4.060792802788196, "grad_norm": 2.280642509460449, "learning_rate": 0.00011314194062277786, "loss": 4.5492, "step": 673450 }, { "epoch": 4.06109429456953, "grad_norm": 3.1010780334472656, "learning_rate": 0.00011312221588067677, "loss": 5.1681, "step": 673500 }, { "epoch": 4.0613957863508645, "grad_norm": 2.625664472579956, "learning_rate": 0.0001131024918172587, "loss": 4.663, "step": 673550 }, { "epoch": 4.061697278132198, "grad_norm": 2.9394779205322266, "learning_rate": 0.00011308276843288664, "loss": 4.6808, "step": 673600 }, { "epoch": 4.061998769913532, "grad_norm": 1.4306871891021729, "learning_rate": 0.0001130630457279235, "loss": 4.9579, "step": 673650 }, { "epoch": 4.062300261694866, "grad_norm": 2.403536319732666, "learning_rate": 0.0001130433237027323, "loss": 4.7495, "step": 673700 }, { "epoch": 4.0626017534762005, "grad_norm": 0.8275432586669922, "learning_rate": 0.00011302360235767602, "loss": 4.5882, "step": 673750 }, { "epoch": 4.062903245257535, "grad_norm": 2.172485113143921, "learning_rate": 0.00011300388169311753, "loss": 4.4759, "step": 673800 }, { "epoch": 4.063204737038868, "grad_norm": 4.171360969543457, "learning_rate": 0.00011298416170941977, "loss": 4.449, "step": 673850 }, { "epoch": 4.063506228820202, "grad_norm": 3.1477866172790527, "learning_rate": 0.00011296444240694578, "loss": 4.44, "step": 673900 }, { "epoch": 4.063807720601536, "grad_norm": 2.1039013862609863, "learning_rate": 0.00011294472378605832, "loss": 4.7015, "step": 673950 }, { "epoch": 4.064109212382871, "grad_norm": 3.674224853515625, "learning_rate": 0.00011292500584712032, "loss": 4.8739, "step": 674000 }, { "epoch": 4.064410704164205, "grad_norm": 1.8096381425857544, "learning_rate": 0.00011290528859049475, "loss": 4.8177, "step": 674050 }, { "epoch": 4.064712195945538, "grad_norm": 0.8193411827087402, "learning_rate": 0.00011288557201654431, "loss": 4.4638, "step": 674100 }, { "epoch": 4.065013687726872, "grad_norm": 2.2382218837738037, "learning_rate": 0.000112865856125632, "loss": 5.024, "step": 674150 }, { "epoch": 4.0653151795082065, "grad_norm": 1.7939825057983398, "learning_rate": 0.00011284614091812066, "loss": 4.5527, "step": 674200 }, { "epoch": 4.065616671289541, "grad_norm": 2.4984240531921387, "learning_rate": 0.00011282642639437301, "loss": 4.5655, "step": 674250 }, { "epoch": 4.065918163070875, "grad_norm": 2.170095682144165, "learning_rate": 0.00011280671255475196, "loss": 4.8412, "step": 674300 }, { "epoch": 4.066219654852209, "grad_norm": 1.1091716289520264, "learning_rate": 0.00011278699939962031, "loss": 4.961, "step": 674350 }, { "epoch": 4.0665211466335425, "grad_norm": 2.214663505554199, "learning_rate": 0.00011276728692934077, "loss": 4.4804, "step": 674400 }, { "epoch": 4.066822638414877, "grad_norm": 2.796013355255127, "learning_rate": 0.00011274757514427618, "loss": 4.4726, "step": 674450 }, { "epoch": 4.067124130196211, "grad_norm": 2.9275081157684326, "learning_rate": 0.00011272786404478938, "loss": 4.808, "step": 674500 }, { "epoch": 4.067425621977545, "grad_norm": 3.0583717823028564, "learning_rate": 0.00011270815363124301, "loss": 4.4643, "step": 674550 }, { "epoch": 4.067727113758879, "grad_norm": 2.0052490234375, "learning_rate": 0.00011268844390399984, "loss": 4.7662, "step": 674600 }, { "epoch": 4.068028605540213, "grad_norm": 2.8118207454681396, "learning_rate": 0.00011266873486342268, "loss": 4.7838, "step": 674650 }, { "epoch": 4.068330097321547, "grad_norm": 2.198965072631836, "learning_rate": 0.00011264902650987407, "loss": 5.0543, "step": 674700 }, { "epoch": 4.068631589102881, "grad_norm": 2.0790855884552, "learning_rate": 0.0001126293188437169, "loss": 4.9603, "step": 674750 }, { "epoch": 4.068933080884215, "grad_norm": 2.009834051132202, "learning_rate": 0.0001126096118653138, "loss": 4.7186, "step": 674800 }, { "epoch": 4.069234572665549, "grad_norm": 3.5452160835266113, "learning_rate": 0.00011258990557502738, "loss": 4.5251, "step": 674850 }, { "epoch": 4.069536064446883, "grad_norm": 2.839146852493286, "learning_rate": 0.0001125701999732204, "loss": 5.0929, "step": 674900 }, { "epoch": 4.069837556228217, "grad_norm": 1.8485090732574463, "learning_rate": 0.00011255049506025547, "loss": 4.6593, "step": 674950 }, { "epoch": 4.070139048009551, "grad_norm": 4.638018608093262, "learning_rate": 0.00011253079083649522, "loss": 5.0745, "step": 675000 }, { "epoch": 4.070139048009551, "eval_loss": 5.203834533691406, "eval_runtime": 39.0543, "eval_samples_per_second": 13.11, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.652630332637699, "step": 675000 }, { "epoch": 4.070440539790885, "grad_norm": 2.041872262954712, "learning_rate": 0.00011251108730230225, "loss": 4.8422, "step": 675050 }, { "epoch": 4.07074203157222, "grad_norm": 2.217167615890503, "learning_rate": 0.0001124913844580393, "loss": 4.7028, "step": 675100 }, { "epoch": 4.071043523353554, "grad_norm": 2.47389817237854, "learning_rate": 0.00011247168230406888, "loss": 4.5064, "step": 675150 }, { "epoch": 4.071345015134887, "grad_norm": 2.4204342365264893, "learning_rate": 0.00011245198084075353, "loss": 5.2867, "step": 675200 }, { "epoch": 4.071646506916221, "grad_norm": 2.2710580825805664, "learning_rate": 0.0001124322800684559, "loss": 4.8486, "step": 675250 }, { "epoch": 4.0719479986975555, "grad_norm": 2.122847080230713, "learning_rate": 0.00011241257998753861, "loss": 4.6577, "step": 675300 }, { "epoch": 4.07224949047889, "grad_norm": 0.9890849590301514, "learning_rate": 0.00011239288059836412, "loss": 4.4481, "step": 675350 }, { "epoch": 4.072550982260224, "grad_norm": 2.188039541244507, "learning_rate": 0.00011237318190129495, "loss": 4.5592, "step": 675400 }, { "epoch": 4.072852474041557, "grad_norm": 2.1579322814941406, "learning_rate": 0.00011235348389669374, "loss": 4.6203, "step": 675450 }, { "epoch": 4.0731539658228915, "grad_norm": 1.5653462409973145, "learning_rate": 0.0001123337865849229, "loss": 4.799, "step": 675500 }, { "epoch": 4.073455457604226, "grad_norm": 2.1483476161956787, "learning_rate": 0.00011231408996634492, "loss": 4.723, "step": 675550 }, { "epoch": 4.07375694938556, "grad_norm": 2.408221960067749, "learning_rate": 0.0001122943940413224, "loss": 4.8466, "step": 675600 }, { "epoch": 4.074058441166894, "grad_norm": 2.110931634902954, "learning_rate": 0.0001122746988102177, "loss": 4.6142, "step": 675650 }, { "epoch": 4.074359932948227, "grad_norm": 1.8772456645965576, "learning_rate": 0.00011225500427339335, "loss": 4.9091, "step": 675700 }, { "epoch": 4.074661424729562, "grad_norm": 2.1431009769439697, "learning_rate": 0.00011223531043121184, "loss": 4.5416, "step": 675750 }, { "epoch": 4.074962916510896, "grad_norm": 2.1430835723876953, "learning_rate": 0.00011221561728403547, "loss": 4.8356, "step": 675800 }, { "epoch": 4.07526440829223, "grad_norm": 2.129171371459961, "learning_rate": 0.00011219592483222673, "loss": 4.6996, "step": 675850 }, { "epoch": 4.075565900073564, "grad_norm": 2.4601008892059326, "learning_rate": 0.00011217623307614811, "loss": 5.0003, "step": 675900 }, { "epoch": 4.075867391854898, "grad_norm": 2.9194281101226807, "learning_rate": 0.00011215654201616192, "loss": 4.4247, "step": 675950 }, { "epoch": 4.076168883636232, "grad_norm": 2.248708963394165, "learning_rate": 0.00011213685165263055, "loss": 4.7629, "step": 676000 }, { "epoch": 4.076470375417566, "grad_norm": 2.246586561203003, "learning_rate": 0.00011211716198591643, "loss": 4.7869, "step": 676050 }, { "epoch": 4.0767718671989, "grad_norm": 3.6592023372650146, "learning_rate": 0.00011209747301638186, "loss": 5.0357, "step": 676100 }, { "epoch": 4.077073358980234, "grad_norm": 2.590548276901245, "learning_rate": 0.00011207778474438919, "loss": 4.5124, "step": 676150 }, { "epoch": 4.077374850761569, "grad_norm": 2.908508062362671, "learning_rate": 0.00011205809717030083, "loss": 5.0121, "step": 676200 }, { "epoch": 4.077676342542902, "grad_norm": 2.601571798324585, "learning_rate": 0.00011203841029447897, "loss": 5.0158, "step": 676250 }, { "epoch": 4.077977834324236, "grad_norm": 2.3615150451660156, "learning_rate": 0.00011201872411728603, "loss": 4.8054, "step": 676300 }, { "epoch": 4.07827932610557, "grad_norm": 2.3309624195098877, "learning_rate": 0.0001119990386390843, "loss": 5.0011, "step": 676350 }, { "epoch": 4.0785808178869045, "grad_norm": 3.1314268112182617, "learning_rate": 0.00011197935386023599, "loss": 4.6965, "step": 676400 }, { "epoch": 4.078882309668239, "grad_norm": 2.5802059173583984, "learning_rate": 0.00011195966978110339, "loss": 5.0804, "step": 676450 }, { "epoch": 4.079183801449572, "grad_norm": 1.8783352375030518, "learning_rate": 0.00011193998640204885, "loss": 4.9353, "step": 676500 }, { "epoch": 4.079485293230906, "grad_norm": 3.0567238330841064, "learning_rate": 0.00011192030372343452, "loss": 4.689, "step": 676550 }, { "epoch": 4.0797867850122405, "grad_norm": 2.7916221618652344, "learning_rate": 0.00011190062174562262, "loss": 4.7025, "step": 676600 }, { "epoch": 4.080088276793575, "grad_norm": 2.3450682163238525, "learning_rate": 0.00011188094046897549, "loss": 4.4587, "step": 676650 }, { "epoch": 4.080389768574909, "grad_norm": 2.84149169921875, "learning_rate": 0.00011186125989385521, "loss": 5.1493, "step": 676700 }, { "epoch": 4.080691260356243, "grad_norm": 3.7212564945220947, "learning_rate": 0.00011184158002062397, "loss": 4.6768, "step": 676750 }, { "epoch": 4.080992752137576, "grad_norm": 2.9141452312469482, "learning_rate": 0.00011182190084964405, "loss": 4.8139, "step": 676800 }, { "epoch": 4.081294243918911, "grad_norm": 2.473752498626709, "learning_rate": 0.00011180222238127752, "loss": 5.2657, "step": 676850 }, { "epoch": 4.081595735700245, "grad_norm": 1.8616830110549927, "learning_rate": 0.0001117825446158866, "loss": 4.8901, "step": 676900 }, { "epoch": 4.081897227481579, "grad_norm": 2.7644340991973877, "learning_rate": 0.00011176286755383344, "loss": 4.678, "step": 676950 }, { "epoch": 4.082198719262913, "grad_norm": 3.3177871704101562, "learning_rate": 0.00011174319119548008, "loss": 4.6754, "step": 677000 }, { "epoch": 4.0825002110442465, "grad_norm": 1.7593082189559937, "learning_rate": 0.00011172351554118867, "loss": 4.8411, "step": 677050 }, { "epoch": 4.082801702825581, "grad_norm": 2.421251058578491, "learning_rate": 0.00011170384059132141, "loss": 5.2106, "step": 677100 }, { "epoch": 4.083103194606915, "grad_norm": 1.6093065738677979, "learning_rate": 0.00011168416634624029, "loss": 4.9617, "step": 677150 }, { "epoch": 4.083404686388249, "grad_norm": 2.7246341705322266, "learning_rate": 0.00011166449280630735, "loss": 4.5686, "step": 677200 }, { "epoch": 4.083706178169583, "grad_norm": 2.280808210372925, "learning_rate": 0.0001116448199718848, "loss": 4.895, "step": 677250 }, { "epoch": 4.084007669950918, "grad_norm": 2.1019654273986816, "learning_rate": 0.00011162514784333453, "loss": 4.7316, "step": 677300 }, { "epoch": 4.084309161732251, "grad_norm": 1.266408920288086, "learning_rate": 0.0001116054764210187, "loss": 4.8483, "step": 677350 }, { "epoch": 4.084610653513585, "grad_norm": 3.533229112625122, "learning_rate": 0.00011158580570529928, "loss": 4.3851, "step": 677400 }, { "epoch": 4.084912145294919, "grad_norm": 2.8167569637298584, "learning_rate": 0.00011156613569653824, "loss": 4.7532, "step": 677450 }, { "epoch": 4.0852136370762535, "grad_norm": 3.70683217048645, "learning_rate": 0.00011154646639509766, "loss": 4.8258, "step": 677500 }, { "epoch": 4.085515128857588, "grad_norm": 1.1664133071899414, "learning_rate": 0.00011152679780133953, "loss": 4.9617, "step": 677550 }, { "epoch": 4.085816620638921, "grad_norm": 2.0384113788604736, "learning_rate": 0.0001115071299156257, "loss": 4.4052, "step": 677600 }, { "epoch": 4.086118112420255, "grad_norm": 2.1968231201171875, "learning_rate": 0.00011148746273831822, "loss": 4.8496, "step": 677650 }, { "epoch": 4.0864196042015895, "grad_norm": 2.4358069896698, "learning_rate": 0.0001114677962697791, "loss": 4.9289, "step": 677700 }, { "epoch": 4.086721095982924, "grad_norm": 3.0097298622131348, "learning_rate": 0.00011144813051037015, "loss": 4.6642, "step": 677750 }, { "epoch": 4.087022587764258, "grad_norm": 2.2997124195098877, "learning_rate": 0.00011142846546045332, "loss": 4.6914, "step": 677800 }, { "epoch": 4.087324079545591, "grad_norm": 2.4854750633239746, "learning_rate": 0.00011140880112039061, "loss": 4.6118, "step": 677850 }, { "epoch": 4.087625571326925, "grad_norm": 2.3813376426696777, "learning_rate": 0.00011138913749054377, "loss": 4.9171, "step": 677900 }, { "epoch": 4.08792706310826, "grad_norm": 2.8866989612579346, "learning_rate": 0.00011136947457127479, "loss": 4.5958, "step": 677950 }, { "epoch": 4.088228554889594, "grad_norm": 0.9880667924880981, "learning_rate": 0.00011134981236294553, "loss": 4.7246, "step": 678000 }, { "epoch": 4.088228554889594, "eval_loss": 5.214890480041504, "eval_runtime": 39.179, "eval_samples_per_second": 13.068, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.576622078035799, "step": 678000 }, { "epoch": 4.088530046670928, "grad_norm": 2.240396738052368, "learning_rate": 0.00011133015086591775, "loss": 5.0549, "step": 678050 }, { "epoch": 4.088831538452262, "grad_norm": 2.2487192153930664, "learning_rate": 0.0001113104900805534, "loss": 4.8495, "step": 678100 }, { "epoch": 4.0891330302335955, "grad_norm": 4.19464111328125, "learning_rate": 0.00011129083000721429, "loss": 5.0405, "step": 678150 }, { "epoch": 4.08943452201493, "grad_norm": 3.124941825866699, "learning_rate": 0.00011127117064626215, "loss": 4.3063, "step": 678200 }, { "epoch": 4.089736013796264, "grad_norm": 2.374706745147705, "learning_rate": 0.00011125151199805885, "loss": 4.7932, "step": 678250 }, { "epoch": 4.090037505577598, "grad_norm": 3.7543132305145264, "learning_rate": 0.00011123185406296623, "loss": 4.9219, "step": 678300 }, { "epoch": 4.090338997358932, "grad_norm": 2.3348093032836914, "learning_rate": 0.000111212196841346, "loss": 4.736, "step": 678350 }, { "epoch": 4.090640489140266, "grad_norm": 1.7853182554244995, "learning_rate": 0.00011119254033355988, "loss": 4.4802, "step": 678400 }, { "epoch": 4.0909419809216, "grad_norm": 2.2580859661102295, "learning_rate": 0.00011117288453996975, "loss": 4.928, "step": 678450 }, { "epoch": 4.091243472702934, "grad_norm": 1.7463061809539795, "learning_rate": 0.00011115322946093721, "loss": 4.7113, "step": 678500 }, { "epoch": 4.091544964484268, "grad_norm": 2.334303379058838, "learning_rate": 0.00011113357509682405, "loss": 4.9042, "step": 678550 }, { "epoch": 4.0918464562656025, "grad_norm": 2.4495725631713867, "learning_rate": 0.00011111392144799197, "loss": 4.4835, "step": 678600 }, { "epoch": 4.092147948046936, "grad_norm": 2.1584506034851074, "learning_rate": 0.00011109426851480274, "loss": 4.5931, "step": 678650 }, { "epoch": 4.09244943982827, "grad_norm": 3.239877939224243, "learning_rate": 0.00011107461629761791, "loss": 4.7463, "step": 678700 }, { "epoch": 4.092750931609604, "grad_norm": 2.0839338302612305, "learning_rate": 0.00011105496479679921, "loss": 5.1442, "step": 678750 }, { "epoch": 4.093052423390938, "grad_norm": 2.9905810356140137, "learning_rate": 0.00011103531401270837, "loss": 5.0018, "step": 678800 }, { "epoch": 4.093353915172273, "grad_norm": 3.9814834594726562, "learning_rate": 0.0001110156639457069, "loss": 4.6079, "step": 678850 }, { "epoch": 4.093655406953607, "grad_norm": 3.9101996421813965, "learning_rate": 0.00011099601459615655, "loss": 4.9241, "step": 678900 }, { "epoch": 4.09395689873494, "grad_norm": 2.59745192527771, "learning_rate": 0.0001109763659644189, "loss": 4.6059, "step": 678950 }, { "epoch": 4.094258390516274, "grad_norm": 3.2980175018310547, "learning_rate": 0.0001109567180508555, "loss": 4.6936, "step": 679000 }, { "epoch": 4.094559882297609, "grad_norm": 2.1383097171783447, "learning_rate": 0.00011093707085582796, "loss": 4.6756, "step": 679050 }, { "epoch": 4.094861374078943, "grad_norm": 3.9186954498291016, "learning_rate": 0.00011091742437969796, "loss": 4.6644, "step": 679100 }, { "epoch": 4.095162865860277, "grad_norm": 4.359733581542969, "learning_rate": 0.00011089777862282696, "loss": 4.5566, "step": 679150 }, { "epoch": 4.09546435764161, "grad_norm": 3.0828402042388916, "learning_rate": 0.00011087813358557652, "loss": 4.9151, "step": 679200 }, { "epoch": 4.0957658494229445, "grad_norm": 1.478723168373108, "learning_rate": 0.00011085848926830826, "loss": 4.7472, "step": 679250 }, { "epoch": 4.096067341204279, "grad_norm": 2.2255542278289795, "learning_rate": 0.00011083884567138361, "loss": 4.2838, "step": 679300 }, { "epoch": 4.096368832985613, "grad_norm": 2.735400676727295, "learning_rate": 0.0001108192027951641, "loss": 4.7624, "step": 679350 }, { "epoch": 4.096670324766947, "grad_norm": 2.352778673171997, "learning_rate": 0.0001107995606400113, "loss": 4.7305, "step": 679400 }, { "epoch": 4.0969718165482805, "grad_norm": 2.8546934127807617, "learning_rate": 0.0001107799192062866, "loss": 4.9371, "step": 679450 }, { "epoch": 4.097273308329615, "grad_norm": 1.7019765377044678, "learning_rate": 0.00011076027849435153, "loss": 4.8895, "step": 679500 }, { "epoch": 4.097574800110949, "grad_norm": 3.4980862140655518, "learning_rate": 0.00011074063850456756, "loss": 4.9387, "step": 679550 }, { "epoch": 4.097876291892283, "grad_norm": 2.963141918182373, "learning_rate": 0.00011072099923729605, "loss": 4.9277, "step": 679600 }, { "epoch": 4.098177783673617, "grad_norm": 2.854771375656128, "learning_rate": 0.0001107013606928985, "loss": 4.8072, "step": 679650 }, { "epoch": 4.0984792754549515, "grad_norm": 1.7308318614959717, "learning_rate": 0.00011068172287173642, "loss": 4.5671, "step": 679700 }, { "epoch": 4.098780767236285, "grad_norm": 2.733274221420288, "learning_rate": 0.00011066208577417106, "loss": 4.3365, "step": 679750 }, { "epoch": 4.099082259017619, "grad_norm": 2.7133424282073975, "learning_rate": 0.00011064244940056384, "loss": 4.5851, "step": 679800 }, { "epoch": 4.099383750798953, "grad_norm": 2.1319186687469482, "learning_rate": 0.00011062281375127624, "loss": 4.7179, "step": 679850 }, { "epoch": 4.099685242580287, "grad_norm": 1.592800498008728, "learning_rate": 0.00011060317882666952, "loss": 4.897, "step": 679900 }, { "epoch": 4.099986734361622, "grad_norm": 2.6649298667907715, "learning_rate": 0.00011058354462710503, "loss": 4.6415, "step": 679950 }, { "epoch": 4.100288226142955, "grad_norm": 2.183281421661377, "learning_rate": 0.00011056391115294425, "loss": 4.7885, "step": 680000 }, { "epoch": 4.100589717924289, "grad_norm": 6.705356121063232, "learning_rate": 0.00011054427840454832, "loss": 4.4928, "step": 680050 }, { "epoch": 4.100891209705623, "grad_norm": 3.215831995010376, "learning_rate": 0.00011052464638227868, "loss": 4.7, "step": 680100 }, { "epoch": 4.101192701486958, "grad_norm": 1.9154548645019531, "learning_rate": 0.00011050501508649661, "loss": 5.0087, "step": 680150 }, { "epoch": 4.101494193268292, "grad_norm": 1.4285478591918945, "learning_rate": 0.00011048538451756332, "loss": 4.8654, "step": 680200 }, { "epoch": 4.101795685049625, "grad_norm": 1.8993496894836426, "learning_rate": 0.00011046575467584014, "loss": 5.0504, "step": 680250 }, { "epoch": 4.102097176830959, "grad_norm": 1.277282953262329, "learning_rate": 0.00011044612556168837, "loss": 4.5435, "step": 680300 }, { "epoch": 4.1023986686122935, "grad_norm": 2.635897636413574, "learning_rate": 0.0001104264971754692, "loss": 4.7827, "step": 680350 }, { "epoch": 4.102700160393628, "grad_norm": 2.8204121589660645, "learning_rate": 0.00011040686951754385, "loss": 4.5891, "step": 680400 }, { "epoch": 4.103001652174962, "grad_norm": 2.27592134475708, "learning_rate": 0.0001103872425882736, "loss": 5.082, "step": 680450 }, { "epoch": 4.103303143956296, "grad_norm": 2.0110342502593994, "learning_rate": 0.00011036761638801961, "loss": 4.9877, "step": 680500 }, { "epoch": 4.1036046357376295, "grad_norm": 3.008265972137451, "learning_rate": 0.00011034799091714305, "loss": 4.9385, "step": 680550 }, { "epoch": 4.103906127518964, "grad_norm": 2.713026285171509, "learning_rate": 0.00011032836617600519, "loss": 4.3654, "step": 680600 }, { "epoch": 4.104207619300298, "grad_norm": 2.7456655502319336, "learning_rate": 0.00011030874216496707, "loss": 4.9411, "step": 680650 }, { "epoch": 4.104509111081632, "grad_norm": 2.1556928157806396, "learning_rate": 0.00011028911888438992, "loss": 4.6566, "step": 680700 }, { "epoch": 4.104810602862966, "grad_norm": 2.0030975341796875, "learning_rate": 0.00011026949633463494, "loss": 4.9877, "step": 680750 }, { "epoch": 4.1051120946443, "grad_norm": 1.8504908084869385, "learning_rate": 0.0001102498745160631, "loss": 4.5091, "step": 680800 }, { "epoch": 4.105413586425634, "grad_norm": 2.469730854034424, "learning_rate": 0.00011023025342903558, "loss": 4.802, "step": 680850 }, { "epoch": 4.105715078206968, "grad_norm": 3.1008031368255615, "learning_rate": 0.00011021063307391357, "loss": 4.756, "step": 680900 }, { "epoch": 4.106016569988302, "grad_norm": 2.794893980026245, "learning_rate": 0.00011019101345105803, "loss": 4.1981, "step": 680950 }, { "epoch": 4.106318061769636, "grad_norm": 2.991787910461426, "learning_rate": 0.00011017139456083007, "loss": 5.1936, "step": 681000 }, { "epoch": 4.106318061769636, "eval_loss": 5.206184387207031, "eval_runtime": 39.0429, "eval_samples_per_second": 13.114, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.598071810868941, "step": 681000 }, { "epoch": 4.106619553550971, "grad_norm": 2.383589029312134, "learning_rate": 0.0001101517764035908, "loss": 4.9537, "step": 681050 }, { "epoch": 4.106921045332304, "grad_norm": 4.014745712280273, "learning_rate": 0.00011013215897970118, "loss": 4.499, "step": 681100 }, { "epoch": 4.107222537113638, "grad_norm": 1.701340913772583, "learning_rate": 0.00011011254228952225, "loss": 4.5932, "step": 681150 }, { "epoch": 4.107524028894972, "grad_norm": 2.905210494995117, "learning_rate": 0.00011009292633341514, "loss": 5.1249, "step": 681200 }, { "epoch": 4.107825520676307, "grad_norm": 2.0629818439483643, "learning_rate": 0.0001100733111117407, "loss": 4.7533, "step": 681250 }, { "epoch": 4.108127012457641, "grad_norm": 3.4807329177856445, "learning_rate": 0.00011005369662486001, "loss": 5.006, "step": 681300 }, { "epoch": 4.108428504238974, "grad_norm": 0.969086766242981, "learning_rate": 0.00011003408287313408, "loss": 4.9389, "step": 681350 }, { "epoch": 4.108729996020308, "grad_norm": 3.303319215774536, "learning_rate": 0.00011001446985692372, "loss": 4.9823, "step": 681400 }, { "epoch": 4.1090314878016425, "grad_norm": 3.3524608612060547, "learning_rate": 0.00010999485757659, "loss": 4.847, "step": 681450 }, { "epoch": 4.109332979582977, "grad_norm": 2.1357598304748535, "learning_rate": 0.00010997524603249391, "loss": 5.1291, "step": 681500 }, { "epoch": 4.109634471364311, "grad_norm": 1.9962010383605957, "learning_rate": 0.00010995563522499626, "loss": 4.7149, "step": 681550 }, { "epoch": 4.109935963145644, "grad_norm": 2.9128785133361816, "learning_rate": 0.00010993602515445796, "loss": 4.6789, "step": 681600 }, { "epoch": 4.1102374549269785, "grad_norm": 2.151736259460449, "learning_rate": 0.00010991641582124003, "loss": 4.6873, "step": 681650 }, { "epoch": 4.110538946708313, "grad_norm": 1.601859450340271, "learning_rate": 0.0001098968072257032, "loss": 4.738, "step": 681700 }, { "epoch": 4.110840438489647, "grad_norm": 1.6740673780441284, "learning_rate": 0.00010987719936820839, "loss": 4.7361, "step": 681750 }, { "epoch": 4.111141930270981, "grad_norm": 3.375270128250122, "learning_rate": 0.0001098575922491165, "loss": 4.82, "step": 681800 }, { "epoch": 4.111443422052315, "grad_norm": 3.6690914630889893, "learning_rate": 0.00010983798586878832, "loss": 4.8081, "step": 681850 }, { "epoch": 4.111744913833649, "grad_norm": 3.332242250442505, "learning_rate": 0.00010981838022758471, "loss": 4.6803, "step": 681900 }, { "epoch": 4.112046405614983, "grad_norm": 2.376007318496704, "learning_rate": 0.00010979877532586641, "loss": 4.7957, "step": 681950 }, { "epoch": 4.112347897396317, "grad_norm": 2.5744283199310303, "learning_rate": 0.00010977917116399439, "loss": 4.8072, "step": 682000 }, { "epoch": 4.112649389177651, "grad_norm": 2.856696128845215, "learning_rate": 0.00010975956774232924, "loss": 4.5509, "step": 682050 }, { "epoch": 4.112950880958985, "grad_norm": 0.8100038766860962, "learning_rate": 0.00010973996506123184, "loss": 4.8557, "step": 682100 }, { "epoch": 4.113252372740319, "grad_norm": 2.50612735748291, "learning_rate": 0.00010972036312106298, "loss": 4.4951, "step": 682150 }, { "epoch": 4.113553864521653, "grad_norm": 2.411978006362915, "learning_rate": 0.00010970076192218327, "loss": 4.8963, "step": 682200 }, { "epoch": 4.113855356302987, "grad_norm": 2.433539628982544, "learning_rate": 0.00010968116146495359, "loss": 4.3672, "step": 682250 }, { "epoch": 4.114156848084321, "grad_norm": 2.316722869873047, "learning_rate": 0.00010966156174973461, "loss": 4.5053, "step": 682300 }, { "epoch": 4.114458339865656, "grad_norm": 2.657097339630127, "learning_rate": 0.00010964196277688696, "loss": 4.7698, "step": 682350 }, { "epoch": 4.114759831646989, "grad_norm": 3.413698434829712, "learning_rate": 0.00010962236454677139, "loss": 4.7458, "step": 682400 }, { "epoch": 4.115061323428323, "grad_norm": 1.6830503940582275, "learning_rate": 0.00010960276705974864, "loss": 4.9322, "step": 682450 }, { "epoch": 4.115362815209657, "grad_norm": 4.450389862060547, "learning_rate": 0.00010958317031617928, "loss": 4.5524, "step": 682500 }, { "epoch": 4.1156643069909915, "grad_norm": 3.5808932781219482, "learning_rate": 0.00010956357431642395, "loss": 5.1474, "step": 682550 }, { "epoch": 4.115965798772326, "grad_norm": 1.805727481842041, "learning_rate": 0.0001095439790608434, "loss": 4.7251, "step": 682600 }, { "epoch": 4.11626729055366, "grad_norm": 3.5251290798187256, "learning_rate": 0.00010952438454979812, "loss": 5.0642, "step": 682650 }, { "epoch": 4.116568782334993, "grad_norm": 3.05717134475708, "learning_rate": 0.00010950479078364881, "loss": 4.5428, "step": 682700 }, { "epoch": 4.116870274116327, "grad_norm": 2.30415940284729, "learning_rate": 0.00010948519776275608, "loss": 4.5613, "step": 682750 }, { "epoch": 4.117171765897662, "grad_norm": 2.8227009773254395, "learning_rate": 0.0001094656054874804, "loss": 4.1657, "step": 682800 }, { "epoch": 4.117473257678996, "grad_norm": 2.444220781326294, "learning_rate": 0.00010944601395818242, "loss": 4.5607, "step": 682850 }, { "epoch": 4.11777474946033, "grad_norm": 2.4840087890625, "learning_rate": 0.00010942642317522269, "loss": 5.0282, "step": 682900 }, { "epoch": 4.118076241241663, "grad_norm": 3.3666412830352783, "learning_rate": 0.00010940683313896175, "loss": 4.8182, "step": 682950 }, { "epoch": 4.118377733022998, "grad_norm": 2.476284980773926, "learning_rate": 0.00010938724384976006, "loss": 4.6093, "step": 683000 }, { "epoch": 4.118679224804332, "grad_norm": 1.4355134963989258, "learning_rate": 0.00010936765530797828, "loss": 4.7606, "step": 683050 }, { "epoch": 4.118980716585666, "grad_norm": 4.440385341644287, "learning_rate": 0.00010934806751397676, "loss": 4.3272, "step": 683100 }, { "epoch": 4.119282208367, "grad_norm": 1.3709688186645508, "learning_rate": 0.00010932848046811605, "loss": 4.8913, "step": 683150 }, { "epoch": 4.1195837001483335, "grad_norm": 2.071300983428955, "learning_rate": 0.00010930889417075665, "loss": 4.8733, "step": 683200 }, { "epoch": 4.119885191929668, "grad_norm": 2.4953811168670654, "learning_rate": 0.00010928930862225892, "loss": 4.5515, "step": 683250 }, { "epoch": 4.120186683711002, "grad_norm": 2.5106537342071533, "learning_rate": 0.00010926972382298343, "loss": 4.4805, "step": 683300 }, { "epoch": 4.120488175492336, "grad_norm": 1.650060772895813, "learning_rate": 0.00010925013977329054, "loss": 4.6686, "step": 683350 }, { "epoch": 4.12078966727367, "grad_norm": 2.747535228729248, "learning_rate": 0.00010923055647354064, "loss": 4.8033, "step": 683400 }, { "epoch": 4.121091159055005, "grad_norm": 2.90147066116333, "learning_rate": 0.00010921097392409417, "loss": 4.7061, "step": 683450 }, { "epoch": 4.121392650836338, "grad_norm": 2.400484085083008, "learning_rate": 0.00010919139212531153, "loss": 4.8346, "step": 683500 }, { "epoch": 4.121694142617672, "grad_norm": 1.3350363969802856, "learning_rate": 0.00010917181107755309, "loss": 4.7935, "step": 683550 }, { "epoch": 4.121995634399006, "grad_norm": 2.7189013957977295, "learning_rate": 0.00010915223078117916, "loss": 4.4157, "step": 683600 }, { "epoch": 4.1222971261803405, "grad_norm": 1.8516401052474976, "learning_rate": 0.0001091326512365502, "loss": 4.786, "step": 683650 }, { "epoch": 4.122598617961675, "grad_norm": 2.4630186557769775, "learning_rate": 0.00010911307244402643, "loss": 4.8835, "step": 683700 }, { "epoch": 4.122900109743008, "grad_norm": 2.278402805328369, "learning_rate": 0.00010909349440396819, "loss": 4.764, "step": 683750 }, { "epoch": 4.123201601524342, "grad_norm": 2.9688100814819336, "learning_rate": 0.00010907391711673587, "loss": 4.706, "step": 683800 }, { "epoch": 4.123503093305676, "grad_norm": 2.559514045715332, "learning_rate": 0.00010905434058268964, "loss": 5.0255, "step": 683850 }, { "epoch": 4.123804585087011, "grad_norm": 2.364409923553467, "learning_rate": 0.00010903476480218988, "loss": 4.6316, "step": 683900 }, { "epoch": 4.124106076868345, "grad_norm": 1.8369014263153076, "learning_rate": 0.00010901518977559681, "loss": 4.3214, "step": 683950 }, { "epoch": 4.124407568649678, "grad_norm": 0.6542109251022339, "learning_rate": 0.00010899561550327066, "loss": 4.6433, "step": 684000 }, { "epoch": 4.124407568649678, "eval_loss": 5.199305057525635, "eval_runtime": 39.0028, "eval_samples_per_second": 13.127, "eval_steps_per_second": 6.564, "eval_tts_loss": 7.643638568451873, "step": 684000 }, { "epoch": 4.124709060431012, "grad_norm": 3.180264711380005, "learning_rate": 0.00010897604198557167, "loss": 4.7332, "step": 684050 }, { "epoch": 4.125010552212347, "grad_norm": 2.0310330390930176, "learning_rate": 0.00010895646922286016, "loss": 4.6569, "step": 684100 }, { "epoch": 4.125312043993681, "grad_norm": 2.51143479347229, "learning_rate": 0.00010893689721549622, "loss": 5.0752, "step": 684150 }, { "epoch": 4.125613535775015, "grad_norm": 2.3370730876922607, "learning_rate": 0.00010891732596384005, "loss": 4.5763, "step": 684200 }, { "epoch": 4.125915027556349, "grad_norm": 0.7779551148414612, "learning_rate": 0.00010889775546825192, "loss": 5.1331, "step": 684250 }, { "epoch": 4.1262165193376825, "grad_norm": 2.5934672355651855, "learning_rate": 0.00010887818572909195, "loss": 4.8522, "step": 684300 }, { "epoch": 4.126518011119017, "grad_norm": 3.549787759780884, "learning_rate": 0.00010885861674672022, "loss": 4.4722, "step": 684350 }, { "epoch": 4.126819502900351, "grad_norm": 2.0223195552825928, "learning_rate": 0.00010883904852149702, "loss": 4.5214, "step": 684400 }, { "epoch": 4.127120994681685, "grad_norm": 3.381974935531616, "learning_rate": 0.00010881948105378233, "loss": 4.6717, "step": 684450 }, { "epoch": 4.127422486463019, "grad_norm": 2.875795602798462, "learning_rate": 0.00010879991434393634, "loss": 4.9713, "step": 684500 }, { "epoch": 4.127723978244353, "grad_norm": 1.1316492557525635, "learning_rate": 0.00010878034839231918, "loss": 4.779, "step": 684550 }, { "epoch": 4.128025470025687, "grad_norm": 0.8576596975326538, "learning_rate": 0.0001087607831992908, "loss": 4.3841, "step": 684600 }, { "epoch": 4.128326961807021, "grad_norm": 2.2817494869232178, "learning_rate": 0.00010874121876521137, "loss": 4.6511, "step": 684650 }, { "epoch": 4.128628453588355, "grad_norm": 2.079725503921509, "learning_rate": 0.00010872165509044099, "loss": 5.2505, "step": 684700 }, { "epoch": 4.1289299453696895, "grad_norm": 2.3977701663970947, "learning_rate": 0.00010870209217533962, "loss": 4.853, "step": 684750 }, { "epoch": 4.129231437151024, "grad_norm": 3.302297353744507, "learning_rate": 0.00010868253002026725, "loss": 4.6679, "step": 684800 }, { "epoch": 4.129532928932357, "grad_norm": 3.168055534362793, "learning_rate": 0.00010866296862558404, "loss": 4.717, "step": 684850 }, { "epoch": 4.129834420713691, "grad_norm": 1.0844883918762207, "learning_rate": 0.00010864340799164985, "loss": 4.4134, "step": 684900 }, { "epoch": 4.130135912495025, "grad_norm": 2.082848072052002, "learning_rate": 0.0001086238481188247, "loss": 4.8456, "step": 684950 }, { "epoch": 4.13043740427636, "grad_norm": 2.7862982749938965, "learning_rate": 0.00010860428900746864, "loss": 4.5251, "step": 685000 }, { "epoch": 4.130738896057694, "grad_norm": 2.6946566104888916, "learning_rate": 0.0001085847306579415, "loss": 4.5308, "step": 685050 }, { "epoch": 4.131040387839027, "grad_norm": 1.4247573614120483, "learning_rate": 0.0001085651730706033, "loss": 4.7873, "step": 685100 }, { "epoch": 4.131341879620361, "grad_norm": 3.202932119369507, "learning_rate": 0.00010854561624581402, "loss": 4.84, "step": 685150 }, { "epoch": 4.131643371401696, "grad_norm": 2.0767922401428223, "learning_rate": 0.00010852606018393344, "loss": 4.5672, "step": 685200 }, { "epoch": 4.13194486318303, "grad_norm": 1.1935964822769165, "learning_rate": 0.00010850650488532153, "loss": 4.6877, "step": 685250 }, { "epoch": 4.132246354964364, "grad_norm": 2.3036749362945557, "learning_rate": 0.00010848695035033824, "loss": 4.8037, "step": 685300 }, { "epoch": 4.132547846745697, "grad_norm": 2.1828672885894775, "learning_rate": 0.0001084673965793434, "loss": 4.4, "step": 685350 }, { "epoch": 4.1328493385270315, "grad_norm": 2.5609538555145264, "learning_rate": 0.00010844784357269678, "loss": 4.8305, "step": 685400 }, { "epoch": 4.133150830308366, "grad_norm": 1.4813770055770874, "learning_rate": 0.00010842829133075834, "loss": 4.4546, "step": 685450 }, { "epoch": 4.1334523220897, "grad_norm": 2.173448324203491, "learning_rate": 0.00010840873985388792, "loss": 4.6704, "step": 685500 }, { "epoch": 4.133753813871034, "grad_norm": 2.047781467437744, "learning_rate": 0.0001083891891424452, "loss": 4.6514, "step": 685550 }, { "epoch": 4.134055305652368, "grad_norm": 2.168635368347168, "learning_rate": 0.00010836963919679011, "loss": 5.0896, "step": 685600 }, { "epoch": 4.134356797433702, "grad_norm": 2.10638165473938, "learning_rate": 0.00010835009001728244, "loss": 4.9674, "step": 685650 }, { "epoch": 4.134658289215036, "grad_norm": 0.38515663146972656, "learning_rate": 0.00010833054160428189, "loss": 4.4541, "step": 685700 }, { "epoch": 4.13495978099637, "grad_norm": 2.2541778087615967, "learning_rate": 0.00010831099395814823, "loss": 4.4987, "step": 685750 }, { "epoch": 4.135261272777704, "grad_norm": 2.680631637573242, "learning_rate": 0.00010829144707924129, "loss": 4.7379, "step": 685800 }, { "epoch": 4.1355627645590385, "grad_norm": 5.808477878570557, "learning_rate": 0.0001082719009679207, "loss": 4.6264, "step": 685850 }, { "epoch": 4.135864256340372, "grad_norm": 1.6789817810058594, "learning_rate": 0.00010825235562454623, "loss": 4.4508, "step": 685900 }, { "epoch": 4.136165748121706, "grad_norm": 3.214477062225342, "learning_rate": 0.00010823281104947762, "loss": 4.4653, "step": 685950 }, { "epoch": 4.13646723990304, "grad_norm": 3.330855369567871, "learning_rate": 0.00010821326724307445, "loss": 4.4649, "step": 686000 }, { "epoch": 4.136768731684374, "grad_norm": 1.9882938861846924, "learning_rate": 0.00010819372420569651, "loss": 4.7548, "step": 686050 }, { "epoch": 4.137070223465709, "grad_norm": 2.479346990585327, "learning_rate": 0.00010817418193770345, "loss": 4.3362, "step": 686100 }, { "epoch": 4.137371715247042, "grad_norm": 2.7720632553100586, "learning_rate": 0.0001081546404394548, "loss": 4.7925, "step": 686150 }, { "epoch": 4.137673207028376, "grad_norm": 2.075752019882202, "learning_rate": 0.00010813509971131027, "loss": 4.9937, "step": 686200 }, { "epoch": 4.13797469880971, "grad_norm": 3.066934585571289, "learning_rate": 0.00010811555975362959, "loss": 4.6703, "step": 686250 }, { "epoch": 4.138276190591045, "grad_norm": 5.195237159729004, "learning_rate": 0.0001080960205667722, "loss": 4.5457, "step": 686300 }, { "epoch": 4.138577682372379, "grad_norm": 1.2115230560302734, "learning_rate": 0.00010807648215109772, "loss": 4.3281, "step": 686350 }, { "epoch": 4.138879174153713, "grad_norm": 5.514028549194336, "learning_rate": 0.00010805694450696583, "loss": 4.4893, "step": 686400 }, { "epoch": 4.139180665935046, "grad_norm": 2.368969202041626, "learning_rate": 0.00010803740763473593, "loss": 5.0179, "step": 686450 }, { "epoch": 4.1394821577163805, "grad_norm": 3.2352993488311768, "learning_rate": 0.00010801787153476771, "loss": 5.0583, "step": 686500 }, { "epoch": 4.139783649497715, "grad_norm": 1.9745943546295166, "learning_rate": 0.00010799833620742067, "loss": 4.5082, "step": 686550 }, { "epoch": 4.140085141279049, "grad_norm": 2.6316001415252686, "learning_rate": 0.00010797880165305428, "loss": 4.5645, "step": 686600 }, { "epoch": 4.140386633060383, "grad_norm": 2.003720760345459, "learning_rate": 0.0001079592678720281, "loss": 4.6081, "step": 686650 }, { "epoch": 4.1406881248417164, "grad_norm": 1.8493223190307617, "learning_rate": 0.00010793973486470161, "loss": 4.4859, "step": 686700 }, { "epoch": 4.140989616623051, "grad_norm": 2.7855405807495117, "learning_rate": 0.00010792020263143423, "loss": 4.7175, "step": 686750 }, { "epoch": 4.141291108404385, "grad_norm": 3.0694618225097656, "learning_rate": 0.00010790067117258548, "loss": 4.8416, "step": 686800 }, { "epoch": 4.141592600185719, "grad_norm": 2.6146657466888428, "learning_rate": 0.00010788114048851484, "loss": 4.885, "step": 686850 }, { "epoch": 4.141894091967053, "grad_norm": 2.56815242767334, "learning_rate": 0.00010786161057958168, "loss": 5.1061, "step": 686900 }, { "epoch": 4.142195583748387, "grad_norm": 2.6529734134674072, "learning_rate": 0.00010784208144614542, "loss": 4.5231, "step": 686950 }, { "epoch": 4.142497075529721, "grad_norm": 2.563652992248535, "learning_rate": 0.00010782255308856553, "loss": 4.542, "step": 687000 }, { "epoch": 4.142497075529721, "eval_loss": 5.2027997970581055, "eval_runtime": 39.2505, "eval_samples_per_second": 13.044, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.617372649123891, "step": 687000 }, { "epoch": 4.142798567311055, "grad_norm": 1.999078392982483, "learning_rate": 0.00010780302550720132, "loss": 4.7767, "step": 687050 }, { "epoch": 4.143100059092389, "grad_norm": 2.167365550994873, "learning_rate": 0.00010778349870241223, "loss": 4.8018, "step": 687100 }, { "epoch": 4.143401550873723, "grad_norm": 3.06697940826416, "learning_rate": 0.00010776397267455763, "loss": 4.1101, "step": 687150 }, { "epoch": 4.143703042655058, "grad_norm": 2.4749860763549805, "learning_rate": 0.00010774444742399675, "loss": 5.003, "step": 687200 }, { "epoch": 4.144004534436391, "grad_norm": 3.0032920837402344, "learning_rate": 0.00010772492295108908, "loss": 4.9725, "step": 687250 }, { "epoch": 4.144306026217725, "grad_norm": 2.898637056350708, "learning_rate": 0.00010770539925619389, "loss": 4.8509, "step": 687300 }, { "epoch": 4.144607517999059, "grad_norm": 3.6455445289611816, "learning_rate": 0.00010768587633967042, "loss": 4.9185, "step": 687350 }, { "epoch": 4.144909009780394, "grad_norm": 3.4613912105560303, "learning_rate": 0.00010766635420187799, "loss": 4.9502, "step": 687400 }, { "epoch": 4.145210501561728, "grad_norm": 3.654153823852539, "learning_rate": 0.000107646832843176, "loss": 4.8488, "step": 687450 }, { "epoch": 4.145511993343061, "grad_norm": 2.573394298553467, "learning_rate": 0.00010762731226392353, "loss": 4.8583, "step": 687500 }, { "epoch": 4.145813485124395, "grad_norm": 2.2537944316864014, "learning_rate": 0.0001076077924644799, "loss": 4.5315, "step": 687550 }, { "epoch": 4.1461149769057295, "grad_norm": 2.0970053672790527, "learning_rate": 0.00010758827344520441, "loss": 4.5423, "step": 687600 }, { "epoch": 4.146416468687064, "grad_norm": 2.4543745517730713, "learning_rate": 0.00010756875520645615, "loss": 4.9029, "step": 687650 }, { "epoch": 4.146717960468398, "grad_norm": 2.4273478984832764, "learning_rate": 0.00010754923774859446, "loss": 4.8805, "step": 687700 }, { "epoch": 4.147019452249731, "grad_norm": 0.9765130877494812, "learning_rate": 0.00010752972107197849, "loss": 5.0847, "step": 687750 }, { "epoch": 4.147320944031065, "grad_norm": 2.15936017036438, "learning_rate": 0.00010751020517696733, "loss": 4.4044, "step": 687800 }, { "epoch": 4.1476224358124, "grad_norm": 2.3559672832489014, "learning_rate": 0.00010749069006392022, "loss": 4.891, "step": 687850 }, { "epoch": 4.147923927593734, "grad_norm": 2.263835906982422, "learning_rate": 0.00010747117573319631, "loss": 4.6932, "step": 687900 }, { "epoch": 4.148225419375068, "grad_norm": 1.7620970010757446, "learning_rate": 0.00010745166218515472, "loss": 4.6268, "step": 687950 }, { "epoch": 4.148526911156402, "grad_norm": 2.699164867401123, "learning_rate": 0.00010743214942015452, "loss": 4.9565, "step": 688000 }, { "epoch": 4.148828402937736, "grad_norm": 2.357858419418335, "learning_rate": 0.00010741263743855495, "loss": 5.0392, "step": 688050 }, { "epoch": 4.14912989471907, "grad_norm": 1.623962163925171, "learning_rate": 0.00010739312624071498, "loss": 4.9427, "step": 688100 }, { "epoch": 4.149431386500404, "grad_norm": 2.695910930633545, "learning_rate": 0.00010737361582699368, "loss": 4.4432, "step": 688150 }, { "epoch": 4.149732878281738, "grad_norm": 2.1297686100006104, "learning_rate": 0.0001073541061977502, "loss": 4.693, "step": 688200 }, { "epoch": 4.150034370063072, "grad_norm": 1.1771509647369385, "learning_rate": 0.00010733459735334348, "loss": 4.6877, "step": 688250 }, { "epoch": 4.150335861844406, "grad_norm": 0.8205311298370361, "learning_rate": 0.00010731508929413265, "loss": 4.7723, "step": 688300 }, { "epoch": 4.15063735362574, "grad_norm": 2.6881072521209717, "learning_rate": 0.00010729558202047672, "loss": 5.1966, "step": 688350 }, { "epoch": 4.150938845407074, "grad_norm": 2.97109055519104, "learning_rate": 0.00010727607553273457, "loss": 4.718, "step": 688400 }, { "epoch": 4.151240337188408, "grad_norm": 2.4324471950531006, "learning_rate": 0.00010725656983126535, "loss": 4.903, "step": 688450 }, { "epoch": 4.1515418289697426, "grad_norm": 3.908446788787842, "learning_rate": 0.00010723706491642794, "loss": 4.5702, "step": 688500 }, { "epoch": 4.151843320751077, "grad_norm": 3.552675485610962, "learning_rate": 0.00010721756078858133, "loss": 4.7585, "step": 688550 }, { "epoch": 4.15214481253241, "grad_norm": 2.2997658252716064, "learning_rate": 0.00010719805744808442, "loss": 4.6092, "step": 688600 }, { "epoch": 4.152446304313744, "grad_norm": 3.0328314304351807, "learning_rate": 0.00010717855489529621, "loss": 4.807, "step": 688650 }, { "epoch": 4.1527477960950785, "grad_norm": 0.7858163118362427, "learning_rate": 0.00010715905313057564, "loss": 4.9083, "step": 688700 }, { "epoch": 4.153049287876413, "grad_norm": 4.083212375640869, "learning_rate": 0.00010713955215428148, "loss": 5.0724, "step": 688750 }, { "epoch": 4.153350779657747, "grad_norm": 1.5858268737792969, "learning_rate": 0.0001071200519667727, "loss": 4.4295, "step": 688800 }, { "epoch": 4.15365227143908, "grad_norm": 2.7806241512298584, "learning_rate": 0.00010710055256840825, "loss": 4.5922, "step": 688850 }, { "epoch": 4.153953763220414, "grad_norm": 2.2447915077209473, "learning_rate": 0.00010708105395954686, "loss": 5.0819, "step": 688900 }, { "epoch": 4.154255255001749, "grad_norm": 3.0279436111450195, "learning_rate": 0.00010706155614054739, "loss": 4.7587, "step": 688950 }, { "epoch": 4.154556746783083, "grad_norm": 3.0637152194976807, "learning_rate": 0.00010704205911176878, "loss": 4.6475, "step": 689000 }, { "epoch": 4.154858238564417, "grad_norm": 3.1151645183563232, "learning_rate": 0.00010702256287356973, "loss": 4.434, "step": 689050 }, { "epoch": 4.15515973034575, "grad_norm": 2.444507598876953, "learning_rate": 0.00010700306742630903, "loss": 4.4555, "step": 689100 }, { "epoch": 4.155461222127085, "grad_norm": 3.8865325450897217, "learning_rate": 0.0001069835727703456, "loss": 4.584, "step": 689150 }, { "epoch": 4.155762713908419, "grad_norm": 2.8212380409240723, "learning_rate": 0.00010696407890603806, "loss": 4.7043, "step": 689200 }, { "epoch": 4.156064205689753, "grad_norm": 2.2810003757476807, "learning_rate": 0.00010694458583374524, "loss": 4.7037, "step": 689250 }, { "epoch": 4.156365697471087, "grad_norm": 2.360203981399536, "learning_rate": 0.00010692509355382594, "loss": 4.8058, "step": 689300 }, { "epoch": 4.156667189252421, "grad_norm": 3.0835390090942383, "learning_rate": 0.00010690560206663875, "loss": 4.8829, "step": 689350 }, { "epoch": 4.156968681033755, "grad_norm": 1.5231467485427856, "learning_rate": 0.00010688611137254243, "loss": 5.0521, "step": 689400 }, { "epoch": 4.157270172815089, "grad_norm": 0.8470848202705383, "learning_rate": 0.0001068666214718958, "loss": 4.3488, "step": 689450 }, { "epoch": 4.157571664596423, "grad_norm": 1.5259891748428345, "learning_rate": 0.0001068471323650574, "loss": 4.2115, "step": 689500 }, { "epoch": 4.157873156377757, "grad_norm": 3.1363210678100586, "learning_rate": 0.00010682764405238593, "loss": 4.6634, "step": 689550 }, { "epoch": 4.1581746481590915, "grad_norm": 3.8235368728637695, "learning_rate": 0.00010680815653424009, "loss": 4.8203, "step": 689600 }, { "epoch": 4.158476139940425, "grad_norm": 2.6873509883880615, "learning_rate": 0.00010678866981097846, "loss": 4.7464, "step": 689650 }, { "epoch": 4.158777631721759, "grad_norm": 2.4626059532165527, "learning_rate": 0.0001067691838829597, "loss": 4.701, "step": 689700 }, { "epoch": 4.159079123503093, "grad_norm": 3.0233047008514404, "learning_rate": 0.00010674969875054246, "loss": 4.8752, "step": 689750 }, { "epoch": 4.1593806152844275, "grad_norm": 2.3416953086853027, "learning_rate": 0.00010673021441408524, "loss": 4.7594, "step": 689800 }, { "epoch": 4.159682107065762, "grad_norm": 3.4878005981445312, "learning_rate": 0.00010671073087394669, "loss": 4.7521, "step": 689850 }, { "epoch": 4.159983598847095, "grad_norm": 1.8264753818511963, "learning_rate": 0.0001066912481304854, "loss": 4.5544, "step": 689900 }, { "epoch": 4.160285090628429, "grad_norm": 4.206552028656006, "learning_rate": 0.00010667176618405984, "loss": 4.4587, "step": 689950 }, { "epoch": 4.160586582409763, "grad_norm": 2.4273672103881836, "learning_rate": 0.00010665228503502856, "loss": 4.96, "step": 690000 }, { "epoch": 4.160586582409763, "eval_loss": 5.205038070678711, "eval_runtime": 39.1526, "eval_samples_per_second": 13.077, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.566505348520309, "step": 690000 }, { "epoch": 4.160888074191098, "grad_norm": 1.8055083751678467, "learning_rate": 0.00010663280468375018, "loss": 5.1712, "step": 690050 }, { "epoch": 4.161189565972432, "grad_norm": 2.6146128177642822, "learning_rate": 0.00010661332513058312, "loss": 4.6741, "step": 690100 }, { "epoch": 4.161491057753766, "grad_norm": 2.84663724899292, "learning_rate": 0.00010659384637588587, "loss": 4.886, "step": 690150 }, { "epoch": 4.161792549535099, "grad_norm": 3.1197192668914795, "learning_rate": 0.000106574368420017, "loss": 4.8479, "step": 690200 }, { "epoch": 4.162094041316434, "grad_norm": 3.1501781940460205, "learning_rate": 0.00010655489126333482, "loss": 4.8578, "step": 690250 }, { "epoch": 4.162395533097768, "grad_norm": 2.1674046516418457, "learning_rate": 0.00010653541490619791, "loss": 4.7704, "step": 690300 }, { "epoch": 4.162697024879102, "grad_norm": 1.741330623626709, "learning_rate": 0.0001065159393489647, "loss": 4.8192, "step": 690350 }, { "epoch": 4.162998516660436, "grad_norm": 3.3975026607513428, "learning_rate": 0.0001064964645919935, "loss": 4.5611, "step": 690400 }, { "epoch": 4.1633000084417695, "grad_norm": 2.0785610675811768, "learning_rate": 0.00010647699063564281, "loss": 4.4613, "step": 690450 }, { "epoch": 4.163601500223104, "grad_norm": 1.375929832458496, "learning_rate": 0.00010645751748027105, "loss": 4.7907, "step": 690500 }, { "epoch": 4.163902992004438, "grad_norm": 2.0214009284973145, "learning_rate": 0.00010643804512623646, "loss": 4.6366, "step": 690550 }, { "epoch": 4.164204483785772, "grad_norm": 1.9132503271102905, "learning_rate": 0.0001064185735738975, "loss": 5.0879, "step": 690600 }, { "epoch": 4.164505975567106, "grad_norm": 2.13230037689209, "learning_rate": 0.00010639910282361255, "loss": 4.8682, "step": 690650 }, { "epoch": 4.1648074673484405, "grad_norm": 2.8488247394561768, "learning_rate": 0.00010637963287573987, "loss": 4.1826, "step": 690700 }, { "epoch": 4.165108959129774, "grad_norm": 2.367830514907837, "learning_rate": 0.00010636016373063775, "loss": 4.2301, "step": 690750 }, { "epoch": 4.165410450911108, "grad_norm": 2.906254291534424, "learning_rate": 0.00010634069538866462, "loss": 4.3823, "step": 690800 }, { "epoch": 4.165711942692442, "grad_norm": 2.8099963665008545, "learning_rate": 0.0001063212278501786, "loss": 4.9951, "step": 690850 }, { "epoch": 4.1660134344737765, "grad_norm": 2.062225818634033, "learning_rate": 0.0001063017611155381, "loss": 4.9577, "step": 690900 }, { "epoch": 4.166314926255111, "grad_norm": 2.9601082801818848, "learning_rate": 0.00010628229518510135, "loss": 4.3797, "step": 690950 }, { "epoch": 4.166616418036444, "grad_norm": 2.830410957336426, "learning_rate": 0.0001062628300592265, "loss": 4.7242, "step": 691000 }, { "epoch": 4.166917909817778, "grad_norm": 2.324172019958496, "learning_rate": 0.0001062433657382719, "loss": 4.7513, "step": 691050 }, { "epoch": 4.167219401599112, "grad_norm": 2.565655469894409, "learning_rate": 0.00010622390222259572, "loss": 4.6356, "step": 691100 }, { "epoch": 4.167520893380447, "grad_norm": 2.6727070808410645, "learning_rate": 0.0001062044395125561, "loss": 4.3408, "step": 691150 }, { "epoch": 4.167822385161781, "grad_norm": 3.504379987716675, "learning_rate": 0.00010618497760851126, "loss": 5.0748, "step": 691200 }, { "epoch": 4.168123876943114, "grad_norm": 2.497933864593506, "learning_rate": 0.00010616551651081944, "loss": 4.7965, "step": 691250 }, { "epoch": 4.168425368724448, "grad_norm": 2.5605692863464355, "learning_rate": 0.00010614605621983872, "loss": 4.7416, "step": 691300 }, { "epoch": 4.168726860505783, "grad_norm": 2.424612045288086, "learning_rate": 0.00010612659673592719, "loss": 4.9306, "step": 691350 }, { "epoch": 4.169028352287117, "grad_norm": 2.4199166297912598, "learning_rate": 0.00010610713805944312, "loss": 4.5903, "step": 691400 }, { "epoch": 4.169329844068451, "grad_norm": 2.137728214263916, "learning_rate": 0.00010608768019074447, "loss": 4.2109, "step": 691450 }, { "epoch": 4.169631335849784, "grad_norm": 3.611952066421509, "learning_rate": 0.00010606822313018943, "loss": 4.8873, "step": 691500 }, { "epoch": 4.1699328276311185, "grad_norm": 2.736083507537842, "learning_rate": 0.00010604876687813608, "loss": 4.7779, "step": 691550 }, { "epoch": 4.170234319412453, "grad_norm": 2.0741305351257324, "learning_rate": 0.0001060293114349424, "loss": 4.8859, "step": 691600 }, { "epoch": 4.170535811193787, "grad_norm": 3.366163730621338, "learning_rate": 0.00010600985680096651, "loss": 4.9602, "step": 691650 }, { "epoch": 4.170837302975121, "grad_norm": 2.822171449661255, "learning_rate": 0.00010599040297656646, "loss": 4.8711, "step": 691700 }, { "epoch": 4.171138794756455, "grad_norm": 3.1501665115356445, "learning_rate": 0.0001059709499621002, "loss": 4.6242, "step": 691750 }, { "epoch": 4.171440286537789, "grad_norm": 2.0132813453674316, "learning_rate": 0.00010595149775792574, "loss": 4.5863, "step": 691800 }, { "epoch": 4.171741778319123, "grad_norm": 3.508176326751709, "learning_rate": 0.00010593204636440118, "loss": 4.9727, "step": 691850 }, { "epoch": 4.172043270100457, "grad_norm": 3.507561206817627, "learning_rate": 0.0001059125957818844, "loss": 4.9679, "step": 691900 }, { "epoch": 4.172344761881791, "grad_norm": 3.137052536010742, "learning_rate": 0.00010589314601073333, "loss": 4.9296, "step": 691950 }, { "epoch": 4.1726462536631255, "grad_norm": 1.874160885810852, "learning_rate": 0.00010587369705130603, "loss": 4.8451, "step": 692000 }, { "epoch": 4.172947745444459, "grad_norm": 2.991257905960083, "learning_rate": 0.00010585424890396029, "loss": 5.0693, "step": 692050 }, { "epoch": 4.173249237225793, "grad_norm": 3.033524751663208, "learning_rate": 0.00010583480156905415, "loss": 5.1438, "step": 692100 }, { "epoch": 4.173550729007127, "grad_norm": 1.4108352661132812, "learning_rate": 0.00010581535504694543, "loss": 4.8725, "step": 692150 }, { "epoch": 4.173852220788461, "grad_norm": 2.4309144020080566, "learning_rate": 0.0001057959093379921, "loss": 4.836, "step": 692200 }, { "epoch": 4.174153712569796, "grad_norm": 2.626051664352417, "learning_rate": 0.00010577646444255196, "loss": 4.6946, "step": 692250 }, { "epoch": 4.17445520435113, "grad_norm": 3.0423760414123535, "learning_rate": 0.00010575702036098281, "loss": 4.9129, "step": 692300 }, { "epoch": 4.174756696132463, "grad_norm": 1.516147494316101, "learning_rate": 0.00010573757709364266, "loss": 5.3011, "step": 692350 }, { "epoch": 4.175058187913797, "grad_norm": 5.1726813316345215, "learning_rate": 0.00010571813464088916, "loss": 4.7161, "step": 692400 }, { "epoch": 4.1753596796951316, "grad_norm": 2.5356550216674805, "learning_rate": 0.00010569869300308024, "loss": 4.9411, "step": 692450 }, { "epoch": 4.175661171476466, "grad_norm": 2.544344425201416, "learning_rate": 0.0001056792521805737, "loss": 5.2431, "step": 692500 }, { "epoch": 4.1759626632578, "grad_norm": 2.651599884033203, "learning_rate": 0.00010565981217372718, "loss": 4.8264, "step": 692550 }, { "epoch": 4.176264155039133, "grad_norm": 2.3862009048461914, "learning_rate": 0.00010564037298289855, "loss": 4.388, "step": 692600 }, { "epoch": 4.1765656468204675, "grad_norm": 2.233193874359131, "learning_rate": 0.00010562093460844562, "loss": 4.4218, "step": 692650 }, { "epoch": 4.176867138601802, "grad_norm": 2.442150592803955, "learning_rate": 0.00010560149705072603, "loss": 5.0105, "step": 692700 }, { "epoch": 4.177168630383136, "grad_norm": 1.5084092617034912, "learning_rate": 0.00010558206031009749, "loss": 4.5019, "step": 692750 }, { "epoch": 4.17747012216447, "grad_norm": 3.014599084854126, "learning_rate": 0.00010556262438691778, "loss": 4.6759, "step": 692800 }, { "epoch": 4.177771613945803, "grad_norm": 2.4104275703430176, "learning_rate": 0.00010554318928154456, "loss": 4.9822, "step": 692850 }, { "epoch": 4.178073105727138, "grad_norm": 2.5390288829803467, "learning_rate": 0.00010552375499433543, "loss": 4.3069, "step": 692900 }, { "epoch": 4.178374597508472, "grad_norm": 1.1503489017486572, "learning_rate": 0.00010550432152564821, "loss": 4.3314, "step": 692950 }, { "epoch": 4.178676089289806, "grad_norm": 2.5787062644958496, "learning_rate": 0.00010548488887584039, "loss": 4.5308, "step": 693000 }, { "epoch": 4.178676089289806, "eval_loss": 5.193739891052246, "eval_runtime": 38.9286, "eval_samples_per_second": 13.152, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.636348424807629, "step": 693000 }, { "epoch": 4.17897758107114, "grad_norm": 2.818854331970215, "learning_rate": 0.00010546545704526967, "loss": 4.6853, "step": 693050 }, { "epoch": 4.1792790728524745, "grad_norm": 2.067502975463867, "learning_rate": 0.00010544602603429372, "loss": 4.5293, "step": 693100 }, { "epoch": 4.179580564633808, "grad_norm": 2.442164421081543, "learning_rate": 0.00010542659584327, "loss": 4.7399, "step": 693150 }, { "epoch": 4.179882056415142, "grad_norm": 2.702706813812256, "learning_rate": 0.00010540716647255616, "loss": 4.6939, "step": 693200 }, { "epoch": 4.180183548196476, "grad_norm": 2.5149574279785156, "learning_rate": 0.00010538773792250987, "loss": 4.7651, "step": 693250 }, { "epoch": 4.18048503997781, "grad_norm": 2.5629513263702393, "learning_rate": 0.00010536831019348856, "loss": 4.875, "step": 693300 }, { "epoch": 4.180786531759145, "grad_norm": 2.0829684734344482, "learning_rate": 0.00010534888328584977, "loss": 4.7517, "step": 693350 }, { "epoch": 4.181088023540478, "grad_norm": 0.8616994619369507, "learning_rate": 0.00010532945719995114, "loss": 4.7003, "step": 693400 }, { "epoch": 4.181389515321812, "grad_norm": 2.149758815765381, "learning_rate": 0.00010531003193615007, "loss": 4.0789, "step": 693450 }, { "epoch": 4.181691007103146, "grad_norm": 1.5314323902130127, "learning_rate": 0.00010529060749480404, "loss": 4.5161, "step": 693500 }, { "epoch": 4.1819924988844805, "grad_norm": 2.744189977645874, "learning_rate": 0.00010527118387627064, "loss": 4.8563, "step": 693550 }, { "epoch": 4.182293990665815, "grad_norm": 0.847869336605072, "learning_rate": 0.00010525176108090723, "loss": 4.7471, "step": 693600 }, { "epoch": 4.182595482447148, "grad_norm": 4.276966094970703, "learning_rate": 0.00010523233910907129, "loss": 4.7977, "step": 693650 }, { "epoch": 4.182896974228482, "grad_norm": 2.6493043899536133, "learning_rate": 0.00010521291796112032, "loss": 4.7401, "step": 693700 }, { "epoch": 4.1831984660098165, "grad_norm": 4.314540863037109, "learning_rate": 0.0001051934976374116, "loss": 5.0121, "step": 693750 }, { "epoch": 4.183499957791151, "grad_norm": 2.1625618934631348, "learning_rate": 0.00010517407813830264, "loss": 4.7886, "step": 693800 }, { "epoch": 4.183801449572485, "grad_norm": 3.282085418701172, "learning_rate": 0.00010515465946415085, "loss": 4.5022, "step": 693850 }, { "epoch": 4.184102941353819, "grad_norm": 2.807111978530884, "learning_rate": 0.00010513524161531352, "loss": 4.8865, "step": 693900 }, { "epoch": 4.184404433135152, "grad_norm": 2.931171417236328, "learning_rate": 0.00010511582459214802, "loss": 4.6503, "step": 693950 }, { "epoch": 4.184705924916487, "grad_norm": 3.9759421348571777, "learning_rate": 0.00010509640839501178, "loss": 4.3436, "step": 694000 }, { "epoch": 4.185007416697821, "grad_norm": 2.231624126434326, "learning_rate": 0.00010507699302426202, "loss": 4.9681, "step": 694050 }, { "epoch": 4.185308908479155, "grad_norm": 3.4949276447296143, "learning_rate": 0.00010505757848025605, "loss": 4.6937, "step": 694100 }, { "epoch": 4.185610400260489, "grad_norm": 2.997511625289917, "learning_rate": 0.0001050381647633513, "loss": 4.7264, "step": 694150 }, { "epoch": 4.185911892041823, "grad_norm": 2.411100149154663, "learning_rate": 0.00010501875187390491, "loss": 4.3708, "step": 694200 }, { "epoch": 4.186213383823157, "grad_norm": 2.8340559005737305, "learning_rate": 0.0001049993398122742, "loss": 4.6922, "step": 694250 }, { "epoch": 4.186514875604491, "grad_norm": 3.3142592906951904, "learning_rate": 0.00010497992857881647, "loss": 4.824, "step": 694300 }, { "epoch": 4.186816367385825, "grad_norm": 2.203728675842285, "learning_rate": 0.00010496051817388886, "loss": 4.3342, "step": 694350 }, { "epoch": 4.187117859167159, "grad_norm": 2.8520524501800537, "learning_rate": 0.0001049411085978486, "loss": 5.1136, "step": 694400 }, { "epoch": 4.187419350948494, "grad_norm": 2.5789499282836914, "learning_rate": 0.00010492169985105301, "loss": 4.4458, "step": 694450 }, { "epoch": 4.187720842729827, "grad_norm": 3.294491767883301, "learning_rate": 0.0001049022919338592, "loss": 4.571, "step": 694500 }, { "epoch": 4.188022334511161, "grad_norm": 1.4093421697616577, "learning_rate": 0.00010488288484662428, "loss": 4.9268, "step": 694550 }, { "epoch": 4.188323826292495, "grad_norm": 2.7615864276885986, "learning_rate": 0.00010486347858970557, "loss": 4.7954, "step": 694600 }, { "epoch": 4.1886253180738295, "grad_norm": 1.9463021755218506, "learning_rate": 0.00010484407316346005, "loss": 4.9433, "step": 694650 }, { "epoch": 4.188926809855164, "grad_norm": 2.0078837871551514, "learning_rate": 0.00010482466856824494, "loss": 4.8165, "step": 694700 }, { "epoch": 4.189228301636497, "grad_norm": 1.8538206815719604, "learning_rate": 0.0001048052648044174, "loss": 4.4713, "step": 694750 }, { "epoch": 4.189529793417831, "grad_norm": 0.6238895058631897, "learning_rate": 0.00010478586187233437, "loss": 4.7122, "step": 694800 }, { "epoch": 4.1898312851991655, "grad_norm": 2.6366560459136963, "learning_rate": 0.00010476645977235308, "loss": 4.7794, "step": 694850 }, { "epoch": 4.1901327769805, "grad_norm": 2.861065149307251, "learning_rate": 0.00010474705850483057, "loss": 4.7317, "step": 694900 }, { "epoch": 4.190434268761834, "grad_norm": 3.1237618923187256, "learning_rate": 0.00010472765807012379, "loss": 4.4725, "step": 694950 }, { "epoch": 4.190735760543167, "grad_norm": 2.4062533378601074, "learning_rate": 0.00010470825846858989, "loss": 4.7696, "step": 695000 }, { "epoch": 4.191037252324501, "grad_norm": 3.234531879425049, "learning_rate": 0.00010468885970058589, "loss": 4.7352, "step": 695050 }, { "epoch": 4.191338744105836, "grad_norm": 1.4544733762741089, "learning_rate": 0.00010466946176646874, "loss": 4.7915, "step": 695100 }, { "epoch": 4.19164023588717, "grad_norm": 5.447114944458008, "learning_rate": 0.00010465006466659544, "loss": 4.6598, "step": 695150 }, { "epoch": 4.191941727668504, "grad_norm": 1.9638711214065552, "learning_rate": 0.00010463066840132305, "loss": 4.7749, "step": 695200 }, { "epoch": 4.192243219449837, "grad_norm": 2.096167802810669, "learning_rate": 0.00010461127297100836, "loss": 4.9947, "step": 695250 }, { "epoch": 4.192544711231172, "grad_norm": 2.243241786956787, "learning_rate": 0.0001045918783760085, "loss": 4.5413, "step": 695300 }, { "epoch": 4.192846203012506, "grad_norm": 2.3561813831329346, "learning_rate": 0.00010457248461668031, "loss": 4.7747, "step": 695350 }, { "epoch": 4.19314769479384, "grad_norm": 2.1519529819488525, "learning_rate": 0.00010455309169338067, "loss": 4.4661, "step": 695400 }, { "epoch": 4.193449186575174, "grad_norm": 2.664015293121338, "learning_rate": 0.00010453369960646655, "loss": 4.4645, "step": 695450 }, { "epoch": 4.193750678356508, "grad_norm": 3.113590717315674, "learning_rate": 0.00010451430835629476, "loss": 4.8944, "step": 695500 }, { "epoch": 4.194052170137842, "grad_norm": 1.9082924127578735, "learning_rate": 0.00010449491794322228, "loss": 4.4933, "step": 695550 }, { "epoch": 4.194353661919176, "grad_norm": 3.247037172317505, "learning_rate": 0.00010447552836760584, "loss": 5.1113, "step": 695600 }, { "epoch": 4.19465515370051, "grad_norm": 0.9739037752151489, "learning_rate": 0.00010445613962980237, "loss": 4.7922, "step": 695650 }, { "epoch": 4.194956645481844, "grad_norm": 2.346708059310913, "learning_rate": 0.00010443675173016866, "loss": 4.8101, "step": 695700 }, { "epoch": 4.1952581372631785, "grad_norm": 1.2074393033981323, "learning_rate": 0.00010441736466906148, "loss": 4.4825, "step": 695750 }, { "epoch": 4.195559629044512, "grad_norm": 4.729082107543945, "learning_rate": 0.00010439797844683765, "loss": 4.6746, "step": 695800 }, { "epoch": 4.195861120825846, "grad_norm": 2.598607301712036, "learning_rate": 0.000104378593063854, "loss": 4.766, "step": 695850 }, { "epoch": 4.19616261260718, "grad_norm": 3.437145233154297, "learning_rate": 0.0001043592085204672, "loss": 4.8191, "step": 695900 }, { "epoch": 4.1964641043885145, "grad_norm": 1.9896808862686157, "learning_rate": 0.00010433982481703403, "loss": 4.765, "step": 695950 }, { "epoch": 4.196765596169849, "grad_norm": 3.193338632583618, "learning_rate": 0.00010432044195391127, "loss": 4.5387, "step": 696000 }, { "epoch": 4.196765596169849, "eval_loss": 5.186921119689941, "eval_runtime": 39.2031, "eval_samples_per_second": 13.06, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.663002760781676, "step": 696000 }, { "epoch": 4.197067087951183, "grad_norm": 2.2313764095306396, "learning_rate": 0.00010430105993145557, "loss": 4.6475, "step": 696050 }, { "epoch": 4.197368579732516, "grad_norm": 2.925079345703125, "learning_rate": 0.00010428167875002359, "loss": 4.992, "step": 696100 }, { "epoch": 4.19767007151385, "grad_norm": 2.7854106426239014, "learning_rate": 0.00010426229840997215, "loss": 4.835, "step": 696150 }, { "epoch": 4.197971563295185, "grad_norm": 1.4167488813400269, "learning_rate": 0.00010424291891165777, "loss": 4.5474, "step": 696200 }, { "epoch": 4.198273055076519, "grad_norm": 2.3581349849700928, "learning_rate": 0.0001042235402554372, "loss": 4.5264, "step": 696250 }, { "epoch": 4.198574546857853, "grad_norm": 3.6917150020599365, "learning_rate": 0.00010420416244166709, "loss": 4.974, "step": 696300 }, { "epoch": 4.198876038639186, "grad_norm": 3.974783182144165, "learning_rate": 0.00010418478547070392, "loss": 4.8101, "step": 696350 }, { "epoch": 4.1991775304205206, "grad_norm": 2.702348232269287, "learning_rate": 0.00010416540934290441, "loss": 4.6079, "step": 696400 }, { "epoch": 4.199479022201855, "grad_norm": 2.595458984375, "learning_rate": 0.0001041460340586252, "loss": 5.0877, "step": 696450 }, { "epoch": 4.199780513983189, "grad_norm": 3.096022605895996, "learning_rate": 0.00010412665961822277, "loss": 4.4168, "step": 696500 }, { "epoch": 4.200082005764523, "grad_norm": 2.331066846847534, "learning_rate": 0.00010410728602205364, "loss": 4.7001, "step": 696550 }, { "epoch": 4.2003834975458565, "grad_norm": 3.0441207885742188, "learning_rate": 0.00010408791327047451, "loss": 4.4675, "step": 696600 }, { "epoch": 4.200684989327191, "grad_norm": 2.184816837310791, "learning_rate": 0.00010406854136384178, "loss": 4.9539, "step": 696650 }, { "epoch": 4.200986481108525, "grad_norm": 2.2186667919158936, "learning_rate": 0.00010404917030251195, "loss": 4.7913, "step": 696700 }, { "epoch": 4.201287972889859, "grad_norm": 1.4307870864868164, "learning_rate": 0.00010402980008684166, "loss": 4.3781, "step": 696750 }, { "epoch": 4.201589464671193, "grad_norm": 3.7533912658691406, "learning_rate": 0.0001040104307171872, "loss": 4.7099, "step": 696800 }, { "epoch": 4.2018909564525275, "grad_norm": 3.2498953342437744, "learning_rate": 0.00010399106219390519, "loss": 4.8481, "step": 696850 }, { "epoch": 4.202192448233861, "grad_norm": 1.9991402626037598, "learning_rate": 0.00010397169451735205, "loss": 4.8911, "step": 696900 }, { "epoch": 4.202493940015195, "grad_norm": 2.4940338134765625, "learning_rate": 0.00010395232768788411, "loss": 4.6112, "step": 696950 }, { "epoch": 4.202795431796529, "grad_norm": 2.2560641765594482, "learning_rate": 0.00010393296170585786, "loss": 4.8336, "step": 697000 }, { "epoch": 4.2030969235778635, "grad_norm": 1.6432448625564575, "learning_rate": 0.0001039135965716298, "loss": 4.6857, "step": 697050 }, { "epoch": 4.203398415359198, "grad_norm": 2.830517530441284, "learning_rate": 0.00010389423228555619, "loss": 4.7945, "step": 697100 }, { "epoch": 4.203699907140531, "grad_norm": 2.805828332901001, "learning_rate": 0.00010387486884799339, "loss": 4.6851, "step": 697150 }, { "epoch": 4.204001398921865, "grad_norm": 2.81343936920166, "learning_rate": 0.00010385550625929786, "loss": 4.5218, "step": 697200 }, { "epoch": 4.204302890703199, "grad_norm": 3.1603288650512695, "learning_rate": 0.00010383614451982588, "loss": 4.3498, "step": 697250 }, { "epoch": 4.204604382484534, "grad_norm": 2.7513904571533203, "learning_rate": 0.00010381678362993373, "loss": 4.507, "step": 697300 }, { "epoch": 4.204905874265868, "grad_norm": 5.106279373168945, "learning_rate": 0.00010379742358997785, "loss": 4.5237, "step": 697350 }, { "epoch": 4.205207366047201, "grad_norm": 2.660198211669922, "learning_rate": 0.00010377806440031438, "loss": 5.0444, "step": 697400 }, { "epoch": 4.205508857828535, "grad_norm": 0.6728161573410034, "learning_rate": 0.0001037587060612997, "loss": 4.5786, "step": 697450 }, { "epoch": 4.2058103496098695, "grad_norm": 1.5495222806930542, "learning_rate": 0.00010373934857329008, "loss": 4.8734, "step": 697500 }, { "epoch": 4.206111841391204, "grad_norm": 3.030790090560913, "learning_rate": 0.00010371999193664167, "loss": 4.755, "step": 697550 }, { "epoch": 4.206413333172538, "grad_norm": 5.359470844268799, "learning_rate": 0.00010370063615171078, "loss": 4.8041, "step": 697600 }, { "epoch": 4.206714824953872, "grad_norm": 3.994887590408325, "learning_rate": 0.00010368128121885364, "loss": 4.4604, "step": 697650 }, { "epoch": 4.2070163167352055, "grad_norm": 2.5006513595581055, "learning_rate": 0.0001036619271384264, "loss": 4.7236, "step": 697700 }, { "epoch": 4.20731780851654, "grad_norm": 3.662703037261963, "learning_rate": 0.00010364257391078524, "loss": 4.7647, "step": 697750 }, { "epoch": 4.207619300297874, "grad_norm": 1.1855446100234985, "learning_rate": 0.0001036232215362864, "loss": 4.7187, "step": 697800 }, { "epoch": 4.207920792079208, "grad_norm": 3.6502418518066406, "learning_rate": 0.00010360387001528597, "loss": 4.1519, "step": 697850 }, { "epoch": 4.208222283860542, "grad_norm": 2.1231868267059326, "learning_rate": 0.00010358451934814006, "loss": 4.7809, "step": 697900 }, { "epoch": 4.208523775641876, "grad_norm": 2.527127265930176, "learning_rate": 0.00010356516953520491, "loss": 4.7617, "step": 697950 }, { "epoch": 4.20882526742321, "grad_norm": 2.1049516201019287, "learning_rate": 0.00010354582057683647, "loss": 4.7181, "step": 698000 }, { "epoch": 4.209126759204544, "grad_norm": 3.5227248668670654, "learning_rate": 0.00010352647247339095, "loss": 4.6745, "step": 698050 }, { "epoch": 4.209428250985878, "grad_norm": 3.545248508453369, "learning_rate": 0.00010350712522522442, "loss": 5.0662, "step": 698100 }, { "epoch": 4.2097297427672125, "grad_norm": 2.4079856872558594, "learning_rate": 0.0001034877788326928, "loss": 4.5893, "step": 698150 }, { "epoch": 4.210031234548547, "grad_norm": 2.3773462772369385, "learning_rate": 0.00010346843329615226, "loss": 4.4132, "step": 698200 }, { "epoch": 4.21033272632988, "grad_norm": 3.352475166320801, "learning_rate": 0.00010344908861595888, "loss": 4.8613, "step": 698250 }, { "epoch": 4.210634218111214, "grad_norm": 3.8606812953948975, "learning_rate": 0.00010342974479246855, "loss": 4.5462, "step": 698300 }, { "epoch": 4.210935709892548, "grad_norm": 3.2988688945770264, "learning_rate": 0.00010341040182603728, "loss": 4.9741, "step": 698350 }, { "epoch": 4.211237201673883, "grad_norm": 3.085075855255127, "learning_rate": 0.00010339105971702112, "loss": 4.8482, "step": 698400 }, { "epoch": 4.211538693455217, "grad_norm": 2.950864315032959, "learning_rate": 0.00010337171846577598, "loss": 4.87, "step": 698450 }, { "epoch": 4.21184018523655, "grad_norm": 2.696718215942383, "learning_rate": 0.00010335237807265777, "loss": 4.9798, "step": 698500 }, { "epoch": 4.212141677017884, "grad_norm": 2.0920605659484863, "learning_rate": 0.00010333303853802257, "loss": 4.9301, "step": 698550 }, { "epoch": 4.2124431687992185, "grad_norm": 2.859076976776123, "learning_rate": 0.00010331369986222609, "loss": 4.8374, "step": 698600 }, { "epoch": 4.212744660580553, "grad_norm": 2.966167688369751, "learning_rate": 0.0001032943620456244, "loss": 4.5169, "step": 698650 }, { "epoch": 4.213046152361887, "grad_norm": 2.1684458255767822, "learning_rate": 0.00010327502508857334, "loss": 4.3189, "step": 698700 }, { "epoch": 4.21334764414322, "grad_norm": 2.223766803741455, "learning_rate": 0.00010325568899142872, "loss": 4.5081, "step": 698750 }, { "epoch": 4.2136491359245545, "grad_norm": 2.6200664043426514, "learning_rate": 0.00010323635375454641, "loss": 4.8831, "step": 698800 }, { "epoch": 4.213950627705889, "grad_norm": 2.132469415664673, "learning_rate": 0.00010321701937828233, "loss": 4.9425, "step": 698850 }, { "epoch": 4.214252119487223, "grad_norm": 2.760289192199707, "learning_rate": 0.00010319768586299226, "loss": 4.9763, "step": 698900 }, { "epoch": 4.214553611268557, "grad_norm": 3.6748979091644287, "learning_rate": 0.00010317835320903192, "loss": 4.968, "step": 698950 }, { "epoch": 4.21485510304989, "grad_norm": 2.6707913875579834, "learning_rate": 0.00010315902141675724, "loss": 5.0056, "step": 699000 }, { "epoch": 4.21485510304989, "eval_loss": 5.188821792602539, "eval_runtime": 38.9998, "eval_samples_per_second": 13.128, "eval_steps_per_second": 6.564, "eval_tts_loss": 7.562493812542797, "step": 699000 }, { "epoch": 4.215156594831225, "grad_norm": 0.9991772770881653, "learning_rate": 0.00010313969048652392, "loss": 4.1999, "step": 699050 }, { "epoch": 4.215458086612559, "grad_norm": 2.19378399848938, "learning_rate": 0.00010312036041868768, "loss": 4.6098, "step": 699100 }, { "epoch": 4.215759578393893, "grad_norm": 1.7783684730529785, "learning_rate": 0.00010310103121360431, "loss": 4.8731, "step": 699150 }, { "epoch": 4.216061070175227, "grad_norm": 1.1397582292556763, "learning_rate": 0.00010308170287162956, "loss": 4.4955, "step": 699200 }, { "epoch": 4.2163625619565614, "grad_norm": 1.3024550676345825, "learning_rate": 0.00010306237539311912, "loss": 4.299, "step": 699250 }, { "epoch": 4.216664053737895, "grad_norm": 2.3209569454193115, "learning_rate": 0.00010304304877842861, "loss": 4.959, "step": 699300 }, { "epoch": 4.216965545519229, "grad_norm": 2.5147483348846436, "learning_rate": 0.00010302372302791387, "loss": 4.996, "step": 699350 }, { "epoch": 4.217267037300563, "grad_norm": 3.758319854736328, "learning_rate": 0.0001030043981419304, "loss": 4.8315, "step": 699400 }, { "epoch": 4.217568529081897, "grad_norm": 2.45930814743042, "learning_rate": 0.00010298507412083395, "loss": 4.9896, "step": 699450 }, { "epoch": 4.217870020863232, "grad_norm": 2.4995994567871094, "learning_rate": 0.00010296575096498014, "loss": 4.6027, "step": 699500 }, { "epoch": 4.218171512644565, "grad_norm": 1.6317754983901978, "learning_rate": 0.0001029464286747245, "loss": 4.622, "step": 699550 }, { "epoch": 4.218473004425899, "grad_norm": 4.5000386238098145, "learning_rate": 0.00010292710725042273, "loss": 5.2847, "step": 699600 }, { "epoch": 4.218774496207233, "grad_norm": 0.689042329788208, "learning_rate": 0.00010290778669243037, "loss": 4.818, "step": 699650 }, { "epoch": 4.2190759879885675, "grad_norm": 2.832915782928467, "learning_rate": 0.000102888467001103, "loss": 4.5983, "step": 699700 }, { "epoch": 4.219377479769902, "grad_norm": 3.199431896209717, "learning_rate": 0.00010286914817679612, "loss": 4.8655, "step": 699750 }, { "epoch": 4.219678971551236, "grad_norm": 2.8288071155548096, "learning_rate": 0.00010284983021986535, "loss": 5.0537, "step": 699800 }, { "epoch": 4.219980463332569, "grad_norm": 0.6634432673454285, "learning_rate": 0.00010283051313066616, "loss": 4.4372, "step": 699850 }, { "epoch": 4.2202819551139035, "grad_norm": 1.3136248588562012, "learning_rate": 0.00010281119690955403, "loss": 4.863, "step": 699900 }, { "epoch": 4.220583446895238, "grad_norm": 2.895793914794922, "learning_rate": 0.00010279188155688456, "loss": 4.8012, "step": 699950 }, { "epoch": 4.220884938676572, "grad_norm": 1.98395574092865, "learning_rate": 0.00010277256707301306, "loss": 5.2088, "step": 700000 }, { "epoch": 4.221186430457906, "grad_norm": 1.306893229484558, "learning_rate": 0.00010275325345829507, "loss": 4.0347, "step": 700050 }, { "epoch": 4.221487922239239, "grad_norm": 2.3978042602539062, "learning_rate": 0.00010273394071308608, "loss": 4.821, "step": 700100 }, { "epoch": 4.221789414020574, "grad_norm": 2.088552236557007, "learning_rate": 0.0001027146288377414, "loss": 4.93, "step": 700150 }, { "epoch": 4.222090905801908, "grad_norm": 2.9557909965515137, "learning_rate": 0.00010269531783261653, "loss": 4.7232, "step": 700200 }, { "epoch": 4.222392397583242, "grad_norm": 1.423657774925232, "learning_rate": 0.00010267600769806682, "loss": 4.7691, "step": 700250 }, { "epoch": 4.222693889364576, "grad_norm": 2.843271017074585, "learning_rate": 0.00010265669843444766, "loss": 4.5999, "step": 700300 }, { "epoch": 4.2229953811459096, "grad_norm": 2.6261775493621826, "learning_rate": 0.00010263739004211438, "loss": 4.9694, "step": 700350 }, { "epoch": 4.223296872927244, "grad_norm": 2.1305389404296875, "learning_rate": 0.00010261808252142239, "loss": 4.8644, "step": 700400 }, { "epoch": 4.223598364708578, "grad_norm": 2.396275281906128, "learning_rate": 0.00010259877587272697, "loss": 4.8415, "step": 700450 }, { "epoch": 4.223899856489912, "grad_norm": 1.6962846517562866, "learning_rate": 0.00010257947009638336, "loss": 4.5721, "step": 700500 }, { "epoch": 4.224201348271246, "grad_norm": 2.0973405838012695, "learning_rate": 0.00010256016519274703, "loss": 4.6684, "step": 700550 }, { "epoch": 4.224502840052581, "grad_norm": 3.6911115646362305, "learning_rate": 0.0001025408611621731, "loss": 4.928, "step": 700600 }, { "epoch": 4.224804331833914, "grad_norm": 2.157688856124878, "learning_rate": 0.00010252155800501688, "loss": 4.6762, "step": 700650 }, { "epoch": 4.225105823615248, "grad_norm": 1.5618550777435303, "learning_rate": 0.0001025022557216337, "loss": 4.9098, "step": 700700 }, { "epoch": 4.225407315396582, "grad_norm": 2.7333295345306396, "learning_rate": 0.00010248295431237864, "loss": 4.7879, "step": 700750 }, { "epoch": 4.2257088071779165, "grad_norm": 3.6258177757263184, "learning_rate": 0.00010246365377760704, "loss": 4.6662, "step": 700800 }, { "epoch": 4.226010298959251, "grad_norm": 2.5870490074157715, "learning_rate": 0.00010244435411767404, "loss": 5.2589, "step": 700850 }, { "epoch": 4.226311790740584, "grad_norm": 3.390024423599243, "learning_rate": 0.00010242505533293483, "loss": 4.7938, "step": 700900 }, { "epoch": 4.226613282521918, "grad_norm": 2.545163631439209, "learning_rate": 0.0001024057574237446, "loss": 4.8529, "step": 700950 }, { "epoch": 4.2269147743032525, "grad_norm": 2.3938982486724854, "learning_rate": 0.0001023864603904585, "loss": 4.8869, "step": 701000 }, { "epoch": 4.227216266084587, "grad_norm": 3.0058913230895996, "learning_rate": 0.00010236716423343163, "loss": 4.8362, "step": 701050 }, { "epoch": 4.227517757865921, "grad_norm": 2.3203468322753906, "learning_rate": 0.00010234786895301909, "loss": 4.9113, "step": 701100 }, { "epoch": 4.227819249647254, "grad_norm": 2.297325372695923, "learning_rate": 0.00010232857454957609, "loss": 4.5518, "step": 701150 }, { "epoch": 4.228120741428588, "grad_norm": 3.1171715259552, "learning_rate": 0.00010230928102345758, "loss": 4.6907, "step": 701200 }, { "epoch": 4.228422233209923, "grad_norm": 1.759853720664978, "learning_rate": 0.00010228998837501872, "loss": 4.7484, "step": 701250 }, { "epoch": 4.228723724991257, "grad_norm": 2.545626163482666, "learning_rate": 0.00010227069660461458, "loss": 4.8228, "step": 701300 }, { "epoch": 4.229025216772591, "grad_norm": 3.9924168586730957, "learning_rate": 0.00010225140571260008, "loss": 4.5423, "step": 701350 }, { "epoch": 4.229326708553925, "grad_norm": 3.3047661781311035, "learning_rate": 0.00010223211569933032, "loss": 5.0787, "step": 701400 }, { "epoch": 4.2296282003352585, "grad_norm": 1.8154199123382568, "learning_rate": 0.00010221282656516037, "loss": 4.9282, "step": 701450 }, { "epoch": 4.229929692116593, "grad_norm": 2.21349835395813, "learning_rate": 0.00010219353831044512, "loss": 4.6537, "step": 701500 }, { "epoch": 4.230231183897927, "grad_norm": 3.4407520294189453, "learning_rate": 0.00010217425093553953, "loss": 5.1919, "step": 701550 }, { "epoch": 4.230532675679261, "grad_norm": 2.311922550201416, "learning_rate": 0.00010215496444079868, "loss": 5.039, "step": 701600 }, { "epoch": 4.230834167460595, "grad_norm": 2.4101510047912598, "learning_rate": 0.0001021356788265774, "loss": 4.7005, "step": 701650 }, { "epoch": 4.231135659241929, "grad_norm": 2.0795674324035645, "learning_rate": 0.00010211639409323059, "loss": 4.8564, "step": 701700 }, { "epoch": 4.231437151023263, "grad_norm": 2.5178182125091553, "learning_rate": 0.00010209711024111331, "loss": 4.6867, "step": 701750 }, { "epoch": 4.231738642804597, "grad_norm": 2.3184380531311035, "learning_rate": 0.00010207782727058026, "loss": 4.9006, "step": 701800 }, { "epoch": 4.232040134585931, "grad_norm": 2.136524200439453, "learning_rate": 0.00010205854518198647, "loss": 4.8309, "step": 701850 }, { "epoch": 4.2323416263672655, "grad_norm": 2.0261120796203613, "learning_rate": 0.00010203926397568674, "loss": 4.9413, "step": 701900 }, { "epoch": 4.2326431181486, "grad_norm": 1.2696630954742432, "learning_rate": 0.00010201998365203588, "loss": 4.6003, "step": 701950 }, { "epoch": 4.232944609929933, "grad_norm": 3.0629682540893555, "learning_rate": 0.00010200070421138875, "loss": 4.6876, "step": 702000 }, { "epoch": 4.232944609929933, "eval_loss": 5.174034595489502, "eval_runtime": 39.0693, "eval_samples_per_second": 13.105, "eval_steps_per_second": 6.552, "eval_tts_loss": 7.66600047969205, "step": 702000 }, { "epoch": 4.233246101711267, "grad_norm": 3.149853467941284, "learning_rate": 0.0001019814256541002, "loss": 4.5683, "step": 702050 }, { "epoch": 4.2335475934926015, "grad_norm": 3.981969118118286, "learning_rate": 0.00010196214798052497, "loss": 4.6069, "step": 702100 }, { "epoch": 4.233849085273936, "grad_norm": 0.3324117064476013, "learning_rate": 0.00010194287119101781, "loss": 4.6309, "step": 702150 }, { "epoch": 4.23415057705527, "grad_norm": 2.538895606994629, "learning_rate": 0.00010192359528593356, "loss": 4.8878, "step": 702200 }, { "epoch": 4.234452068836603, "grad_norm": 2.2750539779663086, "learning_rate": 0.00010190432026562697, "loss": 5.0519, "step": 702250 }, { "epoch": 4.234753560617937, "grad_norm": 2.844101667404175, "learning_rate": 0.00010188504613045265, "loss": 4.8498, "step": 702300 }, { "epoch": 4.235055052399272, "grad_norm": 3.8505501747131348, "learning_rate": 0.00010186577288076541, "loss": 4.9864, "step": 702350 }, { "epoch": 4.235356544180606, "grad_norm": 2.2806801795959473, "learning_rate": 0.00010184650051691998, "loss": 4.7397, "step": 702400 }, { "epoch": 4.23565803596194, "grad_norm": 2.472121477127075, "learning_rate": 0.00010182722903927096, "loss": 4.4904, "step": 702450 }, { "epoch": 4.235959527743273, "grad_norm": 2.3980319499969482, "learning_rate": 0.00010180795844817301, "loss": 4.7786, "step": 702500 }, { "epoch": 4.2362610195246075, "grad_norm": 2.797913074493408, "learning_rate": 0.00010178868874398087, "loss": 4.8542, "step": 702550 }, { "epoch": 4.236562511305942, "grad_norm": 1.863036870956421, "learning_rate": 0.00010176941992704907, "loss": 4.9799, "step": 702600 }, { "epoch": 4.236864003087276, "grad_norm": 2.300180196762085, "learning_rate": 0.00010175015199773228, "loss": 4.6629, "step": 702650 }, { "epoch": 4.23716549486861, "grad_norm": 3.2169764041900635, "learning_rate": 0.00010173088495638514, "loss": 4.7176, "step": 702700 }, { "epoch": 4.2374669866499435, "grad_norm": 2.745842933654785, "learning_rate": 0.00010171161880336208, "loss": 4.7706, "step": 702750 }, { "epoch": 4.237768478431278, "grad_norm": 3.4412198066711426, "learning_rate": 0.00010169235353901782, "loss": 4.8486, "step": 702800 }, { "epoch": 4.238069970212612, "grad_norm": 2.7127695083618164, "learning_rate": 0.0001016730891637069, "loss": 4.7853, "step": 702850 }, { "epoch": 4.238371461993946, "grad_norm": 1.9563605785369873, "learning_rate": 0.00010165382567778372, "loss": 4.0847, "step": 702900 }, { "epoch": 4.23867295377528, "grad_norm": 3.5314927101135254, "learning_rate": 0.00010163456308160289, "loss": 4.7034, "step": 702950 }, { "epoch": 4.2389744455566145, "grad_norm": 0.7932218313217163, "learning_rate": 0.00010161530137551896, "loss": 4.246, "step": 703000 }, { "epoch": 4.239275937337948, "grad_norm": 0.6958334445953369, "learning_rate": 0.00010159604055988635, "loss": 4.8528, "step": 703050 }, { "epoch": 4.239577429119282, "grad_norm": 2.2492001056671143, "learning_rate": 0.00010157678063505947, "loss": 4.405, "step": 703100 }, { "epoch": 4.239878920900616, "grad_norm": 0.5960789322853088, "learning_rate": 0.00010155752160139294, "loss": 4.9208, "step": 703150 }, { "epoch": 4.2401804126819505, "grad_norm": 1.1174126863479614, "learning_rate": 0.00010153826345924101, "loss": 5.0286, "step": 703200 }, { "epoch": 4.240481904463285, "grad_norm": 2.4814937114715576, "learning_rate": 0.00010151900620895821, "loss": 4.3958, "step": 703250 }, { "epoch": 4.240783396244618, "grad_norm": 3.84442400932312, "learning_rate": 0.00010149974985089897, "loss": 4.7181, "step": 703300 }, { "epoch": 4.241084888025952, "grad_norm": 2.4266958236694336, "learning_rate": 0.00010148049438541754, "loss": 4.64, "step": 703350 }, { "epoch": 4.241386379807286, "grad_norm": 2.2543654441833496, "learning_rate": 0.00010146123981286841, "loss": 4.7008, "step": 703400 }, { "epoch": 4.241687871588621, "grad_norm": 2.221843719482422, "learning_rate": 0.0001014419861336059, "loss": 4.6889, "step": 703450 }, { "epoch": 4.241989363369955, "grad_norm": 1.1548045873641968, "learning_rate": 0.0001014227333479843, "loss": 4.698, "step": 703500 }, { "epoch": 4.242290855151289, "grad_norm": 2.9240880012512207, "learning_rate": 0.00010140348145635797, "loss": 5.0145, "step": 703550 }, { "epoch": 4.242592346932622, "grad_norm": 1.8490660190582275, "learning_rate": 0.00010138423045908127, "loss": 4.8824, "step": 703600 }, { "epoch": 4.2428938387139565, "grad_norm": 4.366124153137207, "learning_rate": 0.00010136498035650839, "loss": 4.5448, "step": 703650 }, { "epoch": 4.243195330495291, "grad_norm": 2.984151840209961, "learning_rate": 0.00010134573114899362, "loss": 4.8435, "step": 703700 }, { "epoch": 4.243496822276625, "grad_norm": 2.2963621616363525, "learning_rate": 0.0001013264828368913, "loss": 4.4638, "step": 703750 }, { "epoch": 4.243798314057959, "grad_norm": 1.966739535331726, "learning_rate": 0.00010130723542055552, "loss": 4.4456, "step": 703800 }, { "epoch": 4.2440998058392925, "grad_norm": 0.9139816164970398, "learning_rate": 0.00010128798890034064, "loss": 4.5949, "step": 703850 }, { "epoch": 4.244401297620627, "grad_norm": 2.0985279083251953, "learning_rate": 0.00010126874327660084, "loss": 4.807, "step": 703900 }, { "epoch": 4.244702789401961, "grad_norm": 1.9953666925430298, "learning_rate": 0.00010124949854969018, "loss": 4.3983, "step": 703950 }, { "epoch": 4.245004281183295, "grad_norm": 3.51871657371521, "learning_rate": 0.000101230254719963, "loss": 4.3516, "step": 704000 }, { "epoch": 4.245305772964629, "grad_norm": 3.5818941593170166, "learning_rate": 0.0001012110117877734, "loss": 4.6054, "step": 704050 }, { "epoch": 4.245607264745963, "grad_norm": 3.0096559524536133, "learning_rate": 0.00010119176975347544, "loss": 4.4109, "step": 704100 }, { "epoch": 4.245908756527297, "grad_norm": 1.3176259994506836, "learning_rate": 0.00010117252861742329, "loss": 4.7867, "step": 704150 }, { "epoch": 4.246210248308631, "grad_norm": 1.971431016921997, "learning_rate": 0.00010115328837997116, "loss": 4.3644, "step": 704200 }, { "epoch": 4.246511740089965, "grad_norm": 2.530287981033325, "learning_rate": 0.00010113404904147299, "loss": 4.5698, "step": 704250 }, { "epoch": 4.246813231871299, "grad_norm": 2.5770130157470703, "learning_rate": 0.00010111481060228291, "loss": 4.6666, "step": 704300 }, { "epoch": 4.247114723652634, "grad_norm": 5.076071262359619, "learning_rate": 0.00010109557306275501, "loss": 4.5317, "step": 704350 }, { "epoch": 4.247416215433967, "grad_norm": 4.948085784912109, "learning_rate": 0.00010107633642324326, "loss": 4.7027, "step": 704400 }, { "epoch": 4.247717707215301, "grad_norm": 2.700011730194092, "learning_rate": 0.00010105710068410174, "loss": 4.7153, "step": 704450 }, { "epoch": 4.248019198996635, "grad_norm": 3.079463243484497, "learning_rate": 0.00010103786584568447, "loss": 4.5303, "step": 704500 }, { "epoch": 4.24832069077797, "grad_norm": 1.2913424968719482, "learning_rate": 0.00010101863190834534, "loss": 4.5394, "step": 704550 }, { "epoch": 4.248622182559304, "grad_norm": 2.707845687866211, "learning_rate": 0.00010099939887243843, "loss": 4.8577, "step": 704600 }, { "epoch": 4.248923674340637, "grad_norm": 2.95833158493042, "learning_rate": 0.00010098016673831768, "loss": 4.9311, "step": 704650 }, { "epoch": 4.249225166121971, "grad_norm": 1.1252719163894653, "learning_rate": 0.00010096093550633694, "loss": 4.4218, "step": 704700 }, { "epoch": 4.2495266579033055, "grad_norm": 0.5454203486442566, "learning_rate": 0.00010094170517685022, "loss": 4.8527, "step": 704750 }, { "epoch": 4.24982814968464, "grad_norm": 2.5548813343048096, "learning_rate": 0.00010092247575021144, "loss": 5.0093, "step": 704800 }, { "epoch": 4.250129641465974, "grad_norm": 1.892993688583374, "learning_rate": 0.00010090324722677445, "loss": 4.9673, "step": 704850 }, { "epoch": 4.250431133247307, "grad_norm": 1.5638048648834229, "learning_rate": 0.00010088401960689307, "loss": 5.0593, "step": 704900 }, { "epoch": 4.2507326250286415, "grad_norm": 1.4448860883712769, "learning_rate": 0.0001008647928909213, "loss": 4.9181, "step": 704950 }, { "epoch": 4.251034116809976, "grad_norm": 3.5389065742492676, "learning_rate": 0.00010084556707921283, "loss": 4.8971, "step": 705000 }, { "epoch": 4.251034116809976, "eval_loss": 5.186065673828125, "eval_runtime": 39.0952, "eval_samples_per_second": 13.096, "eval_steps_per_second": 6.548, "eval_tts_loss": 7.665960902378954, "step": 705000 }, { "epoch": 4.25133560859131, "grad_norm": 3.5628793239593506, "learning_rate": 0.00010082634217212158, "loss": 4.83, "step": 705050 }, { "epoch": 4.251637100372644, "grad_norm": 2.437213182449341, "learning_rate": 0.00010080711817000137, "loss": 4.4403, "step": 705100 }, { "epoch": 4.251938592153978, "grad_norm": 3.4003593921661377, "learning_rate": 0.00010078789507320588, "loss": 4.9708, "step": 705150 }, { "epoch": 4.252240083935312, "grad_norm": 2.8077456951141357, "learning_rate": 0.00010076867288208899, "loss": 5.1868, "step": 705200 }, { "epoch": 4.252541575716646, "grad_norm": 4.286981582641602, "learning_rate": 0.0001007494515970044, "loss": 4.8285, "step": 705250 }, { "epoch": 4.25284306749798, "grad_norm": 2.3197453022003174, "learning_rate": 0.00010073023121830589, "loss": 4.9913, "step": 705300 }, { "epoch": 4.253144559279314, "grad_norm": 1.913227915763855, "learning_rate": 0.0001007110117463471, "loss": 4.421, "step": 705350 }, { "epoch": 4.253446051060648, "grad_norm": 1.5668054819107056, "learning_rate": 0.0001006917931814819, "loss": 4.8873, "step": 705400 }, { "epoch": 4.253747542841982, "grad_norm": 3.4812750816345215, "learning_rate": 0.00010067257552406382, "loss": 4.8744, "step": 705450 }, { "epoch": 4.254049034623316, "grad_norm": 4.1607985496521, "learning_rate": 0.00010065335877444657, "loss": 4.6258, "step": 705500 }, { "epoch": 4.25435052640465, "grad_norm": 2.621382713317871, "learning_rate": 0.00010063414293298385, "loss": 4.6688, "step": 705550 }, { "epoch": 4.254652018185984, "grad_norm": 2.0969929695129395, "learning_rate": 0.00010061492800002932, "loss": 4.7223, "step": 705600 }, { "epoch": 4.254953509967319, "grad_norm": 2.08976149559021, "learning_rate": 0.00010059571397593654, "loss": 4.7798, "step": 705650 }, { "epoch": 4.255255001748653, "grad_norm": 4.121768951416016, "learning_rate": 0.00010057650086105911, "loss": 4.8476, "step": 705700 }, { "epoch": 4.255556493529986, "grad_norm": 1.2088338136672974, "learning_rate": 0.0001005572886557507, "loss": 4.7227, "step": 705750 }, { "epoch": 4.25585798531132, "grad_norm": 2.0656850337982178, "learning_rate": 0.00010053807736036481, "loss": 4.7708, "step": 705800 }, { "epoch": 4.2561594770926545, "grad_norm": 4.144597053527832, "learning_rate": 0.000100518866975255, "loss": 4.7447, "step": 705850 }, { "epoch": 4.256460968873989, "grad_norm": 3.071720838546753, "learning_rate": 0.00010049965750077488, "loss": 4.5672, "step": 705900 }, { "epoch": 4.256762460655323, "grad_norm": 3.622666597366333, "learning_rate": 0.00010048044893727787, "loss": 4.9783, "step": 705950 }, { "epoch": 4.257063952436656, "grad_norm": 2.312563896179199, "learning_rate": 0.00010046124128511756, "loss": 4.8978, "step": 706000 }, { "epoch": 4.2573654442179905, "grad_norm": 2.7083709239959717, "learning_rate": 0.00010044203454464744, "loss": 5.2023, "step": 706050 }, { "epoch": 4.257666935999325, "grad_norm": 2.4521639347076416, "learning_rate": 0.00010042282871622089, "loss": 4.4511, "step": 706100 }, { "epoch": 4.257968427780659, "grad_norm": 2.9809930324554443, "learning_rate": 0.00010040362380019141, "loss": 4.806, "step": 706150 }, { "epoch": 4.258269919561993, "grad_norm": 6.499049663543701, "learning_rate": 0.00010038441979691252, "loss": 4.7041, "step": 706200 }, { "epoch": 4.258571411343326, "grad_norm": 1.7049752473831177, "learning_rate": 0.00010036521670673756, "loss": 4.944, "step": 706250 }, { "epoch": 4.258872903124661, "grad_norm": 3.3883907794952393, "learning_rate": 0.00010034601453001992, "loss": 4.8402, "step": 706300 }, { "epoch": 4.259174394905995, "grad_norm": 2.923070192337036, "learning_rate": 0.00010032681326711309, "loss": 5.2328, "step": 706350 }, { "epoch": 4.259475886687329, "grad_norm": 1.8245385885238647, "learning_rate": 0.0001003076129183703, "loss": 4.7974, "step": 706400 }, { "epoch": 4.259777378468663, "grad_norm": 2.216585636138916, "learning_rate": 0.00010028841348414503, "loss": 4.8755, "step": 706450 }, { "epoch": 4.2600788702499965, "grad_norm": 2.665278196334839, "learning_rate": 0.00010026921496479058, "loss": 4.7854, "step": 706500 }, { "epoch": 4.260380362031331, "grad_norm": 3.243947744369507, "learning_rate": 0.00010025001736066022, "loss": 4.8852, "step": 706550 }, { "epoch": 4.260681853812665, "grad_norm": 2.8944225311279297, "learning_rate": 0.00010023082067210728, "loss": 4.98, "step": 706600 }, { "epoch": 4.260983345593999, "grad_norm": 2.3983700275421143, "learning_rate": 0.00010021162489948514, "loss": 4.8069, "step": 706650 }, { "epoch": 4.261284837375333, "grad_norm": 2.2720415592193604, "learning_rate": 0.0001001924300431469, "loss": 5.014, "step": 706700 }, { "epoch": 4.261586329156668, "grad_norm": 2.3384337425231934, "learning_rate": 0.00010017323610344591, "loss": 4.8317, "step": 706750 }, { "epoch": 4.261887820938001, "grad_norm": 2.1654040813446045, "learning_rate": 0.00010015404308073549, "loss": 4.6444, "step": 706800 }, { "epoch": 4.262189312719335, "grad_norm": 2.49297833442688, "learning_rate": 0.00010013485097536872, "loss": 4.6754, "step": 706850 }, { "epoch": 4.262490804500669, "grad_norm": 2.238334894180298, "learning_rate": 0.00010011565978769881, "loss": 4.6958, "step": 706900 }, { "epoch": 4.2627922962820035, "grad_norm": 2.1906285285949707, "learning_rate": 0.00010009646951807908, "loss": 4.6825, "step": 706950 }, { "epoch": 4.263093788063338, "grad_norm": 2.5279393196105957, "learning_rate": 0.00010007728016686255, "loss": 4.9102, "step": 707000 }, { "epoch": 4.263395279844671, "grad_norm": 4.101426124572754, "learning_rate": 0.00010005809173440246, "loss": 4.6481, "step": 707050 }, { "epoch": 4.263696771626005, "grad_norm": 2.8275609016418457, "learning_rate": 0.00010003890422105195, "loss": 4.5735, "step": 707100 }, { "epoch": 4.2639982634073395, "grad_norm": 1.8398064374923706, "learning_rate": 0.00010001971762716405, "loss": 4.5501, "step": 707150 }, { "epoch": 4.264299755188674, "grad_norm": 3.2277090549468994, "learning_rate": 0.00010000053195309195, "loss": 5.0225, "step": 707200 }, { "epoch": 4.264601246970008, "grad_norm": 3.2110159397125244, "learning_rate": 9.998134719918873e-05, "loss": 4.5522, "step": 707250 }, { "epoch": 4.264902738751342, "grad_norm": 1.3334808349609375, "learning_rate": 9.996216336580739e-05, "loss": 4.518, "step": 707300 }, { "epoch": 4.265204230532675, "grad_norm": 6.586574554443359, "learning_rate": 9.9942980453301e-05, "loss": 4.7778, "step": 707350 }, { "epoch": 4.26550572231401, "grad_norm": 2.619018077850342, "learning_rate": 9.99237984620227e-05, "loss": 4.8438, "step": 707400 }, { "epoch": 4.265807214095344, "grad_norm": 1.469612717628479, "learning_rate": 9.99046173923254e-05, "loss": 4.5476, "step": 707450 }, { "epoch": 4.266108705876678, "grad_norm": 1.9392155408859253, "learning_rate": 9.988543724456209e-05, "loss": 4.7183, "step": 707500 }, { "epoch": 4.266410197658012, "grad_norm": 3.5590970516204834, "learning_rate": 9.986625801908586e-05, "loss": 4.7265, "step": 707550 }, { "epoch": 4.2667116894393455, "grad_norm": 1.9027332067489624, "learning_rate": 9.984707971624953e-05, "loss": 4.6796, "step": 707600 }, { "epoch": 4.26701318122068, "grad_norm": 6.712109565734863, "learning_rate": 9.982790233640616e-05, "loss": 4.7673, "step": 707650 }, { "epoch": 4.267314673002014, "grad_norm": 2.2522850036621094, "learning_rate": 9.980872587990867e-05, "loss": 4.4924, "step": 707700 }, { "epoch": 4.267616164783348, "grad_norm": 3.707045793533325, "learning_rate": 9.97895503471099e-05, "loss": 4.757, "step": 707750 }, { "epoch": 4.267917656564682, "grad_norm": 0.8400431275367737, "learning_rate": 9.977037573836286e-05, "loss": 4.3139, "step": 707800 }, { "epoch": 4.268219148346016, "grad_norm": 1.8134489059448242, "learning_rate": 9.975120205402037e-05, "loss": 4.6261, "step": 707850 }, { "epoch": 4.26852064012735, "grad_norm": 2.658914566040039, "learning_rate": 9.973202929443524e-05, "loss": 4.6482, "step": 707900 }, { "epoch": 4.268822131908684, "grad_norm": 2.50075626373291, "learning_rate": 9.971285745996039e-05, "loss": 4.7384, "step": 707950 }, { "epoch": 4.269123623690018, "grad_norm": 2.57889986038208, "learning_rate": 9.969368655094868e-05, "loss": 4.7467, "step": 708000 }, { "epoch": 4.269123623690018, "eval_loss": 5.181478500366211, "eval_runtime": 38.9883, "eval_samples_per_second": 13.132, "eval_steps_per_second": 6.566, "eval_tts_loss": 7.6561053149111, "step": 708000 }, { "epoch": 4.2694251154713525, "grad_norm": 3.4149599075317383, "learning_rate": 9.967451656775284e-05, "loss": 4.5769, "step": 708050 }, { "epoch": 4.269726607252686, "grad_norm": 3.0119595527648926, "learning_rate": 9.96553475107257e-05, "loss": 4.8677, "step": 708100 }, { "epoch": 4.27002809903402, "grad_norm": 3.0339319705963135, "learning_rate": 9.963617938022009e-05, "loss": 4.5482, "step": 708150 }, { "epoch": 4.270329590815354, "grad_norm": 1.928401231765747, "learning_rate": 9.961701217658867e-05, "loss": 4.3225, "step": 708200 }, { "epoch": 4.270631082596688, "grad_norm": 1.5995205640792847, "learning_rate": 9.959784590018426e-05, "loss": 4.6034, "step": 708250 }, { "epoch": 4.270932574378023, "grad_norm": 3.33396315574646, "learning_rate": 9.95786805513596e-05, "loss": 4.4991, "step": 708300 }, { "epoch": 4.271234066159357, "grad_norm": 2.824087381362915, "learning_rate": 9.95595161304673e-05, "loss": 4.7172, "step": 708350 }, { "epoch": 4.27153555794069, "grad_norm": 2.941904306411743, "learning_rate": 9.95403526378602e-05, "loss": 5.1527, "step": 708400 }, { "epoch": 4.271837049722024, "grad_norm": 2.10632061958313, "learning_rate": 9.952119007389088e-05, "loss": 5.0616, "step": 708450 }, { "epoch": 4.272138541503359, "grad_norm": 3.6873645782470703, "learning_rate": 9.950202843891198e-05, "loss": 4.5779, "step": 708500 }, { "epoch": 4.272440033284693, "grad_norm": 3.870919704437256, "learning_rate": 9.948286773327618e-05, "loss": 4.6657, "step": 708550 }, { "epoch": 4.272741525066027, "grad_norm": 0.7358741760253906, "learning_rate": 9.946370795733617e-05, "loss": 4.7163, "step": 708600 }, { "epoch": 4.27304301684736, "grad_norm": 4.293385982513428, "learning_rate": 9.944454911144447e-05, "loss": 4.6743, "step": 708650 }, { "epoch": 4.2733445086286945, "grad_norm": 2.4817047119140625, "learning_rate": 9.942539119595365e-05, "loss": 4.7103, "step": 708700 }, { "epoch": 4.273646000410029, "grad_norm": 3.37007212638855, "learning_rate": 9.94062342112164e-05, "loss": 5.0185, "step": 708750 }, { "epoch": 4.273947492191363, "grad_norm": 4.38665771484375, "learning_rate": 9.938707815758516e-05, "loss": 4.1895, "step": 708800 }, { "epoch": 4.274248983972697, "grad_norm": 2.7925782203674316, "learning_rate": 9.936792303541252e-05, "loss": 4.733, "step": 708850 }, { "epoch": 4.274550475754031, "grad_norm": 1.9996559619903564, "learning_rate": 9.934876884505099e-05, "loss": 5.3333, "step": 708900 }, { "epoch": 4.274851967535365, "grad_norm": 2.045900344848633, "learning_rate": 9.932961558685316e-05, "loss": 4.7137, "step": 708950 }, { "epoch": 4.275153459316699, "grad_norm": 2.611680507659912, "learning_rate": 9.93104632611714e-05, "loss": 4.6479, "step": 709000 }, { "epoch": 4.275454951098033, "grad_norm": 2.548659086227417, "learning_rate": 9.929131186835819e-05, "loss": 4.5836, "step": 709050 }, { "epoch": 4.275756442879367, "grad_norm": 2.836858034133911, "learning_rate": 9.927216140876608e-05, "loss": 4.9818, "step": 709100 }, { "epoch": 4.2760579346607015, "grad_norm": 1.848268985748291, "learning_rate": 9.92530118827474e-05, "loss": 4.8594, "step": 709150 }, { "epoch": 4.276359426442035, "grad_norm": 3.6788973808288574, "learning_rate": 9.923386329065464e-05, "loss": 4.6977, "step": 709200 }, { "epoch": 4.276660918223369, "grad_norm": 2.0592315196990967, "learning_rate": 9.921471563284021e-05, "loss": 4.7557, "step": 709250 }, { "epoch": 4.276962410004703, "grad_norm": 2.2955098152160645, "learning_rate": 9.919556890965638e-05, "loss": 4.6134, "step": 709300 }, { "epoch": 4.277263901786037, "grad_norm": 1.595200777053833, "learning_rate": 9.917642312145564e-05, "loss": 4.7985, "step": 709350 }, { "epoch": 4.277565393567372, "grad_norm": 1.7660497426986694, "learning_rate": 9.915727826859034e-05, "loss": 4.9311, "step": 709400 }, { "epoch": 4.277866885348706, "grad_norm": 2.0876171588897705, "learning_rate": 9.913813435141275e-05, "loss": 5.0633, "step": 709450 }, { "epoch": 4.278168377130039, "grad_norm": 3.2934441566467285, "learning_rate": 9.911899137027519e-05, "loss": 5.0154, "step": 709500 }, { "epoch": 4.278469868911373, "grad_norm": 2.2405736446380615, "learning_rate": 9.909984932553004e-05, "loss": 4.6064, "step": 709550 }, { "epoch": 4.278771360692708, "grad_norm": 2.411634683609009, "learning_rate": 9.90807082175295e-05, "loss": 4.7942, "step": 709600 }, { "epoch": 4.279072852474042, "grad_norm": 0.6038468480110168, "learning_rate": 9.906156804662582e-05, "loss": 4.5898, "step": 709650 }, { "epoch": 4.279374344255376, "grad_norm": 3.349702835083008, "learning_rate": 9.904242881317136e-05, "loss": 4.9003, "step": 709700 }, { "epoch": 4.279675836036709, "grad_norm": 1.99384605884552, "learning_rate": 9.902329051751821e-05, "loss": 4.6539, "step": 709750 }, { "epoch": 4.2799773278180435, "grad_norm": 2.207571506500244, "learning_rate": 9.900415316001867e-05, "loss": 4.6653, "step": 709800 }, { "epoch": 4.280278819599378, "grad_norm": 1.8638545274734497, "learning_rate": 9.898501674102495e-05, "loss": 4.1085, "step": 709850 }, { "epoch": 4.280580311380712, "grad_norm": 2.6123921871185303, "learning_rate": 9.896588126088914e-05, "loss": 4.8407, "step": 709900 }, { "epoch": 4.280881803162046, "grad_norm": 2.1159229278564453, "learning_rate": 9.894674671996345e-05, "loss": 4.5852, "step": 709950 }, { "epoch": 4.2811832949433795, "grad_norm": 3.3445487022399902, "learning_rate": 9.892761311860009e-05, "loss": 5.1739, "step": 710000 }, { "epoch": 4.281484786724714, "grad_norm": 2.833780288696289, "learning_rate": 9.890848045715111e-05, "loss": 4.585, "step": 710050 }, { "epoch": 4.281786278506048, "grad_norm": 2.333061456680298, "learning_rate": 9.888934873596859e-05, "loss": 4.7169, "step": 710100 }, { "epoch": 4.282087770287382, "grad_norm": 2.1474697589874268, "learning_rate": 9.887021795540475e-05, "loss": 4.756, "step": 710150 }, { "epoch": 4.282389262068716, "grad_norm": 1.0074903964996338, "learning_rate": 9.885108811581153e-05, "loss": 4.748, "step": 710200 }, { "epoch": 4.28269075385005, "grad_norm": 3.041964530944824, "learning_rate": 9.883195921754101e-05, "loss": 4.5476, "step": 710250 }, { "epoch": 4.282992245631384, "grad_norm": 3.1038990020751953, "learning_rate": 9.881283126094532e-05, "loss": 4.8637, "step": 710300 }, { "epoch": 4.283293737412718, "grad_norm": 2.7506515979766846, "learning_rate": 9.879370424637639e-05, "loss": 5.0577, "step": 710350 }, { "epoch": 4.283595229194052, "grad_norm": 2.087938070297241, "learning_rate": 9.877457817418626e-05, "loss": 4.7095, "step": 710400 }, { "epoch": 4.283896720975386, "grad_norm": 2.178609609603882, "learning_rate": 9.875545304472696e-05, "loss": 4.9477, "step": 710450 }, { "epoch": 4.284198212756721, "grad_norm": 0.44041335582733154, "learning_rate": 9.873632885835035e-05, "loss": 4.8514, "step": 710500 }, { "epoch": 4.284499704538054, "grad_norm": 2.5772781372070312, "learning_rate": 9.871720561540846e-05, "loss": 4.42, "step": 710550 }, { "epoch": 4.284801196319388, "grad_norm": 2.9747204780578613, "learning_rate": 9.869808331625327e-05, "loss": 4.2493, "step": 710600 }, { "epoch": 4.285102688100722, "grad_norm": 1.7756397724151611, "learning_rate": 9.86789619612366e-05, "loss": 4.8322, "step": 710650 }, { "epoch": 4.285404179882057, "grad_norm": 2.5421130657196045, "learning_rate": 9.865984155071038e-05, "loss": 4.5968, "step": 710700 }, { "epoch": 4.285705671663391, "grad_norm": 2.046505928039551, "learning_rate": 9.864072208502656e-05, "loss": 4.5391, "step": 710750 }, { "epoch": 4.286007163444724, "grad_norm": 2.187206983566284, "learning_rate": 9.862160356453695e-05, "loss": 4.5234, "step": 710800 }, { "epoch": 4.286308655226058, "grad_norm": 3.2481088638305664, "learning_rate": 9.860248598959334e-05, "loss": 4.8664, "step": 710850 }, { "epoch": 4.2866101470073925, "grad_norm": 2.9539685249328613, "learning_rate": 9.858336936054772e-05, "loss": 4.6827, "step": 710900 }, { "epoch": 4.286911638788727, "grad_norm": 2.5393497943878174, "learning_rate": 9.856425367775172e-05, "loss": 4.6471, "step": 710950 }, { "epoch": 4.287213130570061, "grad_norm": 3.0102615356445312, "learning_rate": 9.854513894155729e-05, "loss": 4.7952, "step": 711000 }, { "epoch": 4.287213130570061, "eval_loss": 5.174557685852051, "eval_runtime": 39.1404, "eval_samples_per_second": 13.081, "eval_steps_per_second": 6.541, "eval_tts_loss": 7.629475632010296, "step": 711000 }, { "epoch": 4.287514622351395, "grad_norm": 2.6466026306152344, "learning_rate": 9.852602515231617e-05, "loss": 4.6941, "step": 711050 }, { "epoch": 4.2878161141327285, "grad_norm": 4.652525901794434, "learning_rate": 9.850691231038001e-05, "loss": 4.8291, "step": 711100 }, { "epoch": 4.288117605914063, "grad_norm": 2.7124905586242676, "learning_rate": 9.848780041610068e-05, "loss": 4.5773, "step": 711150 }, { "epoch": 4.288419097695397, "grad_norm": 1.9735133647918701, "learning_rate": 9.846868946982992e-05, "loss": 4.8335, "step": 711200 }, { "epoch": 4.288720589476731, "grad_norm": 0.4471640884876251, "learning_rate": 9.844957947191936e-05, "loss": 4.8102, "step": 711250 }, { "epoch": 4.289022081258065, "grad_norm": 3.610971450805664, "learning_rate": 9.84304704227207e-05, "loss": 5.0581, "step": 711300 }, { "epoch": 4.289323573039399, "grad_norm": 2.442675828933716, "learning_rate": 9.841136232258569e-05, "loss": 4.983, "step": 711350 }, { "epoch": 4.289625064820733, "grad_norm": 3.55012845993042, "learning_rate": 9.839225517186591e-05, "loss": 4.5734, "step": 711400 }, { "epoch": 4.289926556602067, "grad_norm": 2.7842371463775635, "learning_rate": 9.837314897091299e-05, "loss": 4.6507, "step": 711450 }, { "epoch": 4.290228048383401, "grad_norm": 2.8844523429870605, "learning_rate": 9.835404372007866e-05, "loss": 4.6511, "step": 711500 }, { "epoch": 4.290529540164735, "grad_norm": 3.7983791828155518, "learning_rate": 9.833493941971438e-05, "loss": 4.5287, "step": 711550 }, { "epoch": 4.290831031946069, "grad_norm": 2.006723403930664, "learning_rate": 9.831583607017184e-05, "loss": 4.5298, "step": 711600 }, { "epoch": 4.291132523727403, "grad_norm": 1.3920950889587402, "learning_rate": 9.829673367180261e-05, "loss": 4.7608, "step": 711650 }, { "epoch": 4.291434015508737, "grad_norm": 0.7624642848968506, "learning_rate": 9.827763222495816e-05, "loss": 4.6413, "step": 711700 }, { "epoch": 4.291735507290071, "grad_norm": 3.3368656635284424, "learning_rate": 9.825853172999008e-05, "loss": 4.8478, "step": 711750 }, { "epoch": 4.292036999071406, "grad_norm": 2.2370855808258057, "learning_rate": 9.823943218724994e-05, "loss": 4.7902, "step": 711800 }, { "epoch": 4.292338490852739, "grad_norm": 2.424220085144043, "learning_rate": 9.822033359708915e-05, "loss": 4.7317, "step": 711850 }, { "epoch": 4.292639982634073, "grad_norm": 2.239017963409424, "learning_rate": 9.820123595985919e-05, "loss": 4.8069, "step": 711900 }, { "epoch": 4.292941474415407, "grad_norm": 3.325279951095581, "learning_rate": 9.818213927591163e-05, "loss": 4.4569, "step": 711950 }, { "epoch": 4.2932429661967415, "grad_norm": 1.4681965112686157, "learning_rate": 9.816304354559778e-05, "loss": 4.4744, "step": 712000 }, { "epoch": 4.293544457978076, "grad_norm": 1.3552666902542114, "learning_rate": 9.814394876926919e-05, "loss": 4.342, "step": 712050 }, { "epoch": 4.29384594975941, "grad_norm": 3.8629815578460693, "learning_rate": 9.812485494727724e-05, "loss": 4.6051, "step": 712100 }, { "epoch": 4.294147441540743, "grad_norm": 4.899465560913086, "learning_rate": 9.810576207997323e-05, "loss": 4.8353, "step": 712150 }, { "epoch": 4.294448933322077, "grad_norm": 2.037003517150879, "learning_rate": 9.808667016770866e-05, "loss": 5.0067, "step": 712200 }, { "epoch": 4.294750425103412, "grad_norm": 2.602713108062744, "learning_rate": 9.806757921083486e-05, "loss": 4.4289, "step": 712250 }, { "epoch": 4.295051916884746, "grad_norm": 4.70314884185791, "learning_rate": 9.80484892097031e-05, "loss": 4.7208, "step": 712300 }, { "epoch": 4.29535340866608, "grad_norm": 2.3157567977905273, "learning_rate": 9.802940016466476e-05, "loss": 4.5629, "step": 712350 }, { "epoch": 4.295654900447413, "grad_norm": 2.2477385997772217, "learning_rate": 9.801031207607117e-05, "loss": 4.7547, "step": 712400 }, { "epoch": 4.295956392228748, "grad_norm": 2.0317585468292236, "learning_rate": 9.799122494427362e-05, "loss": 4.557, "step": 712450 }, { "epoch": 4.296257884010082, "grad_norm": 1.895647644996643, "learning_rate": 9.79721387696233e-05, "loss": 4.4003, "step": 712500 }, { "epoch": 4.296559375791416, "grad_norm": 3.709216356277466, "learning_rate": 9.795305355247154e-05, "loss": 5.0617, "step": 712550 }, { "epoch": 4.29686086757275, "grad_norm": 2.549639940261841, "learning_rate": 9.793396929316956e-05, "loss": 4.8264, "step": 712600 }, { "epoch": 4.297162359354084, "grad_norm": 2.1858437061309814, "learning_rate": 9.791488599206857e-05, "loss": 4.8163, "step": 712650 }, { "epoch": 4.297463851135418, "grad_norm": 1.4119913578033447, "learning_rate": 9.789580364951974e-05, "loss": 5.0851, "step": 712700 }, { "epoch": 4.297765342916752, "grad_norm": 2.7440831661224365, "learning_rate": 9.787672226587436e-05, "loss": 4.6542, "step": 712750 }, { "epoch": 4.298066834698086, "grad_norm": 0.9639298319816589, "learning_rate": 9.785764184148348e-05, "loss": 5.2052, "step": 712800 }, { "epoch": 4.29836832647942, "grad_norm": 2.346362352371216, "learning_rate": 9.783856237669828e-05, "loss": 4.7931, "step": 712850 }, { "epoch": 4.298669818260755, "grad_norm": 2.675394296646118, "learning_rate": 9.781948387186995e-05, "loss": 5.2807, "step": 712900 }, { "epoch": 4.298971310042088, "grad_norm": 1.1857417821884155, "learning_rate": 9.780040632734949e-05, "loss": 5.0557, "step": 712950 }, { "epoch": 4.299272801823422, "grad_norm": 1.4175390005111694, "learning_rate": 9.778132974348808e-05, "loss": 4.3235, "step": 713000 }, { "epoch": 4.299574293604756, "grad_norm": 1.309189796447754, "learning_rate": 9.776225412063682e-05, "loss": 4.6556, "step": 713050 }, { "epoch": 4.2998757853860905, "grad_norm": 2.8169898986816406, "learning_rate": 9.774317945914667e-05, "loss": 4.9179, "step": 713100 }, { "epoch": 4.300177277167425, "grad_norm": 2.8877718448638916, "learning_rate": 9.772410575936871e-05, "loss": 4.6531, "step": 713150 }, { "epoch": 4.300478768948759, "grad_norm": 2.9365744590759277, "learning_rate": 9.770503302165407e-05, "loss": 4.9913, "step": 713200 }, { "epoch": 4.300780260730092, "grad_norm": 2.895537853240967, "learning_rate": 9.76859612463536e-05, "loss": 4.407, "step": 713250 }, { "epoch": 4.301081752511426, "grad_norm": 3.5473251342773438, "learning_rate": 9.766689043381835e-05, "loss": 4.8122, "step": 713300 }, { "epoch": 4.301383244292761, "grad_norm": 2.0994367599487305, "learning_rate": 9.764782058439936e-05, "loss": 4.7684, "step": 713350 }, { "epoch": 4.301684736074095, "grad_norm": 1.9731109142303467, "learning_rate": 9.762875169844749e-05, "loss": 4.6167, "step": 713400 }, { "epoch": 4.301986227855429, "grad_norm": 2.7579762935638428, "learning_rate": 9.760968377631367e-05, "loss": 4.3743, "step": 713450 }, { "epoch": 4.302287719636762, "grad_norm": 3.1516637802124023, "learning_rate": 9.759061681834891e-05, "loss": 5.0386, "step": 713500 }, { "epoch": 4.302589211418097, "grad_norm": 3.232808828353882, "learning_rate": 9.7571550824904e-05, "loss": 4.5518, "step": 713550 }, { "epoch": 4.302890703199431, "grad_norm": 1.7031458616256714, "learning_rate": 9.755248579632991e-05, "loss": 4.7731, "step": 713600 }, { "epoch": 4.303192194980765, "grad_norm": 2.1909139156341553, "learning_rate": 9.753342173297752e-05, "loss": 4.3782, "step": 713650 }, { "epoch": 4.303493686762099, "grad_norm": 2.9629874229431152, "learning_rate": 9.751435863519754e-05, "loss": 4.7423, "step": 713700 }, { "epoch": 4.3037951785434325, "grad_norm": 2.7453010082244873, "learning_rate": 9.749529650334091e-05, "loss": 4.139, "step": 713750 }, { "epoch": 4.304096670324767, "grad_norm": 2.5607216358184814, "learning_rate": 9.747623533775846e-05, "loss": 5.1235, "step": 713800 }, { "epoch": 4.304398162106101, "grad_norm": 2.408045530319214, "learning_rate": 9.745717513880094e-05, "loss": 4.2831, "step": 713850 }, { "epoch": 4.304699653887435, "grad_norm": 2.4617271423339844, "learning_rate": 9.743811590681908e-05, "loss": 4.2692, "step": 713900 }, { "epoch": 4.305001145668769, "grad_norm": 2.4701876640319824, "learning_rate": 9.741905764216376e-05, "loss": 4.5531, "step": 713950 }, { "epoch": 4.305302637450103, "grad_norm": 0.9967435002326965, "learning_rate": 9.74000003451856e-05, "loss": 4.7397, "step": 714000 }, { "epoch": 4.305302637450103, "eval_loss": 5.182185173034668, "eval_runtime": 38.9834, "eval_samples_per_second": 13.134, "eval_steps_per_second": 6.567, "eval_tts_loss": 7.608413420528045, "step": 714000 }, { "epoch": 4.305604129231437, "grad_norm": 0.9842645525932312, "learning_rate": 9.738094401623535e-05, "loss": 4.9626, "step": 714050 }, { "epoch": 4.305905621012771, "grad_norm": 1.2926639318466187, "learning_rate": 9.736188865566379e-05, "loss": 4.6607, "step": 714100 }, { "epoch": 4.306207112794105, "grad_norm": 1.7518136501312256, "learning_rate": 9.734283426382148e-05, "loss": 4.5409, "step": 714150 }, { "epoch": 4.3065086045754395, "grad_norm": 2.8872499465942383, "learning_rate": 9.73237808410592e-05, "loss": 4.5489, "step": 714200 }, { "epoch": 4.306810096356774, "grad_norm": 2.4966611862182617, "learning_rate": 9.730472838772759e-05, "loss": 4.8319, "step": 714250 }, { "epoch": 4.307111588138107, "grad_norm": 2.931084632873535, "learning_rate": 9.728567690417722e-05, "loss": 4.8211, "step": 714300 }, { "epoch": 4.307413079919441, "grad_norm": 3.7921195030212402, "learning_rate": 9.726662639075871e-05, "loss": 4.6329, "step": 714350 }, { "epoch": 4.307714571700775, "grad_norm": 2.8435897827148438, "learning_rate": 9.724757684782272e-05, "loss": 4.6109, "step": 714400 }, { "epoch": 4.30801606348211, "grad_norm": 2.1151316165924072, "learning_rate": 9.722852827571982e-05, "loss": 4.911, "step": 714450 }, { "epoch": 4.308317555263444, "grad_norm": 2.870115280151367, "learning_rate": 9.72094806748005e-05, "loss": 4.53, "step": 714500 }, { "epoch": 4.308619047044777, "grad_norm": 2.1460180282592773, "learning_rate": 9.719043404541541e-05, "loss": 4.6333, "step": 714550 }, { "epoch": 4.308920538826111, "grad_norm": 2.2019779682159424, "learning_rate": 9.717138838791498e-05, "loss": 4.8997, "step": 714600 }, { "epoch": 4.309222030607446, "grad_norm": 2.494694948196411, "learning_rate": 9.715234370264974e-05, "loss": 5.0103, "step": 714650 }, { "epoch": 4.30952352238878, "grad_norm": 2.1821279525756836, "learning_rate": 9.713329998997024e-05, "loss": 5.0276, "step": 714700 }, { "epoch": 4.309825014170114, "grad_norm": 3.679837465286255, "learning_rate": 9.711425725022686e-05, "loss": 4.9443, "step": 714750 }, { "epoch": 4.310126505951448, "grad_norm": 2.1741576194763184, "learning_rate": 9.709521548377011e-05, "loss": 4.2464, "step": 714800 }, { "epoch": 4.3104279977327815, "grad_norm": 2.624433755874634, "learning_rate": 9.707617469095048e-05, "loss": 4.5196, "step": 714850 }, { "epoch": 4.310729489514116, "grad_norm": 3.5315091609954834, "learning_rate": 9.705713487211826e-05, "loss": 4.6291, "step": 714900 }, { "epoch": 4.31103098129545, "grad_norm": 1.5585439205169678, "learning_rate": 9.703809602762394e-05, "loss": 4.9239, "step": 714950 }, { "epoch": 4.311332473076784, "grad_norm": 4.352867603302002, "learning_rate": 9.70190581578179e-05, "loss": 4.4157, "step": 715000 }, { "epoch": 4.311633964858118, "grad_norm": 3.625305652618408, "learning_rate": 9.700002126305048e-05, "loss": 4.9386, "step": 715050 }, { "epoch": 4.311935456639452, "grad_norm": 2.7639992237091064, "learning_rate": 9.6980985343672e-05, "loss": 4.9961, "step": 715100 }, { "epoch": 4.312236948420786, "grad_norm": 0.787243127822876, "learning_rate": 9.696195040003288e-05, "loss": 4.7189, "step": 715150 }, { "epoch": 4.31253844020212, "grad_norm": 2.7304556369781494, "learning_rate": 9.694291643248335e-05, "loss": 4.7101, "step": 715200 }, { "epoch": 4.312839931983454, "grad_norm": 1.5067551136016846, "learning_rate": 9.692388344137368e-05, "loss": 4.5285, "step": 715250 }, { "epoch": 4.3131414237647885, "grad_norm": 4.553102016448975, "learning_rate": 9.690485142705429e-05, "loss": 4.8755, "step": 715300 }, { "epoch": 4.313442915546122, "grad_norm": 4.905494689941406, "learning_rate": 9.688582038987525e-05, "loss": 4.612, "step": 715350 }, { "epoch": 4.313744407327456, "grad_norm": 2.3931641578674316, "learning_rate": 9.686679033018692e-05, "loss": 4.4928, "step": 715400 }, { "epoch": 4.31404589910879, "grad_norm": 2.2965683937072754, "learning_rate": 9.684776124833951e-05, "loss": 4.949, "step": 715450 }, { "epoch": 4.314347390890124, "grad_norm": 1.7373476028442383, "learning_rate": 9.682873314468316e-05, "loss": 4.8809, "step": 715500 }, { "epoch": 4.314648882671459, "grad_norm": 4.971594333648682, "learning_rate": 9.680970601956809e-05, "loss": 4.7474, "step": 715550 }, { "epoch": 4.314950374452792, "grad_norm": 2.357334613800049, "learning_rate": 9.679067987334453e-05, "loss": 5.2746, "step": 715600 }, { "epoch": 4.315251866234126, "grad_norm": 3.0043585300445557, "learning_rate": 9.677165470636255e-05, "loss": 4.7257, "step": 715650 }, { "epoch": 4.31555335801546, "grad_norm": 2.295711040496826, "learning_rate": 9.675263051897228e-05, "loss": 4.8967, "step": 715700 }, { "epoch": 4.315854849796795, "grad_norm": 2.6464908123016357, "learning_rate": 9.673360731152385e-05, "loss": 4.86, "step": 715750 }, { "epoch": 4.316156341578129, "grad_norm": 2.2302329540252686, "learning_rate": 9.671458508436743e-05, "loss": 4.7241, "step": 715800 }, { "epoch": 4.316457833359463, "grad_norm": 3.0578465461730957, "learning_rate": 9.669556383785296e-05, "loss": 4.8867, "step": 715850 }, { "epoch": 4.316759325140796, "grad_norm": 2.4565486907958984, "learning_rate": 9.667654357233056e-05, "loss": 4.7114, "step": 715900 }, { "epoch": 4.3170608169221305, "grad_norm": 1.9457927942276, "learning_rate": 9.665752428815035e-05, "loss": 5.0405, "step": 715950 }, { "epoch": 4.317362308703465, "grad_norm": 2.875690460205078, "learning_rate": 9.663850598566226e-05, "loss": 4.8543, "step": 716000 }, { "epoch": 4.317663800484799, "grad_norm": 0.8346315622329712, "learning_rate": 9.661948866521626e-05, "loss": 4.6184, "step": 716050 }, { "epoch": 4.317965292266133, "grad_norm": 3.3564293384552, "learning_rate": 9.660047232716248e-05, "loss": 4.4319, "step": 716100 }, { "epoch": 4.3182667840474664, "grad_norm": 3.0325093269348145, "learning_rate": 9.658145697185072e-05, "loss": 4.8937, "step": 716150 }, { "epoch": 4.318568275828801, "grad_norm": 3.046290159225464, "learning_rate": 9.656244259963105e-05, "loss": 4.9999, "step": 716200 }, { "epoch": 4.318869767610135, "grad_norm": 3.185093402862549, "learning_rate": 9.65434292108534e-05, "loss": 4.8047, "step": 716250 }, { "epoch": 4.319171259391469, "grad_norm": 2.6638216972351074, "learning_rate": 9.652441680586758e-05, "loss": 4.6413, "step": 716300 }, { "epoch": 4.319472751172803, "grad_norm": 1.2239147424697876, "learning_rate": 9.650540538502359e-05, "loss": 4.8521, "step": 716350 }, { "epoch": 4.3197742429541375, "grad_norm": 2.3252477645874023, "learning_rate": 9.64863949486713e-05, "loss": 4.6043, "step": 716400 }, { "epoch": 4.320075734735471, "grad_norm": 2.4772961139678955, "learning_rate": 9.646738549716048e-05, "loss": 4.7807, "step": 716450 }, { "epoch": 4.320377226516805, "grad_norm": 3.5734927654266357, "learning_rate": 9.644837703084105e-05, "loss": 4.4026, "step": 716500 }, { "epoch": 4.320678718298139, "grad_norm": 2.8455264568328857, "learning_rate": 9.642936955006287e-05, "loss": 4.9609, "step": 716550 }, { "epoch": 4.320980210079473, "grad_norm": 4.51336145401001, "learning_rate": 9.641036305517565e-05, "loss": 4.9075, "step": 716600 }, { "epoch": 4.321281701860808, "grad_norm": 1.5206691026687622, "learning_rate": 9.639135754652921e-05, "loss": 4.6189, "step": 716650 }, { "epoch": 4.321583193642141, "grad_norm": 3.698639392852783, "learning_rate": 9.637235302447337e-05, "loss": 4.7052, "step": 716700 }, { "epoch": 4.321884685423475, "grad_norm": 3.0867786407470703, "learning_rate": 9.635334948935781e-05, "loss": 4.5938, "step": 716750 }, { "epoch": 4.322186177204809, "grad_norm": 2.973306179046631, "learning_rate": 9.633434694153231e-05, "loss": 4.8187, "step": 716800 }, { "epoch": 4.322487668986144, "grad_norm": 2.412606716156006, "learning_rate": 9.63153453813466e-05, "loss": 5.0045, "step": 716850 }, { "epoch": 4.322789160767478, "grad_norm": 2.7685723304748535, "learning_rate": 9.629634480915028e-05, "loss": 4.5811, "step": 716900 }, { "epoch": 4.323090652548812, "grad_norm": 3.38759708404541, "learning_rate": 9.627734522529315e-05, "loss": 4.9054, "step": 716950 }, { "epoch": 4.323392144330145, "grad_norm": 1.3287413120269775, "learning_rate": 9.625834663012478e-05, "loss": 4.8128, "step": 717000 }, { "epoch": 4.323392144330145, "eval_loss": 5.170957565307617, "eval_runtime": 39.0108, "eval_samples_per_second": 13.125, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.6261745176517035, "step": 717000 }, { "epoch": 4.3236936361114795, "grad_norm": 3.9343106746673584, "learning_rate": 9.623934902399487e-05, "loss": 4.5999, "step": 717050 }, { "epoch": 4.323995127892814, "grad_norm": 2.745187997817993, "learning_rate": 9.622035240725297e-05, "loss": 4.831, "step": 717100 }, { "epoch": 4.324296619674148, "grad_norm": 2.5254034996032715, "learning_rate": 9.62013567802488e-05, "loss": 4.9536, "step": 717150 }, { "epoch": 4.324598111455482, "grad_norm": 2.2552008628845215, "learning_rate": 9.618236214333189e-05, "loss": 4.6765, "step": 717200 }, { "epoch": 4.324899603236815, "grad_norm": 2.6366608142852783, "learning_rate": 9.616336849685175e-05, "loss": 4.7675, "step": 717250 }, { "epoch": 4.32520109501815, "grad_norm": 2.5742695331573486, "learning_rate": 9.614437584115802e-05, "loss": 4.579, "step": 717300 }, { "epoch": 4.325502586799484, "grad_norm": 1.8542312383651733, "learning_rate": 9.612538417660016e-05, "loss": 4.4733, "step": 717350 }, { "epoch": 4.325804078580818, "grad_norm": 1.8817063570022583, "learning_rate": 9.610639350352775e-05, "loss": 4.615, "step": 717400 }, { "epoch": 4.326105570362152, "grad_norm": 0.9942830801010132, "learning_rate": 9.60874038222903e-05, "loss": 5.1153, "step": 717450 }, { "epoch": 4.326407062143486, "grad_norm": 2.2675585746765137, "learning_rate": 9.606841513323716e-05, "loss": 4.6716, "step": 717500 }, { "epoch": 4.32670855392482, "grad_norm": 2.0544395446777344, "learning_rate": 9.604942743671791e-05, "loss": 4.2536, "step": 717550 }, { "epoch": 4.327010045706154, "grad_norm": 2.357954978942871, "learning_rate": 9.603044073308197e-05, "loss": 4.4864, "step": 717600 }, { "epoch": 4.327311537487488, "grad_norm": 3.314614772796631, "learning_rate": 9.601145502267873e-05, "loss": 4.6778, "step": 717650 }, { "epoch": 4.327613029268822, "grad_norm": 2.3637030124664307, "learning_rate": 9.59924703058576e-05, "loss": 5.0102, "step": 717700 }, { "epoch": 4.327914521050156, "grad_norm": 3.121788263320923, "learning_rate": 9.597348658296803e-05, "loss": 4.5553, "step": 717750 }, { "epoch": 4.32821601283149, "grad_norm": 1.3155609369277954, "learning_rate": 9.595450385435931e-05, "loss": 5.0006, "step": 717800 }, { "epoch": 4.328517504612824, "grad_norm": 3.1640827655792236, "learning_rate": 9.593552212038082e-05, "loss": 5.1586, "step": 717850 }, { "epoch": 4.328818996394158, "grad_norm": 2.7046196460723877, "learning_rate": 9.591654138138191e-05, "loss": 4.7497, "step": 717900 }, { "epoch": 4.3291204881754926, "grad_norm": 2.338474988937378, "learning_rate": 9.589756163771184e-05, "loss": 4.8044, "step": 717950 }, { "epoch": 4.329421979956827, "grad_norm": 2.28448748588562, "learning_rate": 9.587858288971995e-05, "loss": 4.7853, "step": 718000 }, { "epoch": 4.32972347173816, "grad_norm": 1.9289921522140503, "learning_rate": 9.585960513775556e-05, "loss": 4.8723, "step": 718050 }, { "epoch": 4.330024963519494, "grad_norm": 3.204669952392578, "learning_rate": 9.58406283821678e-05, "loss": 4.512, "step": 718100 }, { "epoch": 4.3303264553008285, "grad_norm": 2.3417906761169434, "learning_rate": 9.582165262330599e-05, "loss": 4.8567, "step": 718150 }, { "epoch": 4.330627947082163, "grad_norm": 1.4244639873504639, "learning_rate": 9.58026778615194e-05, "loss": 4.5475, "step": 718200 }, { "epoch": 4.330929438863497, "grad_norm": 2.806122303009033, "learning_rate": 9.578370409715717e-05, "loss": 5.1043, "step": 718250 }, { "epoch": 4.33123093064483, "grad_norm": 5.033783435821533, "learning_rate": 9.576473133056844e-05, "loss": 4.2839, "step": 718300 }, { "epoch": 4.331532422426164, "grad_norm": 2.512505054473877, "learning_rate": 9.574575956210249e-05, "loss": 4.6984, "step": 718350 }, { "epoch": 4.331833914207499, "grad_norm": 0.7910401821136475, "learning_rate": 9.57267887921084e-05, "loss": 5.0798, "step": 718400 }, { "epoch": 4.332135405988833, "grad_norm": 2.407626152038574, "learning_rate": 9.570781902093526e-05, "loss": 4.5984, "step": 718450 }, { "epoch": 4.332436897770167, "grad_norm": 2.741011142730713, "learning_rate": 9.568885024893232e-05, "loss": 4.7554, "step": 718500 }, { "epoch": 4.332738389551501, "grad_norm": 2.6053078174591064, "learning_rate": 9.56698824764485e-05, "loss": 4.7405, "step": 718550 }, { "epoch": 4.333039881332835, "grad_norm": 2.5729591846466064, "learning_rate": 9.565091570383301e-05, "loss": 4.4851, "step": 718600 }, { "epoch": 4.333341373114169, "grad_norm": 3.1873209476470947, "learning_rate": 9.563194993143487e-05, "loss": 4.5811, "step": 718650 }, { "epoch": 4.333642864895503, "grad_norm": 2.577958345413208, "learning_rate": 9.561298515960304e-05, "loss": 4.7632, "step": 718700 }, { "epoch": 4.333944356676837, "grad_norm": 2.949540853500366, "learning_rate": 9.559402138868662e-05, "loss": 4.492, "step": 718750 }, { "epoch": 4.334245848458171, "grad_norm": 1.0339434146881104, "learning_rate": 9.557505861903465e-05, "loss": 4.8881, "step": 718800 }, { "epoch": 4.334547340239505, "grad_norm": 2.5077555179595947, "learning_rate": 9.555609685099602e-05, "loss": 4.8111, "step": 718850 }, { "epoch": 4.334848832020839, "grad_norm": 3.19290828704834, "learning_rate": 9.55371360849197e-05, "loss": 4.8167, "step": 718900 }, { "epoch": 4.335150323802173, "grad_norm": 3.098287582397461, "learning_rate": 9.551817632115474e-05, "loss": 4.9368, "step": 718950 }, { "epoch": 4.335451815583507, "grad_norm": 6.361212253570557, "learning_rate": 9.549921756004997e-05, "loss": 4.6258, "step": 719000 }, { "epoch": 4.3357533073648415, "grad_norm": 3.683593273162842, "learning_rate": 9.548025980195428e-05, "loss": 4.2153, "step": 719050 }, { "epoch": 4.336054799146175, "grad_norm": 3.1489944458007812, "learning_rate": 9.546130304721659e-05, "loss": 3.9992, "step": 719100 }, { "epoch": 4.336356290927509, "grad_norm": 2.4215433597564697, "learning_rate": 9.54423472961859e-05, "loss": 4.8513, "step": 719150 }, { "epoch": 4.336657782708843, "grad_norm": 2.601766347885132, "learning_rate": 9.542339254921087e-05, "loss": 4.8143, "step": 719200 }, { "epoch": 4.3369592744901775, "grad_norm": 2.0101065635681152, "learning_rate": 9.54044388066404e-05, "loss": 4.6403, "step": 719250 }, { "epoch": 4.337260766271512, "grad_norm": 3.5244925022125244, "learning_rate": 9.538548606882338e-05, "loss": 4.8556, "step": 719300 }, { "epoch": 4.337562258052845, "grad_norm": 1.172239065170288, "learning_rate": 9.53665343361085e-05, "loss": 4.4821, "step": 719350 }, { "epoch": 4.337863749834179, "grad_norm": 3.307169198989868, "learning_rate": 9.534758360884461e-05, "loss": 4.421, "step": 719400 }, { "epoch": 4.338165241615513, "grad_norm": 2.347896099090576, "learning_rate": 9.532863388738047e-05, "loss": 4.4623, "step": 719450 }, { "epoch": 4.338466733396848, "grad_norm": 0.580107569694519, "learning_rate": 9.530968517206476e-05, "loss": 4.6836, "step": 719500 }, { "epoch": 4.338768225178182, "grad_norm": 2.293987512588501, "learning_rate": 9.529073746324628e-05, "loss": 4.7198, "step": 719550 }, { "epoch": 4.339069716959516, "grad_norm": 2.2669405937194824, "learning_rate": 9.527179076127373e-05, "loss": 4.9763, "step": 719600 }, { "epoch": 4.339371208740849, "grad_norm": 3.529038906097412, "learning_rate": 9.525284506649569e-05, "loss": 4.819, "step": 719650 }, { "epoch": 4.339672700522184, "grad_norm": 2.7724080085754395, "learning_rate": 9.523390037926094e-05, "loss": 4.7278, "step": 719700 }, { "epoch": 4.339974192303518, "grad_norm": 4.135274410247803, "learning_rate": 9.521495669991813e-05, "loss": 4.7234, "step": 719750 }, { "epoch": 4.340275684084852, "grad_norm": 3.9059877395629883, "learning_rate": 9.519601402881584e-05, "loss": 4.7501, "step": 719800 }, { "epoch": 4.340577175866186, "grad_norm": 6.111696243286133, "learning_rate": 9.517707236630266e-05, "loss": 4.5959, "step": 719850 }, { "epoch": 4.3408786676475195, "grad_norm": 3.110992193222046, "learning_rate": 9.515813171272731e-05, "loss": 4.5961, "step": 719900 }, { "epoch": 4.341180159428854, "grad_norm": 2.362874746322632, "learning_rate": 9.513919206843822e-05, "loss": 4.6273, "step": 719950 }, { "epoch": 4.341481651210188, "grad_norm": 2.954364776611328, "learning_rate": 9.512025343378402e-05, "loss": 5.0585, "step": 720000 }, { "epoch": 4.341481651210188, "eval_loss": 5.162813186645508, "eval_runtime": 38.9267, "eval_samples_per_second": 13.153, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.568778134314009, "step": 720000 }, { "epoch": 4.341783142991522, "grad_norm": 2.240941047668457, "learning_rate": 9.510131580911327e-05, "loss": 4.5295, "step": 720050 }, { "epoch": 4.342084634772856, "grad_norm": 2.2103734016418457, "learning_rate": 9.508237919477441e-05, "loss": 4.5995, "step": 720100 }, { "epoch": 4.3423861265541905, "grad_norm": 3.1373772621154785, "learning_rate": 9.506344359111602e-05, "loss": 4.1482, "step": 720150 }, { "epoch": 4.342687618335524, "grad_norm": 2.4369912147521973, "learning_rate": 9.504450899848657e-05, "loss": 5.0096, "step": 720200 }, { "epoch": 4.342989110116858, "grad_norm": 3.767908811569214, "learning_rate": 9.502557541723446e-05, "loss": 4.7874, "step": 720250 }, { "epoch": 4.343290601898192, "grad_norm": 2.1039388179779053, "learning_rate": 9.500664284770818e-05, "loss": 4.7604, "step": 720300 }, { "epoch": 4.3435920936795265, "grad_norm": 2.5961966514587402, "learning_rate": 9.498771129025622e-05, "loss": 4.761, "step": 720350 }, { "epoch": 4.343893585460861, "grad_norm": 3.2604141235351562, "learning_rate": 9.496878074522688e-05, "loss": 4.9806, "step": 720400 }, { "epoch": 4.344195077242194, "grad_norm": 2.3198297023773193, "learning_rate": 9.494985121296858e-05, "loss": 4.4452, "step": 720450 }, { "epoch": 4.344496569023528, "grad_norm": 4.571316719055176, "learning_rate": 9.493092269382978e-05, "loss": 4.8891, "step": 720500 }, { "epoch": 4.344798060804862, "grad_norm": 2.1339738368988037, "learning_rate": 9.491199518815869e-05, "loss": 4.9741, "step": 720550 }, { "epoch": 4.345099552586197, "grad_norm": 2.3913686275482178, "learning_rate": 9.489306869630373e-05, "loss": 4.6062, "step": 720600 }, { "epoch": 4.345401044367531, "grad_norm": 2.774994134902954, "learning_rate": 9.487414321861322e-05, "loss": 4.3956, "step": 720650 }, { "epoch": 4.345702536148865, "grad_norm": 2.1806256771087646, "learning_rate": 9.48552187554354e-05, "loss": 4.7694, "step": 720700 }, { "epoch": 4.346004027930198, "grad_norm": 2.890753746032715, "learning_rate": 9.48362953071186e-05, "loss": 4.8081, "step": 720750 }, { "epoch": 4.346305519711533, "grad_norm": 2.7176597118377686, "learning_rate": 9.481737287401112e-05, "loss": 4.7155, "step": 720800 }, { "epoch": 4.346607011492867, "grad_norm": 3.97745680809021, "learning_rate": 9.479845145646105e-05, "loss": 4.8062, "step": 720850 }, { "epoch": 4.346908503274201, "grad_norm": 3.9035847187042236, "learning_rate": 9.477953105481672e-05, "loss": 4.5107, "step": 720900 }, { "epoch": 4.347209995055535, "grad_norm": 2.863163948059082, "learning_rate": 9.476061166942635e-05, "loss": 4.7429, "step": 720950 }, { "epoch": 4.3475114868368685, "grad_norm": 3.363940477371216, "learning_rate": 9.474169330063811e-05, "loss": 4.7299, "step": 721000 }, { "epoch": 4.347812978618203, "grad_norm": 1.880793809890747, "learning_rate": 9.472277594880008e-05, "loss": 4.355, "step": 721050 }, { "epoch": 4.348114470399537, "grad_norm": 2.573233127593994, "learning_rate": 9.470385961426054e-05, "loss": 4.647, "step": 721100 }, { "epoch": 4.348415962180871, "grad_norm": 4.0052642822265625, "learning_rate": 9.46849442973675e-05, "loss": 4.7135, "step": 721150 }, { "epoch": 4.348717453962205, "grad_norm": 3.6015353202819824, "learning_rate": 9.466602999846914e-05, "loss": 4.7944, "step": 721200 }, { "epoch": 4.349018945743539, "grad_norm": 2.7268803119659424, "learning_rate": 9.464711671791358e-05, "loss": 4.4019, "step": 721250 }, { "epoch": 4.349320437524873, "grad_norm": 1.9279701709747314, "learning_rate": 9.462820445604877e-05, "loss": 5.0311, "step": 721300 }, { "epoch": 4.349621929306207, "grad_norm": 2.221153736114502, "learning_rate": 9.460929321322288e-05, "loss": 4.6208, "step": 721350 }, { "epoch": 4.349923421087541, "grad_norm": 1.2087466716766357, "learning_rate": 9.459038298978392e-05, "loss": 4.7958, "step": 721400 }, { "epoch": 4.3502249128688755, "grad_norm": 1.0575978755950928, "learning_rate": 9.457147378607983e-05, "loss": 4.7783, "step": 721450 }, { "epoch": 4.350526404650209, "grad_norm": 3.1031813621520996, "learning_rate": 9.455256560245865e-05, "loss": 5.1455, "step": 721500 }, { "epoch": 4.350827896431543, "grad_norm": 1.7848800420761108, "learning_rate": 9.453365843926847e-05, "loss": 4.3176, "step": 721550 }, { "epoch": 4.351129388212877, "grad_norm": 2.546030044555664, "learning_rate": 9.451475229685708e-05, "loss": 4.7035, "step": 721600 }, { "epoch": 4.351430879994211, "grad_norm": 1.245720386505127, "learning_rate": 9.449584717557249e-05, "loss": 4.851, "step": 721650 }, { "epoch": 4.351732371775546, "grad_norm": 3.8161017894744873, "learning_rate": 9.447694307576267e-05, "loss": 4.6379, "step": 721700 }, { "epoch": 4.35203386355688, "grad_norm": 2.716601610183716, "learning_rate": 9.445803999777542e-05, "loss": 4.8804, "step": 721750 }, { "epoch": 4.352335355338213, "grad_norm": 6.102606773376465, "learning_rate": 9.443913794195871e-05, "loss": 4.5522, "step": 721800 }, { "epoch": 4.352636847119547, "grad_norm": 2.672772169113159, "learning_rate": 9.442023690866041e-05, "loss": 5.1068, "step": 721850 }, { "epoch": 4.3529383389008816, "grad_norm": 2.0009734630584717, "learning_rate": 9.440133689822827e-05, "loss": 4.5108, "step": 721900 }, { "epoch": 4.353239830682216, "grad_norm": 2.8890106678009033, "learning_rate": 9.438243791101022e-05, "loss": 4.6261, "step": 721950 }, { "epoch": 4.35354132246355, "grad_norm": 5.037230014801025, "learning_rate": 9.436353994735402e-05, "loss": 4.3404, "step": 722000 }, { "epoch": 4.353842814244883, "grad_norm": 2.989157199859619, "learning_rate": 9.434464300760748e-05, "loss": 4.4998, "step": 722050 }, { "epoch": 4.3541443060262175, "grad_norm": 2.3491997718811035, "learning_rate": 9.432574709211832e-05, "loss": 4.2285, "step": 722100 }, { "epoch": 4.354445797807552, "grad_norm": 2.7942140102386475, "learning_rate": 9.430685220123439e-05, "loss": 4.5251, "step": 722150 }, { "epoch": 4.354747289588886, "grad_norm": 3.091571807861328, "learning_rate": 9.428795833530335e-05, "loss": 4.5517, "step": 722200 }, { "epoch": 4.35504878137022, "grad_norm": 2.6991872787475586, "learning_rate": 9.426906549467289e-05, "loss": 4.9531, "step": 722250 }, { "epoch": 4.355350273151554, "grad_norm": 3.8501510620117188, "learning_rate": 9.425017367969081e-05, "loss": 4.97, "step": 722300 }, { "epoch": 4.355651764932888, "grad_norm": 2.5917136669158936, "learning_rate": 9.423128289070468e-05, "loss": 4.6497, "step": 722350 }, { "epoch": 4.355953256714222, "grad_norm": 3.125113010406494, "learning_rate": 9.421239312806222e-05, "loss": 4.3089, "step": 722400 }, { "epoch": 4.356254748495556, "grad_norm": 3.8060617446899414, "learning_rate": 9.4193504392111e-05, "loss": 4.9459, "step": 722450 }, { "epoch": 4.35655624027689, "grad_norm": 3.1272342205047607, "learning_rate": 9.417461668319878e-05, "loss": 4.8666, "step": 722500 }, { "epoch": 4.3568577320582245, "grad_norm": 3.0530197620391846, "learning_rate": 9.415573000167304e-05, "loss": 4.7881, "step": 722550 }, { "epoch": 4.357159223839558, "grad_norm": 2.6573071479797363, "learning_rate": 9.413684434788137e-05, "loss": 4.9589, "step": 722600 }, { "epoch": 4.357460715620892, "grad_norm": 2.661449909210205, "learning_rate": 9.41179597221714e-05, "loss": 4.4868, "step": 722650 }, { "epoch": 4.357762207402226, "grad_norm": 1.986692190170288, "learning_rate": 9.409907612489057e-05, "loss": 4.9188, "step": 722700 }, { "epoch": 4.35806369918356, "grad_norm": 3.316995620727539, "learning_rate": 9.408019355638653e-05, "loss": 4.9861, "step": 722750 }, { "epoch": 4.358365190964895, "grad_norm": 1.572320818901062, "learning_rate": 9.406131201700673e-05, "loss": 4.3979, "step": 722800 }, { "epoch": 4.358666682746228, "grad_norm": 3.537719488143921, "learning_rate": 9.404243150709859e-05, "loss": 4.6287, "step": 722850 }, { "epoch": 4.358968174527562, "grad_norm": 2.8730249404907227, "learning_rate": 9.402355202700965e-05, "loss": 4.77, "step": 722900 }, { "epoch": 4.359269666308896, "grad_norm": 3.0835745334625244, "learning_rate": 9.400467357708743e-05, "loss": 4.6422, "step": 722950 }, { "epoch": 4.3595711580902305, "grad_norm": 1.6526914834976196, "learning_rate": 9.398579615767923e-05, "loss": 4.3887, "step": 723000 }, { "epoch": 4.3595711580902305, "eval_loss": 5.165388107299805, "eval_runtime": 39.0052, "eval_samples_per_second": 13.126, "eval_steps_per_second": 6.563, "eval_tts_loss": 7.679113872992344, "step": 723000 }, { "epoch": 4.359872649871565, "grad_norm": 2.6577584743499756, "learning_rate": 9.39669197691325e-05, "loss": 4.7835, "step": 723050 }, { "epoch": 4.360174141652898, "grad_norm": 2.2761669158935547, "learning_rate": 9.394804441179469e-05, "loss": 4.5597, "step": 723100 }, { "epoch": 4.360475633434232, "grad_norm": 2.6436240673065186, "learning_rate": 9.392917008601312e-05, "loss": 4.7089, "step": 723150 }, { "epoch": 4.3607771252155665, "grad_norm": 1.5801481008529663, "learning_rate": 9.391029679213513e-05, "loss": 4.7687, "step": 723200 }, { "epoch": 4.361078616996901, "grad_norm": 2.4465856552124023, "learning_rate": 9.389142453050813e-05, "loss": 4.4482, "step": 723250 }, { "epoch": 4.361380108778235, "grad_norm": 1.6596699953079224, "learning_rate": 9.387255330147933e-05, "loss": 5.0703, "step": 723300 }, { "epoch": 4.361681600559569, "grad_norm": 2.732975721359253, "learning_rate": 9.385368310539611e-05, "loss": 5.0038, "step": 723350 }, { "epoch": 4.361983092340902, "grad_norm": 2.331568717956543, "learning_rate": 9.383481394260579e-05, "loss": 4.303, "step": 723400 }, { "epoch": 4.362284584122237, "grad_norm": 4.212088584899902, "learning_rate": 9.381594581345548e-05, "loss": 4.9321, "step": 723450 }, { "epoch": 4.362586075903571, "grad_norm": 2.587789535522461, "learning_rate": 9.379707871829251e-05, "loss": 5.1626, "step": 723500 }, { "epoch": 4.362887567684905, "grad_norm": 2.542355537414551, "learning_rate": 9.377821265746418e-05, "loss": 4.5815, "step": 723550 }, { "epoch": 4.363189059466239, "grad_norm": 3.2887015342712402, "learning_rate": 9.375934763131757e-05, "loss": 4.782, "step": 723600 }, { "epoch": 4.363490551247573, "grad_norm": 2.35770320892334, "learning_rate": 9.374048364019986e-05, "loss": 4.825, "step": 723650 }, { "epoch": 4.363792043028907, "grad_norm": 1.790919303894043, "learning_rate": 9.372162068445834e-05, "loss": 5.2931, "step": 723700 }, { "epoch": 4.364093534810241, "grad_norm": 3.4893345832824707, "learning_rate": 9.370275876444002e-05, "loss": 4.4883, "step": 723750 }, { "epoch": 4.364395026591575, "grad_norm": 1.4743590354919434, "learning_rate": 9.368389788049212e-05, "loss": 5.2088, "step": 723800 }, { "epoch": 4.364696518372909, "grad_norm": 1.7669752836227417, "learning_rate": 9.366503803296173e-05, "loss": 4.8167, "step": 723850 }, { "epoch": 4.364998010154244, "grad_norm": 2.35231614112854, "learning_rate": 9.364617922219586e-05, "loss": 4.7709, "step": 723900 }, { "epoch": 4.365299501935577, "grad_norm": 3.290828227996826, "learning_rate": 9.362732144854167e-05, "loss": 4.6641, "step": 723950 }, { "epoch": 4.365600993716911, "grad_norm": 3.354526996612549, "learning_rate": 9.360846471234622e-05, "loss": 4.84, "step": 724000 }, { "epoch": 4.365902485498245, "grad_norm": 3.155391216278076, "learning_rate": 9.358960901395644e-05, "loss": 4.306, "step": 724050 }, { "epoch": 4.3662039772795795, "grad_norm": 3.334977626800537, "learning_rate": 9.35707543537194e-05, "loss": 4.8348, "step": 724100 }, { "epoch": 4.366505469060914, "grad_norm": 2.2852072715759277, "learning_rate": 9.355190073198215e-05, "loss": 5.0501, "step": 724150 }, { "epoch": 4.366806960842247, "grad_norm": 2.5015740394592285, "learning_rate": 9.353304814909158e-05, "loss": 4.6311, "step": 724200 }, { "epoch": 4.367108452623581, "grad_norm": 3.715944290161133, "learning_rate": 9.351419660539467e-05, "loss": 4.4683, "step": 724250 }, { "epoch": 4.3674099444049155, "grad_norm": 2.751494884490967, "learning_rate": 9.34953461012384e-05, "loss": 4.741, "step": 724300 }, { "epoch": 4.36771143618625, "grad_norm": 1.2821906805038452, "learning_rate": 9.34764966369696e-05, "loss": 4.4699, "step": 724350 }, { "epoch": 4.368012927967584, "grad_norm": 1.33090341091156, "learning_rate": 9.345764821293523e-05, "loss": 4.5841, "step": 724400 }, { "epoch": 4.368314419748918, "grad_norm": 2.394224166870117, "learning_rate": 9.34388008294822e-05, "loss": 4.6614, "step": 724450 }, { "epoch": 4.368615911530251, "grad_norm": 3.190747022628784, "learning_rate": 9.341995448695727e-05, "loss": 4.4745, "step": 724500 }, { "epoch": 4.368917403311586, "grad_norm": 2.536766529083252, "learning_rate": 9.340110918570735e-05, "loss": 4.8494, "step": 724550 }, { "epoch": 4.36921889509292, "grad_norm": 1.1866681575775146, "learning_rate": 9.338226492607929e-05, "loss": 4.7745, "step": 724600 }, { "epoch": 4.369520386874254, "grad_norm": 2.3405747413635254, "learning_rate": 9.336342170841975e-05, "loss": 4.8246, "step": 724650 }, { "epoch": 4.369821878655588, "grad_norm": 1.8394532203674316, "learning_rate": 9.334457953307564e-05, "loss": 4.7944, "step": 724700 }, { "epoch": 4.370123370436922, "grad_norm": 1.0287443399429321, "learning_rate": 9.332573840039373e-05, "loss": 4.523, "step": 724750 }, { "epoch": 4.370424862218256, "grad_norm": 2.843190908432007, "learning_rate": 9.330689831072072e-05, "loss": 4.7534, "step": 724800 }, { "epoch": 4.37072635399959, "grad_norm": 3.371138095855713, "learning_rate": 9.328805926440331e-05, "loss": 4.913, "step": 724850 }, { "epoch": 4.371027845780924, "grad_norm": 2.340508222579956, "learning_rate": 9.32692212617883e-05, "loss": 4.3811, "step": 724900 }, { "epoch": 4.371329337562258, "grad_norm": 2.1314754486083984, "learning_rate": 9.325038430322222e-05, "loss": 4.9148, "step": 724950 }, { "epoch": 4.371630829343592, "grad_norm": 1.543102502822876, "learning_rate": 9.323154838905189e-05, "loss": 4.7892, "step": 725000 }, { "epoch": 4.371932321124926, "grad_norm": 0.6191011667251587, "learning_rate": 9.321271351962392e-05, "loss": 4.3717, "step": 725050 }, { "epoch": 4.37223381290626, "grad_norm": 2.584470748901367, "learning_rate": 9.319387969528485e-05, "loss": 4.5081, "step": 725100 }, { "epoch": 4.372535304687594, "grad_norm": 3.1909902095794678, "learning_rate": 9.31750469163814e-05, "loss": 4.7088, "step": 725150 }, { "epoch": 4.3728367964689285, "grad_norm": 2.3234102725982666, "learning_rate": 9.315621518326015e-05, "loss": 4.6649, "step": 725200 }, { "epoch": 4.373138288250262, "grad_norm": 2.027341365814209, "learning_rate": 9.313738449626757e-05, "loss": 4.5195, "step": 725250 }, { "epoch": 4.373439780031596, "grad_norm": 2.944190740585327, "learning_rate": 9.311855485575029e-05, "loss": 5.0731, "step": 725300 }, { "epoch": 4.37374127181293, "grad_norm": 2.1163687705993652, "learning_rate": 9.309972626205489e-05, "loss": 4.5339, "step": 725350 }, { "epoch": 4.3740427635942645, "grad_norm": 3.3805854320526123, "learning_rate": 9.30808987155278e-05, "loss": 4.7093, "step": 725400 }, { "epoch": 4.374344255375599, "grad_norm": 1.9648762941360474, "learning_rate": 9.306207221651552e-05, "loss": 4.7043, "step": 725450 }, { "epoch": 4.374645747156933, "grad_norm": 1.7468327283859253, "learning_rate": 9.304324676536458e-05, "loss": 4.8285, "step": 725500 }, { "epoch": 4.374947238938266, "grad_norm": 2.489166259765625, "learning_rate": 9.302442236242137e-05, "loss": 4.7993, "step": 725550 }, { "epoch": 4.3752487307196, "grad_norm": 1.2454092502593994, "learning_rate": 9.300559900803237e-05, "loss": 4.9842, "step": 725600 }, { "epoch": 4.375550222500935, "grad_norm": 1.6755627393722534, "learning_rate": 9.298677670254402e-05, "loss": 4.9409, "step": 725650 }, { "epoch": 4.375851714282269, "grad_norm": 1.5381758213043213, "learning_rate": 9.29679554463026e-05, "loss": 5.1891, "step": 725700 }, { "epoch": 4.376153206063603, "grad_norm": 1.9647923707962036, "learning_rate": 9.294913523965463e-05, "loss": 4.623, "step": 725750 }, { "epoch": 4.376454697844936, "grad_norm": 3.248677968978882, "learning_rate": 9.293031608294636e-05, "loss": 4.5718, "step": 725800 }, { "epoch": 4.3767561896262706, "grad_norm": 2.5702431201934814, "learning_rate": 9.291149797652428e-05, "loss": 4.4137, "step": 725850 }, { "epoch": 4.377057681407605, "grad_norm": 2.184096574783325, "learning_rate": 9.28926809207345e-05, "loss": 4.5905, "step": 725900 }, { "epoch": 4.377359173188939, "grad_norm": 2.0180044174194336, "learning_rate": 9.287386491592349e-05, "loss": 4.8285, "step": 725950 }, { "epoch": 4.377660664970273, "grad_norm": 2.5953469276428223, "learning_rate": 9.28550499624375e-05, "loss": 4.5112, "step": 726000 }, { "epoch": 4.377660664970273, "eval_loss": 5.167557716369629, "eval_runtime": 39.252, "eval_samples_per_second": 13.044, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.640573530937952, "step": 726000 }, { "epoch": 4.377962156751607, "grad_norm": 2.3578197956085205, "learning_rate": 9.28362360606227e-05, "loss": 4.5935, "step": 726050 }, { "epoch": 4.378263648532941, "grad_norm": 1.9903579950332642, "learning_rate": 9.28174232108254e-05, "loss": 4.5581, "step": 726100 }, { "epoch": 4.378565140314275, "grad_norm": 2.3900694847106934, "learning_rate": 9.279861141339187e-05, "loss": 4.6583, "step": 726150 }, { "epoch": 4.378866632095609, "grad_norm": 2.1742091178894043, "learning_rate": 9.277980066866826e-05, "loss": 4.704, "step": 726200 }, { "epoch": 4.379168123876943, "grad_norm": 3.1034436225891113, "learning_rate": 9.276099097700071e-05, "loss": 4.9005, "step": 726250 }, { "epoch": 4.3794696156582775, "grad_norm": 1.7342768907546997, "learning_rate": 9.274218233873549e-05, "loss": 4.5034, "step": 726300 }, { "epoch": 4.379771107439611, "grad_norm": 2.296762228012085, "learning_rate": 9.272337475421868e-05, "loss": 4.517, "step": 726350 }, { "epoch": 4.380072599220945, "grad_norm": 2.3879737854003906, "learning_rate": 9.27045682237964e-05, "loss": 4.5789, "step": 726400 }, { "epoch": 4.380374091002279, "grad_norm": 3.8798398971557617, "learning_rate": 9.268576274781482e-05, "loss": 4.6722, "step": 726450 }, { "epoch": 4.3806755827836135, "grad_norm": 2.1087167263031006, "learning_rate": 9.266695832661993e-05, "loss": 4.1988, "step": 726500 }, { "epoch": 4.380977074564948, "grad_norm": 3.372636079788208, "learning_rate": 9.264815496055788e-05, "loss": 4.7073, "step": 726550 }, { "epoch": 4.381278566346281, "grad_norm": 3.0764236450195312, "learning_rate": 9.262935264997473e-05, "loss": 4.3875, "step": 726600 }, { "epoch": 4.381580058127615, "grad_norm": 2.289529323577881, "learning_rate": 9.26105513952164e-05, "loss": 5.0483, "step": 726650 }, { "epoch": 4.381881549908949, "grad_norm": 2.423722982406616, "learning_rate": 9.259175119662897e-05, "loss": 5.0972, "step": 726700 }, { "epoch": 4.382183041690284, "grad_norm": 2.70343279838562, "learning_rate": 9.257295205455853e-05, "loss": 4.9051, "step": 726750 }, { "epoch": 4.382484533471618, "grad_norm": 2.620906114578247, "learning_rate": 9.255415396935088e-05, "loss": 4.9458, "step": 726800 }, { "epoch": 4.382786025252951, "grad_norm": 0.6170827150344849, "learning_rate": 9.253535694135204e-05, "loss": 4.3183, "step": 726850 }, { "epoch": 4.383087517034285, "grad_norm": 2.7400577068328857, "learning_rate": 9.2516560970908e-05, "loss": 4.9754, "step": 726900 }, { "epoch": 4.3833890088156195, "grad_norm": 3.0270087718963623, "learning_rate": 9.249776605836459e-05, "loss": 5.0811, "step": 726950 }, { "epoch": 4.383690500596954, "grad_norm": 1.9369744062423706, "learning_rate": 9.24789722040677e-05, "loss": 4.4148, "step": 727000 }, { "epoch": 4.383991992378288, "grad_norm": 3.944664239883423, "learning_rate": 9.246017940836333e-05, "loss": 4.8304, "step": 727050 }, { "epoch": 4.384293484159622, "grad_norm": 2.6746397018432617, "learning_rate": 9.244138767159717e-05, "loss": 4.886, "step": 727100 }, { "epoch": 4.3845949759409555, "grad_norm": 2.812448263168335, "learning_rate": 9.242259699411513e-05, "loss": 4.5951, "step": 727150 }, { "epoch": 4.38489646772229, "grad_norm": 1.0662102699279785, "learning_rate": 9.240380737626308e-05, "loss": 4.6182, "step": 727200 }, { "epoch": 4.385197959503624, "grad_norm": 1.6854817867279053, "learning_rate": 9.238501881838671e-05, "loss": 4.8186, "step": 727250 }, { "epoch": 4.385499451284958, "grad_norm": 2.9579079151153564, "learning_rate": 9.236623132083184e-05, "loss": 4.7977, "step": 727300 }, { "epoch": 4.385800943066292, "grad_norm": 2.1192450523376465, "learning_rate": 9.234744488394427e-05, "loss": 4.8704, "step": 727350 }, { "epoch": 4.386102434847626, "grad_norm": 2.798943042755127, "learning_rate": 9.232865950806971e-05, "loss": 4.4597, "step": 727400 }, { "epoch": 4.38640392662896, "grad_norm": 1.4135874509811401, "learning_rate": 9.230987519355382e-05, "loss": 4.8599, "step": 727450 }, { "epoch": 4.386705418410294, "grad_norm": 1.6909023523330688, "learning_rate": 9.229109194074241e-05, "loss": 4.6298, "step": 727500 }, { "epoch": 4.387006910191628, "grad_norm": 3.012303590774536, "learning_rate": 9.227230974998106e-05, "loss": 4.2974, "step": 727550 }, { "epoch": 4.3873084019729625, "grad_norm": 2.472628593444824, "learning_rate": 9.225352862161546e-05, "loss": 4.2746, "step": 727600 }, { "epoch": 4.387609893754297, "grad_norm": 2.393801212310791, "learning_rate": 9.22347485559913e-05, "loss": 4.7848, "step": 727650 }, { "epoch": 4.38791138553563, "grad_norm": 2.2793805599212646, "learning_rate": 9.221596955345412e-05, "loss": 4.4496, "step": 727700 }, { "epoch": 4.388212877316964, "grad_norm": 1.8415274620056152, "learning_rate": 9.219719161434956e-05, "loss": 5.1799, "step": 727750 }, { "epoch": 4.388514369098298, "grad_norm": 2.691713333129883, "learning_rate": 9.217841473902325e-05, "loss": 4.7268, "step": 727800 }, { "epoch": 4.388815860879633, "grad_norm": 1.7271431684494019, "learning_rate": 9.215963892782063e-05, "loss": 4.5637, "step": 727850 }, { "epoch": 4.389117352660967, "grad_norm": 3.031466007232666, "learning_rate": 9.214086418108731e-05, "loss": 4.9832, "step": 727900 }, { "epoch": 4.3894188444423, "grad_norm": 3.0507919788360596, "learning_rate": 9.212209049916888e-05, "loss": 4.9466, "step": 727950 }, { "epoch": 4.389720336223634, "grad_norm": 3.6687843799591064, "learning_rate": 9.210331788241075e-05, "loss": 5.1739, "step": 728000 }, { "epoch": 4.3900218280049685, "grad_norm": 3.247682809829712, "learning_rate": 9.20845463311584e-05, "loss": 4.5975, "step": 728050 }, { "epoch": 4.390323319786303, "grad_norm": 2.552607536315918, "learning_rate": 9.206577584575738e-05, "loss": 4.8864, "step": 728100 }, { "epoch": 4.390624811567637, "grad_norm": 2.787658452987671, "learning_rate": 9.204700642655305e-05, "loss": 4.4114, "step": 728150 }, { "epoch": 4.390926303348971, "grad_norm": 2.82926607131958, "learning_rate": 9.202823807389084e-05, "loss": 4.8845, "step": 728200 }, { "epoch": 4.3912277951303045, "grad_norm": 3.0674169063568115, "learning_rate": 9.200947078811623e-05, "loss": 4.9998, "step": 728250 }, { "epoch": 4.391529286911639, "grad_norm": 0.6170461177825928, "learning_rate": 9.199070456957451e-05, "loss": 4.8693, "step": 728300 }, { "epoch": 4.391830778692973, "grad_norm": 2.440502882003784, "learning_rate": 9.19719394186111e-05, "loss": 4.7111, "step": 728350 }, { "epoch": 4.392132270474307, "grad_norm": 2.19840931892395, "learning_rate": 9.195317533557137e-05, "loss": 4.8995, "step": 728400 }, { "epoch": 4.392433762255641, "grad_norm": 3.6024882793426514, "learning_rate": 9.193441232080055e-05, "loss": 4.6528, "step": 728450 }, { "epoch": 4.392735254036975, "grad_norm": 2.0969910621643066, "learning_rate": 9.1915650374644e-05, "loss": 4.6266, "step": 728500 }, { "epoch": 4.393036745818309, "grad_norm": 2.5734407901763916, "learning_rate": 9.18968894974471e-05, "loss": 5.0441, "step": 728550 }, { "epoch": 4.393338237599643, "grad_norm": 2.8516736030578613, "learning_rate": 9.187812968955498e-05, "loss": 4.3526, "step": 728600 }, { "epoch": 4.393639729380977, "grad_norm": 3.040524482727051, "learning_rate": 9.18593709513129e-05, "loss": 4.1518, "step": 728650 }, { "epoch": 4.3939412211623115, "grad_norm": 1.0237643718719482, "learning_rate": 9.18406132830662e-05, "loss": 4.9243, "step": 728700 }, { "epoch": 4.394242712943645, "grad_norm": 2.736860752105713, "learning_rate": 9.182185668515996e-05, "loss": 4.5664, "step": 728750 }, { "epoch": 4.394544204724979, "grad_norm": 1.86416494846344, "learning_rate": 9.180310115793945e-05, "loss": 4.5849, "step": 728800 }, { "epoch": 4.394845696506313, "grad_norm": 2.3429791927337646, "learning_rate": 9.178434670174984e-05, "loss": 4.7857, "step": 728850 }, { "epoch": 4.395147188287647, "grad_norm": 0.9514143466949463, "learning_rate": 9.176559331693619e-05, "loss": 4.5068, "step": 728900 }, { "epoch": 4.395448680068982, "grad_norm": 2.574784278869629, "learning_rate": 9.174684100384371e-05, "loss": 4.8801, "step": 728950 }, { "epoch": 4.395750171850315, "grad_norm": 2.524415969848633, "learning_rate": 9.172808976281753e-05, "loss": 4.6492, "step": 729000 }, { "epoch": 4.395750171850315, "eval_loss": 5.155083179473877, "eval_runtime": 39.2276, "eval_samples_per_second": 13.052, "eval_steps_per_second": 6.526, "eval_tts_loss": 7.682585675191986, "step": 729000 }, { "epoch": 4.396051663631649, "grad_norm": 3.0003538131713867, "learning_rate": 9.170933959420264e-05, "loss": 5.0528, "step": 729050 }, { "epoch": 4.396353155412983, "grad_norm": 3.3681039810180664, "learning_rate": 9.169059049834418e-05, "loss": 4.7896, "step": 729100 }, { "epoch": 4.3966546471943175, "grad_norm": 3.3659756183624268, "learning_rate": 9.167184247558721e-05, "loss": 4.5382, "step": 729150 }, { "epoch": 4.396956138975652, "grad_norm": 2.0545976161956787, "learning_rate": 9.165309552627675e-05, "loss": 4.8715, "step": 729200 }, { "epoch": 4.397257630756986, "grad_norm": 2.5132930278778076, "learning_rate": 9.163434965075777e-05, "loss": 4.6198, "step": 729250 }, { "epoch": 4.397559122538319, "grad_norm": 2.8963847160339355, "learning_rate": 9.161560484937531e-05, "loss": 4.6786, "step": 729300 }, { "epoch": 4.3978606143196535, "grad_norm": 3.9841549396514893, "learning_rate": 9.159686112247433e-05, "loss": 4.4434, "step": 729350 }, { "epoch": 4.398162106100988, "grad_norm": 5.082950592041016, "learning_rate": 9.157811847039976e-05, "loss": 5.067, "step": 729400 }, { "epoch": 4.398463597882322, "grad_norm": 1.3922876119613647, "learning_rate": 9.155937689349654e-05, "loss": 4.5341, "step": 729450 }, { "epoch": 4.398765089663656, "grad_norm": 3.5363306999206543, "learning_rate": 9.154063639210963e-05, "loss": 4.8967, "step": 729500 }, { "epoch": 4.399066581444989, "grad_norm": 2.3571789264678955, "learning_rate": 9.152189696658385e-05, "loss": 4.4677, "step": 729550 }, { "epoch": 4.399368073226324, "grad_norm": 3.5324435234069824, "learning_rate": 9.150315861726408e-05, "loss": 4.7043, "step": 729600 }, { "epoch": 4.399669565007658, "grad_norm": 2.963977336883545, "learning_rate": 9.148442134449526e-05, "loss": 4.4752, "step": 729650 }, { "epoch": 4.399971056788992, "grad_norm": 0.8790406584739685, "learning_rate": 9.14656851486221e-05, "loss": 4.733, "step": 729700 }, { "epoch": 4.400272548570326, "grad_norm": 2.1959948539733887, "learning_rate": 9.14469500299895e-05, "loss": 4.6138, "step": 729750 }, { "epoch": 4.40057404035166, "grad_norm": 2.527451515197754, "learning_rate": 9.142821598894225e-05, "loss": 4.7305, "step": 729800 }, { "epoch": 4.400875532132994, "grad_norm": 2.8056395053863525, "learning_rate": 9.140948302582504e-05, "loss": 4.7754, "step": 729850 }, { "epoch": 4.401177023914328, "grad_norm": 2.7161026000976562, "learning_rate": 9.13907511409827e-05, "loss": 4.7008, "step": 729900 }, { "epoch": 4.401478515695662, "grad_norm": 1.7175933122634888, "learning_rate": 9.137202033475998e-05, "loss": 4.5783, "step": 729950 }, { "epoch": 4.401780007476996, "grad_norm": 1.7261857986450195, "learning_rate": 9.135329060750154e-05, "loss": 4.661, "step": 730000 }, { "epoch": 4.402081499258331, "grad_norm": 2.742246150970459, "learning_rate": 9.133456195955206e-05, "loss": 4.6884, "step": 730050 }, { "epoch": 4.402382991039664, "grad_norm": 1.419171690940857, "learning_rate": 9.131583439125631e-05, "loss": 4.6642, "step": 730100 }, { "epoch": 4.402684482820998, "grad_norm": 2.918523073196411, "learning_rate": 9.129710790295884e-05, "loss": 4.9016, "step": 730150 }, { "epoch": 4.402985974602332, "grad_norm": 2.2235803604125977, "learning_rate": 9.127838249500433e-05, "loss": 4.4728, "step": 730200 }, { "epoch": 4.4032874663836665, "grad_norm": 2.3665201663970947, "learning_rate": 9.125965816773741e-05, "loss": 4.6238, "step": 730250 }, { "epoch": 4.403588958165001, "grad_norm": 2.6487395763397217, "learning_rate": 9.124093492150262e-05, "loss": 4.4159, "step": 730300 }, { "epoch": 4.403890449946334, "grad_norm": 2.8708624839782715, "learning_rate": 9.122221275664459e-05, "loss": 5.1068, "step": 730350 }, { "epoch": 4.404191941727668, "grad_norm": 2.831989049911499, "learning_rate": 9.120349167350787e-05, "loss": 4.3579, "step": 730400 }, { "epoch": 4.4044934335090025, "grad_norm": 3.1682446002960205, "learning_rate": 9.118477167243691e-05, "loss": 4.6301, "step": 730450 }, { "epoch": 4.404794925290337, "grad_norm": 3.097691535949707, "learning_rate": 9.116605275377634e-05, "loss": 4.5757, "step": 730500 }, { "epoch": 4.405096417071671, "grad_norm": 6.669034004211426, "learning_rate": 9.114733491787063e-05, "loss": 4.4717, "step": 730550 }, { "epoch": 4.405397908853004, "grad_norm": 2.6532886028289795, "learning_rate": 9.112861816506422e-05, "loss": 4.6832, "step": 730600 }, { "epoch": 4.405699400634338, "grad_norm": 2.7228143215179443, "learning_rate": 9.110990249570154e-05, "loss": 4.6983, "step": 730650 }, { "epoch": 4.406000892415673, "grad_norm": 4.240285873413086, "learning_rate": 9.109118791012715e-05, "loss": 5.0237, "step": 730700 }, { "epoch": 4.406302384197007, "grad_norm": 3.2182371616363525, "learning_rate": 9.107247440868532e-05, "loss": 4.4115, "step": 730750 }, { "epoch": 4.406603875978341, "grad_norm": 2.5155861377716064, "learning_rate": 9.105376199172048e-05, "loss": 4.5867, "step": 730800 }, { "epoch": 4.406905367759675, "grad_norm": 3.1187186241149902, "learning_rate": 9.103505065957709e-05, "loss": 4.4408, "step": 730850 }, { "epoch": 4.4072068595410085, "grad_norm": 1.5863568782806396, "learning_rate": 9.101634041259939e-05, "loss": 4.5355, "step": 730900 }, { "epoch": 4.407508351322343, "grad_norm": 2.528533458709717, "learning_rate": 9.099763125113181e-05, "loss": 4.8864, "step": 730950 }, { "epoch": 4.407809843103677, "grad_norm": 1.8105577230453491, "learning_rate": 9.097892317551864e-05, "loss": 4.5768, "step": 731000 }, { "epoch": 4.408111334885011, "grad_norm": 3.3250253200531006, "learning_rate": 9.09602161861041e-05, "loss": 4.6037, "step": 731050 }, { "epoch": 4.408412826666345, "grad_norm": 2.3776819705963135, "learning_rate": 9.094151028323252e-05, "loss": 4.5762, "step": 731100 }, { "epoch": 4.408714318447679, "grad_norm": 2.1212594509124756, "learning_rate": 9.092280546724825e-05, "loss": 4.3774, "step": 731150 }, { "epoch": 4.409015810229013, "grad_norm": 2.8042144775390625, "learning_rate": 9.090410173849539e-05, "loss": 4.2687, "step": 731200 }, { "epoch": 4.409317302010347, "grad_norm": 3.151850461959839, "learning_rate": 9.088539909731816e-05, "loss": 4.6615, "step": 731250 }, { "epoch": 4.409618793791681, "grad_norm": 1.6075741052627563, "learning_rate": 9.086669754406089e-05, "loss": 5.0177, "step": 731300 }, { "epoch": 4.4099202855730155, "grad_norm": 5.634591102600098, "learning_rate": 9.08479970790676e-05, "loss": 4.7364, "step": 731350 }, { "epoch": 4.41022177735435, "grad_norm": 3.069274663925171, "learning_rate": 9.082929770268251e-05, "loss": 4.8684, "step": 731400 }, { "epoch": 4.410523269135683, "grad_norm": 2.6808886528015137, "learning_rate": 9.08105994152498e-05, "loss": 4.8248, "step": 731450 }, { "epoch": 4.410824760917017, "grad_norm": 2.899038553237915, "learning_rate": 9.079190221711348e-05, "loss": 5.0796, "step": 731500 }, { "epoch": 4.4111262526983515, "grad_norm": 2.013148307800293, "learning_rate": 9.077320610861774e-05, "loss": 4.5538, "step": 731550 }, { "epoch": 4.411427744479686, "grad_norm": 3.256438732147217, "learning_rate": 9.075451109010666e-05, "loss": 4.2909, "step": 731600 }, { "epoch": 4.41172923626102, "grad_norm": 4.0654401779174805, "learning_rate": 9.073581716192416e-05, "loss": 4.8548, "step": 731650 }, { "epoch": 4.412030728042353, "grad_norm": 2.8267476558685303, "learning_rate": 9.07171243244144e-05, "loss": 4.8706, "step": 731700 }, { "epoch": 4.412332219823687, "grad_norm": 0.5863635540008545, "learning_rate": 9.069843257792143e-05, "loss": 4.1355, "step": 731750 }, { "epoch": 4.412633711605022, "grad_norm": 1.7532801628112793, "learning_rate": 9.067974192278913e-05, "loss": 4.8948, "step": 731800 }, { "epoch": 4.412935203386356, "grad_norm": 3.6218485832214355, "learning_rate": 9.066105235936152e-05, "loss": 4.9577, "step": 731850 }, { "epoch": 4.41323669516769, "grad_norm": 3.3840014934539795, "learning_rate": 9.064236388798259e-05, "loss": 4.5096, "step": 731900 }, { "epoch": 4.413538186949024, "grad_norm": 0.6442146897315979, "learning_rate": 9.062367650899625e-05, "loss": 4.6501, "step": 731950 }, { "epoch": 4.4138396787303575, "grad_norm": 2.8033368587493896, "learning_rate": 9.060499022274633e-05, "loss": 4.3856, "step": 732000 }, { "epoch": 4.4138396787303575, "eval_loss": 5.160950660705566, "eval_runtime": 39.0225, "eval_samples_per_second": 13.121, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.7211393272604605, "step": 732000 }, { "epoch": 4.414141170511692, "grad_norm": 3.1485536098480225, "learning_rate": 9.058630502957692e-05, "loss": 4.7963, "step": 732050 }, { "epoch": 4.414442662293026, "grad_norm": 1.3078370094299316, "learning_rate": 9.056762092983167e-05, "loss": 4.4838, "step": 732100 }, { "epoch": 4.41474415407436, "grad_norm": 2.663663387298584, "learning_rate": 9.054893792385459e-05, "loss": 4.8554, "step": 732150 }, { "epoch": 4.415045645855694, "grad_norm": 3.1166279315948486, "learning_rate": 9.053025601198952e-05, "loss": 4.9424, "step": 732200 }, { "epoch": 4.415347137637028, "grad_norm": 2.2913594245910645, "learning_rate": 9.051157519458012e-05, "loss": 4.4805, "step": 732250 }, { "epoch": 4.415648629418362, "grad_norm": 3.1822879314422607, "learning_rate": 9.049289547197031e-05, "loss": 4.7553, "step": 732300 }, { "epoch": 4.415950121199696, "grad_norm": 3.3875033855438232, "learning_rate": 9.04742168445039e-05, "loss": 4.2603, "step": 732350 }, { "epoch": 4.41625161298103, "grad_norm": 2.526395797729492, "learning_rate": 9.045553931252453e-05, "loss": 4.5428, "step": 732400 }, { "epoch": 4.4165531047623645, "grad_norm": 3.885284662246704, "learning_rate": 9.043686287637598e-05, "loss": 5.0709, "step": 732450 }, { "epoch": 4.416854596543698, "grad_norm": 3.4177918434143066, "learning_rate": 9.041818753640202e-05, "loss": 4.7827, "step": 732500 }, { "epoch": 4.417156088325032, "grad_norm": 2.657383680343628, "learning_rate": 9.039951329294623e-05, "loss": 4.7253, "step": 732550 }, { "epoch": 4.417457580106366, "grad_norm": 2.2711081504821777, "learning_rate": 9.038084014635235e-05, "loss": 5.1121, "step": 732600 }, { "epoch": 4.4177590718877005, "grad_norm": 3.045454740524292, "learning_rate": 9.036216809696403e-05, "loss": 4.8823, "step": 732650 }, { "epoch": 4.418060563669035, "grad_norm": 1.1965452432632446, "learning_rate": 9.034349714512492e-05, "loss": 4.8168, "step": 732700 }, { "epoch": 4.418362055450368, "grad_norm": 2.5595784187316895, "learning_rate": 9.032482729117861e-05, "loss": 4.574, "step": 732750 }, { "epoch": 4.418663547231702, "grad_norm": 2.4944732189178467, "learning_rate": 9.030615853546866e-05, "loss": 4.6073, "step": 732800 }, { "epoch": 4.418965039013036, "grad_norm": 2.4573662281036377, "learning_rate": 9.028749087833873e-05, "loss": 4.4555, "step": 732850 }, { "epoch": 4.419266530794371, "grad_norm": 2.6630680561065674, "learning_rate": 9.026882432013225e-05, "loss": 4.9377, "step": 732900 }, { "epoch": 4.419568022575705, "grad_norm": 3.2839677333831787, "learning_rate": 9.025015886119286e-05, "loss": 4.9204, "step": 732950 }, { "epoch": 4.419869514357039, "grad_norm": 4.125365257263184, "learning_rate": 9.023149450186404e-05, "loss": 4.5213, "step": 733000 }, { "epoch": 4.420171006138372, "grad_norm": 2.2882182598114014, "learning_rate": 9.021283124248924e-05, "loss": 4.633, "step": 733050 }, { "epoch": 4.4204724979197065, "grad_norm": 2.933661937713623, "learning_rate": 9.019416908341197e-05, "loss": 5.0203, "step": 733100 }, { "epoch": 4.420773989701041, "grad_norm": 4.007972717285156, "learning_rate": 9.017550802497571e-05, "loss": 4.6095, "step": 733150 }, { "epoch": 4.421075481482375, "grad_norm": 3.7316606044769287, "learning_rate": 9.015684806752377e-05, "loss": 4.768, "step": 733200 }, { "epoch": 4.421376973263709, "grad_norm": 2.1541404724121094, "learning_rate": 9.013818921139966e-05, "loss": 4.0075, "step": 733250 }, { "epoch": 4.4216784650450425, "grad_norm": 2.335796356201172, "learning_rate": 9.01195314569468e-05, "loss": 4.342, "step": 733300 }, { "epoch": 4.421979956826377, "grad_norm": 3.3615429401397705, "learning_rate": 9.010087480450849e-05, "loss": 4.7082, "step": 733350 }, { "epoch": 4.422281448607711, "grad_norm": 3.729644775390625, "learning_rate": 9.008221925442806e-05, "loss": 4.9078, "step": 733400 }, { "epoch": 4.422582940389045, "grad_norm": 2.5042667388916016, "learning_rate": 9.006356480704894e-05, "loss": 4.9024, "step": 733450 }, { "epoch": 4.422884432170379, "grad_norm": 2.3630940914154053, "learning_rate": 9.004491146271431e-05, "loss": 4.6815, "step": 733500 }, { "epoch": 4.4231859239517135, "grad_norm": 3.8230600357055664, "learning_rate": 9.002625922176755e-05, "loss": 4.3803, "step": 733550 }, { "epoch": 4.423487415733047, "grad_norm": 2.599881649017334, "learning_rate": 9.000760808455196e-05, "loss": 5.0476, "step": 733600 }, { "epoch": 4.423788907514381, "grad_norm": 1.5492150783538818, "learning_rate": 8.998895805141064e-05, "loss": 4.5764, "step": 733650 }, { "epoch": 4.424090399295715, "grad_norm": 2.081578254699707, "learning_rate": 8.997030912268696e-05, "loss": 4.3159, "step": 733700 }, { "epoch": 4.424391891077049, "grad_norm": 2.2614808082580566, "learning_rate": 8.995166129872408e-05, "loss": 4.464, "step": 733750 }, { "epoch": 4.424693382858384, "grad_norm": 3.3845012187957764, "learning_rate": 8.993301457986513e-05, "loss": 5.2022, "step": 733800 }, { "epoch": 4.424994874639717, "grad_norm": 3.7302629947662354, "learning_rate": 8.991436896645333e-05, "loss": 4.6091, "step": 733850 }, { "epoch": 4.425296366421051, "grad_norm": 3.193152666091919, "learning_rate": 8.989572445883188e-05, "loss": 4.7623, "step": 733900 }, { "epoch": 4.425597858202385, "grad_norm": 2.1024255752563477, "learning_rate": 8.987708105734381e-05, "loss": 4.7059, "step": 733950 }, { "epoch": 4.42589934998372, "grad_norm": 2.666133403778076, "learning_rate": 8.985843876233225e-05, "loss": 4.8385, "step": 734000 }, { "epoch": 4.426200841765054, "grad_norm": 2.058633327484131, "learning_rate": 8.983979757414035e-05, "loss": 4.3978, "step": 734050 }, { "epoch": 4.426502333546387, "grad_norm": 2.0811386108398438, "learning_rate": 8.982115749311105e-05, "loss": 5.4677, "step": 734100 }, { "epoch": 4.426803825327721, "grad_norm": 4.339588642120361, "learning_rate": 8.980251851958753e-05, "loss": 4.8474, "step": 734150 }, { "epoch": 4.4271053171090555, "grad_norm": 4.855749607086182, "learning_rate": 8.978388065391274e-05, "loss": 4.8566, "step": 734200 }, { "epoch": 4.42740680889039, "grad_norm": 3.0064291954040527, "learning_rate": 8.976524389642966e-05, "loss": 4.6878, "step": 734250 }, { "epoch": 4.427708300671724, "grad_norm": 2.4126133918762207, "learning_rate": 8.974660824748132e-05, "loss": 4.7035, "step": 734300 }, { "epoch": 4.428009792453057, "grad_norm": 3.6844844818115234, "learning_rate": 8.972797370741068e-05, "loss": 4.9312, "step": 734350 }, { "epoch": 4.4283112842343915, "grad_norm": 3.9642889499664307, "learning_rate": 8.970934027656068e-05, "loss": 4.7276, "step": 734400 }, { "epoch": 4.428612776015726, "grad_norm": 4.526881694793701, "learning_rate": 8.969070795527417e-05, "loss": 4.8146, "step": 734450 }, { "epoch": 4.42891426779706, "grad_norm": 1.835123896598816, "learning_rate": 8.967207674389422e-05, "loss": 4.5429, "step": 734500 }, { "epoch": 4.429215759578394, "grad_norm": 2.745274066925049, "learning_rate": 8.965344664276354e-05, "loss": 4.5202, "step": 734550 }, { "epoch": 4.429517251359728, "grad_norm": 1.379771113395691, "learning_rate": 8.963481765222504e-05, "loss": 4.8888, "step": 734600 }, { "epoch": 4.429818743141062, "grad_norm": 3.0571553707122803, "learning_rate": 8.961618977262164e-05, "loss": 4.9393, "step": 734650 }, { "epoch": 4.430120234922396, "grad_norm": 1.1015130281448364, "learning_rate": 8.959756300429602e-05, "loss": 4.6651, "step": 734700 }, { "epoch": 4.43042172670373, "grad_norm": 3.5394253730773926, "learning_rate": 8.957893734759108e-05, "loss": 4.4043, "step": 734750 }, { "epoch": 4.430723218485064, "grad_norm": 0.8000876307487488, "learning_rate": 8.956031280284961e-05, "loss": 4.5484, "step": 734800 }, { "epoch": 4.431024710266398, "grad_norm": 3.3264641761779785, "learning_rate": 8.954168937041428e-05, "loss": 4.8017, "step": 734850 }, { "epoch": 4.431326202047732, "grad_norm": 3.1458632946014404, "learning_rate": 8.95230670506279e-05, "loss": 4.8982, "step": 734900 }, { "epoch": 4.431627693829066, "grad_norm": 2.916969060897827, "learning_rate": 8.950444584383317e-05, "loss": 4.5959, "step": 734950 }, { "epoch": 4.4319291856104, "grad_norm": 2.4612011909484863, "learning_rate": 8.948582575037277e-05, "loss": 4.7759, "step": 735000 }, { "epoch": 4.4319291856104, "eval_loss": 5.148280620574951, "eval_runtime": 39.165, "eval_samples_per_second": 13.073, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.673380095902549, "step": 735000 }, { "epoch": 4.432230677391734, "grad_norm": 3.251333475112915, "learning_rate": 8.946720677058938e-05, "loss": 4.5408, "step": 735050 }, { "epoch": 4.432532169173069, "grad_norm": 2.0458104610443115, "learning_rate": 8.944858890482573e-05, "loss": 4.7679, "step": 735100 }, { "epoch": 4.432833660954403, "grad_norm": 2.5435116291046143, "learning_rate": 8.942997215342432e-05, "loss": 4.5583, "step": 735150 }, { "epoch": 4.433135152735736, "grad_norm": 3.0043132305145264, "learning_rate": 8.941135651672783e-05, "loss": 4.8106, "step": 735200 }, { "epoch": 4.43343664451707, "grad_norm": 2.0823514461517334, "learning_rate": 8.939274199507893e-05, "loss": 4.5949, "step": 735250 }, { "epoch": 4.4337381362984045, "grad_norm": 3.6152024269104004, "learning_rate": 8.937412858882004e-05, "loss": 5.3615, "step": 735300 }, { "epoch": 4.434039628079739, "grad_norm": 2.8636560440063477, "learning_rate": 8.935551629829383e-05, "loss": 4.9305, "step": 735350 }, { "epoch": 4.434341119861073, "grad_norm": 2.3786675930023193, "learning_rate": 8.933690512384283e-05, "loss": 4.9897, "step": 735400 }, { "epoch": 4.434642611642406, "grad_norm": 1.1821208000183105, "learning_rate": 8.931829506580947e-05, "loss": 5.0417, "step": 735450 }, { "epoch": 4.4349441034237405, "grad_norm": 1.9238126277923584, "learning_rate": 8.92996861245363e-05, "loss": 4.3448, "step": 735500 }, { "epoch": 4.435245595205075, "grad_norm": 3.0925657749176025, "learning_rate": 8.928107830036581e-05, "loss": 4.4127, "step": 735550 }, { "epoch": 4.435547086986409, "grad_norm": 2.732633590698242, "learning_rate": 8.92624715936404e-05, "loss": 4.4298, "step": 735600 }, { "epoch": 4.435848578767743, "grad_norm": 3.56687593460083, "learning_rate": 8.924386600470252e-05, "loss": 4.572, "step": 735650 }, { "epoch": 4.436150070549077, "grad_norm": 3.6724050045013428, "learning_rate": 8.922526153389463e-05, "loss": 4.4295, "step": 735700 }, { "epoch": 4.436451562330411, "grad_norm": 3.318871259689331, "learning_rate": 8.920665818155902e-05, "loss": 4.8208, "step": 735750 }, { "epoch": 4.436753054111745, "grad_norm": 3.044710636138916, "learning_rate": 8.918805594803811e-05, "loss": 4.9472, "step": 735800 }, { "epoch": 4.437054545893079, "grad_norm": 4.204291343688965, "learning_rate": 8.916945483367428e-05, "loss": 4.5143, "step": 735850 }, { "epoch": 4.437356037674413, "grad_norm": 2.557617425918579, "learning_rate": 8.915085483880979e-05, "loss": 4.6528, "step": 735900 }, { "epoch": 4.437657529455747, "grad_norm": 3.1083078384399414, "learning_rate": 8.9132255963787e-05, "loss": 4.4702, "step": 735950 }, { "epoch": 4.437959021237081, "grad_norm": 3.0168356895446777, "learning_rate": 8.911365820894813e-05, "loss": 4.7533, "step": 736000 }, { "epoch": 4.438260513018415, "grad_norm": 2.0611941814422607, "learning_rate": 8.909506157463558e-05, "loss": 4.693, "step": 736050 }, { "epoch": 4.438562004799749, "grad_norm": 1.5781112909317017, "learning_rate": 8.907646606119142e-05, "loss": 4.4637, "step": 736100 }, { "epoch": 4.438863496581083, "grad_norm": 5.569937705993652, "learning_rate": 8.9057871668958e-05, "loss": 4.6803, "step": 736150 }, { "epoch": 4.439164988362418, "grad_norm": 2.6510446071624756, "learning_rate": 8.903927839827753e-05, "loss": 4.4644, "step": 736200 }, { "epoch": 4.439466480143751, "grad_norm": 3.80380916595459, "learning_rate": 8.902068624949206e-05, "loss": 4.8856, "step": 736250 }, { "epoch": 4.439767971925085, "grad_norm": 2.8884479999542236, "learning_rate": 8.90020952229439e-05, "loss": 5.0403, "step": 736300 }, { "epoch": 4.440069463706419, "grad_norm": 3.3748221397399902, "learning_rate": 8.898350531897511e-05, "loss": 4.4915, "step": 736350 }, { "epoch": 4.4403709554877535, "grad_norm": 2.2878410816192627, "learning_rate": 8.89649165379278e-05, "loss": 4.6951, "step": 736400 }, { "epoch": 4.440672447269088, "grad_norm": 2.4961938858032227, "learning_rate": 8.894632888014409e-05, "loss": 4.8083, "step": 736450 }, { "epoch": 4.440973939050421, "grad_norm": 2.7630584239959717, "learning_rate": 8.892774234596614e-05, "loss": 4.8613, "step": 736500 }, { "epoch": 4.441275430831755, "grad_norm": 2.074784517288208, "learning_rate": 8.890915693573589e-05, "loss": 4.497, "step": 736550 }, { "epoch": 4.4415769226130895, "grad_norm": 4.0063157081604, "learning_rate": 8.88905726497954e-05, "loss": 4.4442, "step": 736600 }, { "epoch": 4.441878414394424, "grad_norm": 2.064552068710327, "learning_rate": 8.887198948848679e-05, "loss": 4.4953, "step": 736650 }, { "epoch": 4.442179906175758, "grad_norm": 1.8558019399642944, "learning_rate": 8.88534074521519e-05, "loss": 4.5001, "step": 736700 }, { "epoch": 4.442481397957092, "grad_norm": 2.4097177982330322, "learning_rate": 8.883482654113283e-05, "loss": 4.9483, "step": 736750 }, { "epoch": 4.442782889738425, "grad_norm": 2.6072232723236084, "learning_rate": 8.881624675577152e-05, "loss": 5.0738, "step": 736800 }, { "epoch": 4.44308438151976, "grad_norm": 2.1954503059387207, "learning_rate": 8.87976680964098e-05, "loss": 5.3168, "step": 736850 }, { "epoch": 4.443385873301094, "grad_norm": 1.6901317834854126, "learning_rate": 8.877909056338969e-05, "loss": 4.78, "step": 736900 }, { "epoch": 4.443687365082428, "grad_norm": 2.99505352973938, "learning_rate": 8.87605141570531e-05, "loss": 5.0423, "step": 736950 }, { "epoch": 4.443988856863762, "grad_norm": 3.2356812953948975, "learning_rate": 8.874193887774177e-05, "loss": 4.9255, "step": 737000 }, { "epoch": 4.4442903486450955, "grad_norm": 3.1814334392547607, "learning_rate": 8.872336472579764e-05, "loss": 4.7796, "step": 737050 }, { "epoch": 4.44459184042643, "grad_norm": 1.369710922241211, "learning_rate": 8.870479170156262e-05, "loss": 4.7712, "step": 737100 }, { "epoch": 4.444893332207764, "grad_norm": 1.2686833143234253, "learning_rate": 8.868621980537838e-05, "loss": 4.6821, "step": 737150 }, { "epoch": 4.445194823989098, "grad_norm": 0.6496333479881287, "learning_rate": 8.866764903758675e-05, "loss": 4.6777, "step": 737200 }, { "epoch": 4.445496315770432, "grad_norm": 2.459160566329956, "learning_rate": 8.864907939852955e-05, "loss": 5.0909, "step": 737250 }, { "epoch": 4.445797807551767, "grad_norm": 3.778623580932617, "learning_rate": 8.863051088854845e-05, "loss": 5.1601, "step": 737300 }, { "epoch": 4.4460992993331, "grad_norm": 2.5275168418884277, "learning_rate": 8.861194350798524e-05, "loss": 4.7852, "step": 737350 }, { "epoch": 4.446400791114434, "grad_norm": 3.4159045219421387, "learning_rate": 8.859337725718164e-05, "loss": 4.314, "step": 737400 }, { "epoch": 4.446702282895768, "grad_norm": 3.9290738105773926, "learning_rate": 8.857481213647924e-05, "loss": 4.5524, "step": 737450 }, { "epoch": 4.4470037746771025, "grad_norm": 1.2768449783325195, "learning_rate": 8.855624814621978e-05, "loss": 4.7607, "step": 737500 }, { "epoch": 4.447305266458437, "grad_norm": 2.670677900314331, "learning_rate": 8.853768528674494e-05, "loss": 4.2602, "step": 737550 }, { "epoch": 4.44760675823977, "grad_norm": 1.6861268281936646, "learning_rate": 8.85191235583962e-05, "loss": 4.9266, "step": 737600 }, { "epoch": 4.447908250021104, "grad_norm": 1.135789394378662, "learning_rate": 8.850056296151529e-05, "loss": 4.2564, "step": 737650 }, { "epoch": 4.448209741802438, "grad_norm": 2.8916428089141846, "learning_rate": 8.84820034964438e-05, "loss": 4.7426, "step": 737700 }, { "epoch": 4.448511233583773, "grad_norm": 2.312823534011841, "learning_rate": 8.84634451635232e-05, "loss": 4.945, "step": 737750 }, { "epoch": 4.448812725365107, "grad_norm": 3.078472375869751, "learning_rate": 8.844488796309507e-05, "loss": 4.7835, "step": 737800 }, { "epoch": 4.44911421714644, "grad_norm": 0.6682028770446777, "learning_rate": 8.842633189550099e-05, "loss": 4.8861, "step": 737850 }, { "epoch": 4.449415708927774, "grad_norm": 1.9522459506988525, "learning_rate": 8.840777696108233e-05, "loss": 4.7443, "step": 737900 }, { "epoch": 4.449717200709109, "grad_norm": 3.1438393592834473, "learning_rate": 8.838922316018068e-05, "loss": 4.1894, "step": 737950 }, { "epoch": 4.450018692490443, "grad_norm": 2.751772403717041, "learning_rate": 8.837067049313746e-05, "loss": 4.8248, "step": 738000 }, { "epoch": 4.450018692490443, "eval_loss": 5.150801658630371, "eval_runtime": 39.0256, "eval_samples_per_second": 13.12, "eval_steps_per_second": 6.56, "eval_tts_loss": 7.668062559420165, "step": 738000 }, { "epoch": 4.450320184271777, "grad_norm": 2.28674578666687, "learning_rate": 8.835211896029405e-05, "loss": 4.4121, "step": 738050 }, { "epoch": 4.45062167605311, "grad_norm": 1.0844885110855103, "learning_rate": 8.833356856199196e-05, "loss": 4.8319, "step": 738100 }, { "epoch": 4.4509231678344445, "grad_norm": 2.2321736812591553, "learning_rate": 8.831501929857254e-05, "loss": 4.626, "step": 738150 }, { "epoch": 4.451224659615779, "grad_norm": 3.3985066413879395, "learning_rate": 8.829647117037711e-05, "loss": 4.6054, "step": 738200 }, { "epoch": 4.451526151397113, "grad_norm": 2.508453130722046, "learning_rate": 8.82779241777471e-05, "loss": 4.5303, "step": 738250 }, { "epoch": 4.451827643178447, "grad_norm": 2.3660733699798584, "learning_rate": 8.825937832102385e-05, "loss": 4.4974, "step": 738300 }, { "epoch": 4.452129134959781, "grad_norm": 1.655153751373291, "learning_rate": 8.824083360054861e-05, "loss": 4.4613, "step": 738350 }, { "epoch": 4.452430626741115, "grad_norm": 3.4782841205596924, "learning_rate": 8.822229001666265e-05, "loss": 4.6222, "step": 738400 }, { "epoch": 4.452732118522449, "grad_norm": 3.6143553256988525, "learning_rate": 8.820374756970736e-05, "loss": 4.6251, "step": 738450 }, { "epoch": 4.453033610303783, "grad_norm": 1.321074366569519, "learning_rate": 8.818520626002383e-05, "loss": 4.1667, "step": 738500 }, { "epoch": 4.453335102085117, "grad_norm": 1.3639607429504395, "learning_rate": 8.816666608795341e-05, "loss": 5.1938, "step": 738550 }, { "epoch": 4.4536365938664515, "grad_norm": 2.1624536514282227, "learning_rate": 8.814812705383728e-05, "loss": 4.6813, "step": 738600 }, { "epoch": 4.453938085647785, "grad_norm": 1.865592122077942, "learning_rate": 8.812958915801654e-05, "loss": 4.7858, "step": 738650 }, { "epoch": 4.454239577429119, "grad_norm": 2.4939770698547363, "learning_rate": 8.811105240083246e-05, "loss": 4.7424, "step": 738700 }, { "epoch": 4.454541069210453, "grad_norm": 1.9971508979797363, "learning_rate": 8.809251678262616e-05, "loss": 4.5587, "step": 738750 }, { "epoch": 4.454842560991787, "grad_norm": 2.487569808959961, "learning_rate": 8.807398230373868e-05, "loss": 4.6514, "step": 738800 }, { "epoch": 4.455144052773122, "grad_norm": 2.746413469314575, "learning_rate": 8.80554489645112e-05, "loss": 4.467, "step": 738850 }, { "epoch": 4.455445544554456, "grad_norm": 2.964303731918335, "learning_rate": 8.803691676528482e-05, "loss": 4.7125, "step": 738900 }, { "epoch": 4.455747036335789, "grad_norm": 2.4974348545074463, "learning_rate": 8.801838570640054e-05, "loss": 4.4704, "step": 738950 }, { "epoch": 4.456048528117123, "grad_norm": 2.308067560195923, "learning_rate": 8.799985578819939e-05, "loss": 4.5342, "step": 739000 }, { "epoch": 4.456350019898458, "grad_norm": 1.8311940431594849, "learning_rate": 8.798132701102248e-05, "loss": 4.4853, "step": 739050 }, { "epoch": 4.456651511679792, "grad_norm": 4.0856614112854, "learning_rate": 8.796279937521067e-05, "loss": 4.8304, "step": 739100 }, { "epoch": 4.456953003461126, "grad_norm": 2.88189959526062, "learning_rate": 8.794427288110504e-05, "loss": 4.7495, "step": 739150 }, { "epoch": 4.457254495242459, "grad_norm": 5.925548076629639, "learning_rate": 8.792574752904655e-05, "loss": 4.3448, "step": 739200 }, { "epoch": 4.4575559870237935, "grad_norm": 2.3996846675872803, "learning_rate": 8.790722331937601e-05, "loss": 4.7522, "step": 739250 }, { "epoch": 4.457857478805128, "grad_norm": 4.492896556854248, "learning_rate": 8.788870025243446e-05, "loss": 4.4809, "step": 739300 }, { "epoch": 4.458158970586462, "grad_norm": 3.8975038528442383, "learning_rate": 8.787017832856273e-05, "loss": 4.8963, "step": 739350 }, { "epoch": 4.458460462367796, "grad_norm": 3.436450958251953, "learning_rate": 8.785165754810174e-05, "loss": 4.6586, "step": 739400 }, { "epoch": 4.45876195414913, "grad_norm": 1.9447040557861328, "learning_rate": 8.783313791139224e-05, "loss": 4.7106, "step": 739450 }, { "epoch": 4.459063445930464, "grad_norm": 2.8148882389068604, "learning_rate": 8.781461941877514e-05, "loss": 4.7563, "step": 739500 }, { "epoch": 4.459364937711798, "grad_norm": 0.7213157415390015, "learning_rate": 8.77961020705913e-05, "loss": 4.5541, "step": 739550 }, { "epoch": 4.459666429493132, "grad_norm": 3.6166272163391113, "learning_rate": 8.777758586718133e-05, "loss": 4.685, "step": 739600 }, { "epoch": 4.459967921274466, "grad_norm": 2.4036364555358887, "learning_rate": 8.775907080888612e-05, "loss": 4.7668, "step": 739650 }, { "epoch": 4.4602694130558005, "grad_norm": 2.446185350418091, "learning_rate": 8.774055689604643e-05, "loss": 4.7916, "step": 739700 }, { "epoch": 4.460570904837134, "grad_norm": 3.3508567810058594, "learning_rate": 8.772204412900293e-05, "loss": 4.6671, "step": 739750 }, { "epoch": 4.460872396618468, "grad_norm": 1.9922592639923096, "learning_rate": 8.770353250809629e-05, "loss": 4.6002, "step": 739800 }, { "epoch": 4.461173888399802, "grad_norm": 2.782564878463745, "learning_rate": 8.76850220336673e-05, "loss": 4.8875, "step": 739850 }, { "epoch": 4.461475380181136, "grad_norm": 2.580404043197632, "learning_rate": 8.766651270605651e-05, "loss": 4.2538, "step": 739900 }, { "epoch": 4.461776871962471, "grad_norm": 1.2401587963104248, "learning_rate": 8.764800452560458e-05, "loss": 4.675, "step": 739950 }, { "epoch": 4.462078363743804, "grad_norm": 2.212029218673706, "learning_rate": 8.76294974926522e-05, "loss": 4.6079, "step": 740000 }, { "epoch": 4.462379855525138, "grad_norm": 2.2942543029785156, "learning_rate": 8.761099160753985e-05, "loss": 4.6242, "step": 740050 }, { "epoch": 4.462681347306472, "grad_norm": 2.861006259918213, "learning_rate": 8.759248687060821e-05, "loss": 5.0259, "step": 740100 }, { "epoch": 4.462982839087807, "grad_norm": 2.9293019771575928, "learning_rate": 8.757398328219781e-05, "loss": 4.4948, "step": 740150 }, { "epoch": 4.463284330869141, "grad_norm": 3.0895674228668213, "learning_rate": 8.75554808426491e-05, "loss": 4.5149, "step": 740200 }, { "epoch": 4.463585822650474, "grad_norm": 1.8879404067993164, "learning_rate": 8.753697955230266e-05, "loss": 4.2915, "step": 740250 }, { "epoch": 4.463887314431808, "grad_norm": 3.2712180614471436, "learning_rate": 8.7518479411499e-05, "loss": 4.9185, "step": 740300 }, { "epoch": 4.4641888062131425, "grad_norm": 2.2764439582824707, "learning_rate": 8.749998042057853e-05, "loss": 4.6933, "step": 740350 }, { "epoch": 4.464490297994477, "grad_norm": 3.349311351776123, "learning_rate": 8.748148257988177e-05, "loss": 4.4617, "step": 740400 }, { "epoch": 4.464791789775811, "grad_norm": 2.774580955505371, "learning_rate": 8.746298588974914e-05, "loss": 4.7882, "step": 740450 }, { "epoch": 4.465093281557145, "grad_norm": 5.5028076171875, "learning_rate": 8.74444903505209e-05, "loss": 4.4395, "step": 740500 }, { "epoch": 4.4653947733384785, "grad_norm": 3.938523769378662, "learning_rate": 8.742599596253758e-05, "loss": 4.4579, "step": 740550 }, { "epoch": 4.465696265119813, "grad_norm": 3.203143358230591, "learning_rate": 8.740750272613955e-05, "loss": 4.6707, "step": 740600 }, { "epoch": 4.465997756901147, "grad_norm": 3.492706537246704, "learning_rate": 8.738901064166707e-05, "loss": 4.9082, "step": 740650 }, { "epoch": 4.466299248682481, "grad_norm": 3.8651392459869385, "learning_rate": 8.737051970946046e-05, "loss": 4.7023, "step": 740700 }, { "epoch": 4.466600740463815, "grad_norm": 4.106046199798584, "learning_rate": 8.735202992986014e-05, "loss": 4.0848, "step": 740750 }, { "epoch": 4.466902232245149, "grad_norm": 1.068476676940918, "learning_rate": 8.733354130320625e-05, "loss": 4.5452, "step": 740800 }, { "epoch": 4.467203724026483, "grad_norm": 4.2740888595581055, "learning_rate": 8.731505382983916e-05, "loss": 4.698, "step": 740850 }, { "epoch": 4.467505215807817, "grad_norm": 2.4849231243133545, "learning_rate": 8.729656751009905e-05, "loss": 5.12, "step": 740900 }, { "epoch": 4.467806707589151, "grad_norm": 3.4694602489471436, "learning_rate": 8.727808234432606e-05, "loss": 4.7048, "step": 740950 }, { "epoch": 4.468108199370485, "grad_norm": 3.3410420417785645, "learning_rate": 8.725959833286046e-05, "loss": 4.2511, "step": 741000 }, { "epoch": 4.468108199370485, "eval_loss": 5.151421070098877, "eval_runtime": 39.2699, "eval_samples_per_second": 13.038, "eval_steps_per_second": 6.519, "eval_tts_loss": 7.65735623700389, "step": 741000 }, { "epoch": 4.46840969115182, "grad_norm": 3.273918867111206, "learning_rate": 8.724111547604248e-05, "loss": 5.2087, "step": 741050 }, { "epoch": 4.468711182933153, "grad_norm": 2.0582947731018066, "learning_rate": 8.722263377421217e-05, "loss": 4.5557, "step": 741100 }, { "epoch": 4.469012674714487, "grad_norm": 4.125049591064453, "learning_rate": 8.720415322770966e-05, "loss": 4.6128, "step": 741150 }, { "epoch": 4.469314166495821, "grad_norm": 3.0434420108795166, "learning_rate": 8.718567383687519e-05, "loss": 4.2927, "step": 741200 }, { "epoch": 4.469615658277156, "grad_norm": 4.087267875671387, "learning_rate": 8.716719560204872e-05, "loss": 4.6787, "step": 741250 }, { "epoch": 4.46991715005849, "grad_norm": 2.646441698074341, "learning_rate": 8.71487185235703e-05, "loss": 4.7341, "step": 741300 }, { "epoch": 4.470218641839823, "grad_norm": 3.2312469482421875, "learning_rate": 8.713024260178008e-05, "loss": 4.7326, "step": 741350 }, { "epoch": 4.470520133621157, "grad_norm": 0.710115373134613, "learning_rate": 8.711176783701797e-05, "loss": 4.1142, "step": 741400 }, { "epoch": 4.4708216254024915, "grad_norm": 2.286386728286743, "learning_rate": 8.709329422962402e-05, "loss": 4.8804, "step": 741450 }, { "epoch": 4.471123117183826, "grad_norm": 1.5710176229476929, "learning_rate": 8.707482177993829e-05, "loss": 4.9098, "step": 741500 }, { "epoch": 4.47142460896516, "grad_norm": 4.684281349182129, "learning_rate": 8.705635048830059e-05, "loss": 4.7845, "step": 741550 }, { "epoch": 4.471726100746493, "grad_norm": 2.4055705070495605, "learning_rate": 8.7037880355051e-05, "loss": 4.9676, "step": 741600 }, { "epoch": 4.4720275925278274, "grad_norm": 2.533297538757324, "learning_rate": 8.701941138052937e-05, "loss": 4.6376, "step": 741650 }, { "epoch": 4.472329084309162, "grad_norm": 2.5854995250701904, "learning_rate": 8.700094356507555e-05, "loss": 4.6052, "step": 741700 }, { "epoch": 4.472630576090496, "grad_norm": 1.358895182609558, "learning_rate": 8.698247690902945e-05, "loss": 4.5829, "step": 741750 }, { "epoch": 4.47293206787183, "grad_norm": 2.903427839279175, "learning_rate": 8.696401141273098e-05, "loss": 4.617, "step": 741800 }, { "epoch": 4.473233559653164, "grad_norm": 2.6254725456237793, "learning_rate": 8.694554707651986e-05, "loss": 4.6166, "step": 741850 }, { "epoch": 4.473535051434498, "grad_norm": 1.5644662380218506, "learning_rate": 8.692708390073598e-05, "loss": 5.0053, "step": 741900 }, { "epoch": 4.473836543215832, "grad_norm": 2.503444194793701, "learning_rate": 8.690862188571917e-05, "loss": 4.8819, "step": 741950 }, { "epoch": 4.474138034997166, "grad_norm": 2.1918203830718994, "learning_rate": 8.689016103180907e-05, "loss": 4.3349, "step": 742000 }, { "epoch": 4.4744395267785, "grad_norm": 3.8625028133392334, "learning_rate": 8.687170133934556e-05, "loss": 4.5382, "step": 742050 }, { "epoch": 4.474741018559834, "grad_norm": 0.8900120854377747, "learning_rate": 8.685324280866829e-05, "loss": 4.5768, "step": 742100 }, { "epoch": 4.475042510341168, "grad_norm": 2.6550769805908203, "learning_rate": 8.683478544011693e-05, "loss": 4.6736, "step": 742150 }, { "epoch": 4.475344002122502, "grad_norm": 2.5946390628814697, "learning_rate": 8.681632923403117e-05, "loss": 4.683, "step": 742200 }, { "epoch": 4.475645493903836, "grad_norm": 2.891866445541382, "learning_rate": 8.67978741907508e-05, "loss": 4.8777, "step": 742250 }, { "epoch": 4.47594698568517, "grad_norm": 2.520287036895752, "learning_rate": 8.677942031061527e-05, "loss": 4.7298, "step": 742300 }, { "epoch": 4.476248477466505, "grad_norm": 3.3392627239227295, "learning_rate": 8.676096759396428e-05, "loss": 4.9394, "step": 742350 }, { "epoch": 4.476549969247838, "grad_norm": 4.151320934295654, "learning_rate": 8.674251604113753e-05, "loss": 4.8722, "step": 742400 }, { "epoch": 4.476851461029172, "grad_norm": 3.771198034286499, "learning_rate": 8.672406565247448e-05, "loss": 5.0027, "step": 742450 }, { "epoch": 4.477152952810506, "grad_norm": 3.1806957721710205, "learning_rate": 8.670561642831465e-05, "loss": 4.7443, "step": 742500 }, { "epoch": 4.4774544445918405, "grad_norm": 2.157334327697754, "learning_rate": 8.668716836899766e-05, "loss": 4.6411, "step": 742550 }, { "epoch": 4.477755936373175, "grad_norm": 3.863093614578247, "learning_rate": 8.666872147486294e-05, "loss": 4.5415, "step": 742600 }, { "epoch": 4.478057428154509, "grad_norm": 1.6424959897994995, "learning_rate": 8.665027574625004e-05, "loss": 4.9222, "step": 742650 }, { "epoch": 4.478358919935842, "grad_norm": 1.6030774116516113, "learning_rate": 8.663183118349839e-05, "loss": 4.8128, "step": 742700 }, { "epoch": 4.478660411717176, "grad_norm": 2.827099323272705, "learning_rate": 8.661338778694752e-05, "loss": 4.857, "step": 742750 }, { "epoch": 4.478961903498511, "grad_norm": 2.4172744750976562, "learning_rate": 8.659494555693679e-05, "loss": 4.7886, "step": 742800 }, { "epoch": 4.479263395279845, "grad_norm": 3.084484100341797, "learning_rate": 8.657650449380556e-05, "loss": 4.7166, "step": 742850 }, { "epoch": 4.479564887061179, "grad_norm": 2.566051483154297, "learning_rate": 8.655806459789329e-05, "loss": 4.7561, "step": 742900 }, { "epoch": 4.479866378842512, "grad_norm": 4.777130126953125, "learning_rate": 8.653962586953925e-05, "loss": 4.5194, "step": 742950 }, { "epoch": 4.480167870623847, "grad_norm": 2.777535915374756, "learning_rate": 8.652118830908284e-05, "loss": 4.8255, "step": 743000 }, { "epoch": 4.480469362405181, "grad_norm": 1.5810385942459106, "learning_rate": 8.650275191686343e-05, "loss": 4.6871, "step": 743050 }, { "epoch": 4.480770854186515, "grad_norm": 1.3485652208328247, "learning_rate": 8.64843166932202e-05, "loss": 4.7427, "step": 743100 }, { "epoch": 4.481072345967849, "grad_norm": 5.045237064361572, "learning_rate": 8.646588263849245e-05, "loss": 5.0897, "step": 743150 }, { "epoch": 4.481373837749183, "grad_norm": 0.6352879405021667, "learning_rate": 8.644744975301956e-05, "loss": 4.918, "step": 743200 }, { "epoch": 4.481675329530517, "grad_norm": 3.7474586963653564, "learning_rate": 8.642901803714064e-05, "loss": 4.8633, "step": 743250 }, { "epoch": 4.481976821311851, "grad_norm": 5.579488754272461, "learning_rate": 8.641058749119487e-05, "loss": 4.6113, "step": 743300 }, { "epoch": 4.482278313093185, "grad_norm": 3.3112094402313232, "learning_rate": 8.639215811552155e-05, "loss": 4.7477, "step": 743350 }, { "epoch": 4.482579804874519, "grad_norm": 3.334193706512451, "learning_rate": 8.637372991045973e-05, "loss": 4.6416, "step": 743400 }, { "epoch": 4.4828812966558536, "grad_norm": 3.6178948879241943, "learning_rate": 8.63553028763486e-05, "loss": 4.6289, "step": 743450 }, { "epoch": 4.483182788437187, "grad_norm": 3.8473360538482666, "learning_rate": 8.633687701352736e-05, "loss": 4.6959, "step": 743500 }, { "epoch": 4.483484280218521, "grad_norm": 3.3506298065185547, "learning_rate": 8.631845232233499e-05, "loss": 4.863, "step": 743550 }, { "epoch": 4.483785771999855, "grad_norm": 3.3466386795043945, "learning_rate": 8.630002880311069e-05, "loss": 4.4357, "step": 743600 }, { "epoch": 4.4840872637811895, "grad_norm": 1.7409980297088623, "learning_rate": 8.628160645619345e-05, "loss": 4.5804, "step": 743650 }, { "epoch": 4.484388755562524, "grad_norm": 1.977756142616272, "learning_rate": 8.626318528192226e-05, "loss": 4.5978, "step": 743700 }, { "epoch": 4.484690247343857, "grad_norm": 2.506110668182373, "learning_rate": 8.624476528063618e-05, "loss": 4.6117, "step": 743750 }, { "epoch": 4.484991739125191, "grad_norm": 3.3303163051605225, "learning_rate": 8.622634645267427e-05, "loss": 4.6718, "step": 743800 }, { "epoch": 4.485293230906525, "grad_norm": 4.149348735809326, "learning_rate": 8.620792879837538e-05, "loss": 4.8773, "step": 743850 }, { "epoch": 4.48559472268786, "grad_norm": 2.685331106185913, "learning_rate": 8.618951231807853e-05, "loss": 4.3316, "step": 743900 }, { "epoch": 4.485896214469194, "grad_norm": 2.4640345573425293, "learning_rate": 8.61710970121227e-05, "loss": 4.7259, "step": 743950 }, { "epoch": 4.486197706250527, "grad_norm": 2.9042320251464844, "learning_rate": 8.615268288084674e-05, "loss": 4.4394, "step": 744000 }, { "epoch": 4.486197706250527, "eval_loss": 5.147136211395264, "eval_runtime": 38.9671, "eval_samples_per_second": 13.139, "eval_steps_per_second": 6.57, "eval_tts_loss": 7.651533348037893, "step": 744000 }, { "epoch": 4.486499198031861, "grad_norm": 3.3579490184783936, "learning_rate": 8.613426992458949e-05, "loss": 4.5941, "step": 744050 }, { "epoch": 4.486800689813196, "grad_norm": 2.5276870727539062, "learning_rate": 8.611585814368991e-05, "loss": 4.5517, "step": 744100 }, { "epoch": 4.48710218159453, "grad_norm": 3.271212339401245, "learning_rate": 8.609744753848672e-05, "loss": 4.8583, "step": 744150 }, { "epoch": 4.487403673375864, "grad_norm": 3.736088275909424, "learning_rate": 8.607903810931883e-05, "loss": 4.7309, "step": 744200 }, { "epoch": 4.487705165157198, "grad_norm": 3.569761037826538, "learning_rate": 8.606062985652509e-05, "loss": 4.6926, "step": 744250 }, { "epoch": 4.4880066569385315, "grad_norm": 1.7110886573791504, "learning_rate": 8.604222278044414e-05, "loss": 5.1146, "step": 744300 }, { "epoch": 4.488308148719866, "grad_norm": 2.633479356765747, "learning_rate": 8.60238168814148e-05, "loss": 4.8927, "step": 744350 }, { "epoch": 4.4886096405012, "grad_norm": 1.1792649030685425, "learning_rate": 8.600541215977589e-05, "loss": 4.8865, "step": 744400 }, { "epoch": 4.488911132282534, "grad_norm": 2.6441805362701416, "learning_rate": 8.598700861586605e-05, "loss": 4.66, "step": 744450 }, { "epoch": 4.489212624063868, "grad_norm": 2.574748992919922, "learning_rate": 8.596860625002388e-05, "loss": 3.9952, "step": 744500 }, { "epoch": 4.489514115845202, "grad_norm": 2.206308126449585, "learning_rate": 8.59502050625882e-05, "loss": 4.9447, "step": 744550 }, { "epoch": 4.489815607626536, "grad_norm": 2.812774419784546, "learning_rate": 8.593180505389754e-05, "loss": 4.6688, "step": 744600 }, { "epoch": 4.49011709940787, "grad_norm": 2.4287850856781006, "learning_rate": 8.591340622429058e-05, "loss": 4.4819, "step": 744650 }, { "epoch": 4.490418591189204, "grad_norm": 2.3785996437072754, "learning_rate": 8.589500857410599e-05, "loss": 4.9051, "step": 744700 }, { "epoch": 4.4907200829705385, "grad_norm": 2.92625093460083, "learning_rate": 8.58766121036822e-05, "loss": 4.7984, "step": 744750 }, { "epoch": 4.491021574751873, "grad_norm": 2.2404062747955322, "learning_rate": 8.585821681335792e-05, "loss": 4.8472, "step": 744800 }, { "epoch": 4.491323066533206, "grad_norm": 2.287463665008545, "learning_rate": 8.583982270347163e-05, "loss": 4.8002, "step": 744850 }, { "epoch": 4.49162455831454, "grad_norm": 3.25765323638916, "learning_rate": 8.582142977436177e-05, "loss": 4.6898, "step": 744900 }, { "epoch": 4.491926050095874, "grad_norm": 2.3348329067230225, "learning_rate": 8.580303802636693e-05, "loss": 4.4968, "step": 744950 }, { "epoch": 4.492227541877209, "grad_norm": 3.3739964962005615, "learning_rate": 8.578464745982561e-05, "loss": 4.8887, "step": 745000 }, { "epoch": 4.492529033658543, "grad_norm": 3.078284502029419, "learning_rate": 8.576625807507615e-05, "loss": 4.6771, "step": 745050 }, { "epoch": 4.492830525439876, "grad_norm": 2.504326343536377, "learning_rate": 8.574786987245706e-05, "loss": 4.6135, "step": 745100 }, { "epoch": 4.49313201722121, "grad_norm": 1.2954232692718506, "learning_rate": 8.572948285230677e-05, "loss": 4.84, "step": 745150 }, { "epoch": 4.493433509002545, "grad_norm": 2.6079330444335938, "learning_rate": 8.571109701496364e-05, "loss": 4.2656, "step": 745200 }, { "epoch": 4.493735000783879, "grad_norm": 1.1491694450378418, "learning_rate": 8.569271236076598e-05, "loss": 4.663, "step": 745250 }, { "epoch": 4.494036492565213, "grad_norm": 2.8450846672058105, "learning_rate": 8.56743288900522e-05, "loss": 4.5178, "step": 745300 }, { "epoch": 4.494337984346546, "grad_norm": 2.6997835636138916, "learning_rate": 8.565594660316055e-05, "loss": 4.4641, "step": 745350 }, { "epoch": 4.4946394761278805, "grad_norm": 2.332740068435669, "learning_rate": 8.56375655004294e-05, "loss": 4.9062, "step": 745400 }, { "epoch": 4.494940967909215, "grad_norm": 2.4784610271453857, "learning_rate": 8.561918558219705e-05, "loss": 4.6613, "step": 745450 }, { "epoch": 4.495242459690549, "grad_norm": 2.8429603576660156, "learning_rate": 8.560080684880166e-05, "loss": 5.0181, "step": 745500 }, { "epoch": 4.495543951471883, "grad_norm": 3.33609938621521, "learning_rate": 8.558242930058153e-05, "loss": 4.6114, "step": 745550 }, { "epoch": 4.495845443253217, "grad_norm": 3.847149610519409, "learning_rate": 8.55640529378749e-05, "loss": 4.6473, "step": 745600 }, { "epoch": 4.496146935034551, "grad_norm": 3.4361140727996826, "learning_rate": 8.554567776101992e-05, "loss": 4.8535, "step": 745650 }, { "epoch": 4.496448426815885, "grad_norm": 2.2204158306121826, "learning_rate": 8.552730377035472e-05, "loss": 4.3862, "step": 745700 }, { "epoch": 4.496749918597219, "grad_norm": 1.4553638696670532, "learning_rate": 8.550893096621752e-05, "loss": 4.4937, "step": 745750 }, { "epoch": 4.497051410378553, "grad_norm": 4.847972393035889, "learning_rate": 8.549055934894636e-05, "loss": 4.5008, "step": 745800 }, { "epoch": 4.4973529021598875, "grad_norm": 3.510894775390625, "learning_rate": 8.54721889188794e-05, "loss": 4.8645, "step": 745850 }, { "epoch": 4.497654393941221, "grad_norm": 2.848452091217041, "learning_rate": 8.545381967635478e-05, "loss": 4.9684, "step": 745900 }, { "epoch": 4.497955885722555, "grad_norm": 3.068958044052124, "learning_rate": 8.543545162171043e-05, "loss": 4.6998, "step": 745950 }, { "epoch": 4.498257377503889, "grad_norm": 4.579647541046143, "learning_rate": 8.541708475528451e-05, "loss": 4.4266, "step": 746000 }, { "epoch": 4.498558869285223, "grad_norm": 0.56160968542099, "learning_rate": 8.5398719077415e-05, "loss": 4.6784, "step": 746050 }, { "epoch": 4.498860361066558, "grad_norm": 2.1789891719818115, "learning_rate": 8.538035458843979e-05, "loss": 4.8708, "step": 746100 }, { "epoch": 4.499161852847891, "grad_norm": 2.828469753265381, "learning_rate": 8.536199128869695e-05, "loss": 4.5407, "step": 746150 }, { "epoch": 4.499463344629225, "grad_norm": 2.492347240447998, "learning_rate": 8.534362917852442e-05, "loss": 4.6899, "step": 746200 }, { "epoch": 4.499764836410559, "grad_norm": 3.183769702911377, "learning_rate": 8.53252682582602e-05, "loss": 5.0721, "step": 746250 }, { "epoch": 4.500066328191894, "grad_norm": 4.033400535583496, "learning_rate": 8.530690852824203e-05, "loss": 4.389, "step": 746300 }, { "epoch": 4.500367819973228, "grad_norm": 2.703277587890625, "learning_rate": 8.528854998880796e-05, "loss": 4.8259, "step": 746350 }, { "epoch": 4.500669311754562, "grad_norm": 2.57387375831604, "learning_rate": 8.527019264029571e-05, "loss": 4.398, "step": 746400 }, { "epoch": 4.500970803535895, "grad_norm": 2.6476547718048096, "learning_rate": 8.525183648304327e-05, "loss": 4.6671, "step": 746450 }, { "epoch": 4.5012722953172295, "grad_norm": 2.4398107528686523, "learning_rate": 8.52334815173883e-05, "loss": 4.7047, "step": 746500 }, { "epoch": 4.501573787098564, "grad_norm": 3.611738443374634, "learning_rate": 8.521512774366874e-05, "loss": 4.7954, "step": 746550 }, { "epoch": 4.501875278879898, "grad_norm": 2.6434850692749023, "learning_rate": 8.519677516222225e-05, "loss": 5.0223, "step": 746600 }, { "epoch": 4.502176770661232, "grad_norm": 3.6898159980773926, "learning_rate": 8.517842377338662e-05, "loss": 4.9344, "step": 746650 }, { "epoch": 4.502478262442565, "grad_norm": 1.3357945680618286, "learning_rate": 8.516007357749966e-05, "loss": 4.8281, "step": 746700 }, { "epoch": 4.5027797542239, "grad_norm": 1.9712270498275757, "learning_rate": 8.514172457489896e-05, "loss": 4.8545, "step": 746750 }, { "epoch": 4.503081246005234, "grad_norm": 3.968197822570801, "learning_rate": 8.512337676592233e-05, "loss": 4.745, "step": 746800 }, { "epoch": 4.503382737786568, "grad_norm": 4.312568187713623, "learning_rate": 8.510503015090734e-05, "loss": 4.9748, "step": 746850 }, { "epoch": 4.503684229567902, "grad_norm": 2.6251163482666016, "learning_rate": 8.508668473019163e-05, "loss": 4.7338, "step": 746900 }, { "epoch": 4.5039857213492365, "grad_norm": 1.8574389219284058, "learning_rate": 8.506834050411284e-05, "loss": 4.5677, "step": 746950 }, { "epoch": 4.50428721313057, "grad_norm": 3.4464292526245117, "learning_rate": 8.504999747300864e-05, "loss": 4.9528, "step": 747000 }, { "epoch": 4.50428721313057, "eval_loss": 5.1347808837890625, "eval_runtime": 39.0333, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.716089698979179, "step": 747000 }, { "epoch": 4.504588704911904, "grad_norm": 2.4955124855041504, "learning_rate": 8.503165563721648e-05, "loss": 5.2649, "step": 747050 }, { "epoch": 4.504890196693238, "grad_norm": 2.459944248199463, "learning_rate": 8.501331499707398e-05, "loss": 4.6073, "step": 747100 }, { "epoch": 4.505191688474572, "grad_norm": 2.1544995307922363, "learning_rate": 8.499497555291875e-05, "loss": 5.0635, "step": 747150 }, { "epoch": 4.505493180255906, "grad_norm": 3.122258424758911, "learning_rate": 8.497663730508824e-05, "loss": 4.4621, "step": 747200 }, { "epoch": 4.50579467203724, "grad_norm": 1.0993926525115967, "learning_rate": 8.495830025391984e-05, "loss": 4.7256, "step": 747250 }, { "epoch": 4.506096163818574, "grad_norm": 2.4492218494415283, "learning_rate": 8.493996439975116e-05, "loss": 4.4959, "step": 747300 }, { "epoch": 4.506397655599908, "grad_norm": 2.378439426422119, "learning_rate": 8.492162974291955e-05, "loss": 4.6752, "step": 747350 }, { "epoch": 4.5066991473812426, "grad_norm": 2.549159049987793, "learning_rate": 8.490329628376248e-05, "loss": 5.1198, "step": 747400 }, { "epoch": 4.507000639162577, "grad_norm": 2.1666009426116943, "learning_rate": 8.488496402261739e-05, "loss": 4.8067, "step": 747450 }, { "epoch": 4.507302130943911, "grad_norm": 3.818594455718994, "learning_rate": 8.486663295982153e-05, "loss": 4.4858, "step": 747500 }, { "epoch": 4.507603622725244, "grad_norm": 2.696765184402466, "learning_rate": 8.484830309571239e-05, "loss": 4.8136, "step": 747550 }, { "epoch": 4.5079051145065785, "grad_norm": 3.6566803455352783, "learning_rate": 8.482997443062728e-05, "loss": 4.5851, "step": 747600 }, { "epoch": 4.508206606287913, "grad_norm": 2.6796040534973145, "learning_rate": 8.48116469649035e-05, "loss": 4.7798, "step": 747650 }, { "epoch": 4.508508098069247, "grad_norm": 2.188861608505249, "learning_rate": 8.479332069887825e-05, "loss": 5.0719, "step": 747700 }, { "epoch": 4.50880958985058, "grad_norm": 2.5705981254577637, "learning_rate": 8.477499563288896e-05, "loss": 4.7855, "step": 747750 }, { "epoch": 4.509111081631914, "grad_norm": 2.3394253253936768, "learning_rate": 8.475667176727274e-05, "loss": 4.7715, "step": 747800 }, { "epoch": 4.509412573413249, "grad_norm": 1.1597111225128174, "learning_rate": 8.473834910236685e-05, "loss": 4.931, "step": 747850 }, { "epoch": 4.509714065194583, "grad_norm": 3.1534247398376465, "learning_rate": 8.472002763850858e-05, "loss": 4.4529, "step": 747900 }, { "epoch": 4.510015556975917, "grad_norm": 2.748473644256592, "learning_rate": 8.470170737603497e-05, "loss": 5.0539, "step": 747950 }, { "epoch": 4.510317048757251, "grad_norm": 2.314173460006714, "learning_rate": 8.468338831528332e-05, "loss": 4.6793, "step": 748000 }, { "epoch": 4.510618540538585, "grad_norm": 3.4006378650665283, "learning_rate": 8.466507045659069e-05, "loss": 4.9831, "step": 748050 }, { "epoch": 4.510920032319919, "grad_norm": 4.007464408874512, "learning_rate": 8.464675380029416e-05, "loss": 4.7723, "step": 748100 }, { "epoch": 4.511221524101253, "grad_norm": 2.7982289791107178, "learning_rate": 8.462843834673084e-05, "loss": 4.5646, "step": 748150 }, { "epoch": 4.511523015882587, "grad_norm": 2.6883044242858887, "learning_rate": 8.461012409623789e-05, "loss": 4.5594, "step": 748200 }, { "epoch": 4.511824507663921, "grad_norm": 3.091454267501831, "learning_rate": 8.459181104915223e-05, "loss": 4.8533, "step": 748250 }, { "epoch": 4.512125999445255, "grad_norm": 2.4809484481811523, "learning_rate": 8.457349920581094e-05, "loss": 4.7412, "step": 748300 }, { "epoch": 4.512427491226589, "grad_norm": 2.513885498046875, "learning_rate": 8.455518856655112e-05, "loss": 4.4279, "step": 748350 }, { "epoch": 4.512728983007923, "grad_norm": 3.130606174468994, "learning_rate": 8.453687913170961e-05, "loss": 4.5425, "step": 748400 }, { "epoch": 4.513030474789257, "grad_norm": 5.582848072052002, "learning_rate": 8.451857090162339e-05, "loss": 4.8712, "step": 748450 }, { "epoch": 4.5133319665705915, "grad_norm": 2.3750247955322266, "learning_rate": 8.450026387662947e-05, "loss": 4.3668, "step": 748500 }, { "epoch": 4.513633458351926, "grad_norm": 2.646960496902466, "learning_rate": 8.448195805706466e-05, "loss": 4.9337, "step": 748550 }, { "epoch": 4.513934950133259, "grad_norm": 3.4145443439483643, "learning_rate": 8.446365344326593e-05, "loss": 4.9508, "step": 748600 }, { "epoch": 4.514236441914593, "grad_norm": 3.3771259784698486, "learning_rate": 8.444535003557019e-05, "loss": 4.9668, "step": 748650 }, { "epoch": 4.5145379336959275, "grad_norm": 3.215470790863037, "learning_rate": 8.442704783431416e-05, "loss": 4.8325, "step": 748700 }, { "epoch": 4.514839425477262, "grad_norm": 3.750737428665161, "learning_rate": 8.440874683983475e-05, "loss": 4.4714, "step": 748750 }, { "epoch": 4.515140917258596, "grad_norm": 7.9590229988098145, "learning_rate": 8.43904470524688e-05, "loss": 4.5811, "step": 748800 }, { "epoch": 4.515442409039929, "grad_norm": 2.725734233856201, "learning_rate": 8.437214847255304e-05, "loss": 4.5536, "step": 748850 }, { "epoch": 4.515743900821263, "grad_norm": 3.602144956588745, "learning_rate": 8.435385110042418e-05, "loss": 4.7686, "step": 748900 }, { "epoch": 4.516045392602598, "grad_norm": 3.7120068073272705, "learning_rate": 8.433555493641907e-05, "loss": 4.8694, "step": 748950 }, { "epoch": 4.516346884383932, "grad_norm": 2.767381191253662, "learning_rate": 8.431725998087428e-05, "loss": 4.4581, "step": 749000 }, { "epoch": 4.516648376165266, "grad_norm": 1.4434665441513062, "learning_rate": 8.429896623412662e-05, "loss": 4.082, "step": 749050 }, { "epoch": 4.5169498679466, "grad_norm": 2.6615145206451416, "learning_rate": 8.428067369651276e-05, "loss": 4.7864, "step": 749100 }, { "epoch": 4.517251359727934, "grad_norm": 3.0697288513183594, "learning_rate": 8.426238236836925e-05, "loss": 4.6371, "step": 749150 }, { "epoch": 4.517552851509268, "grad_norm": 0.7111141085624695, "learning_rate": 8.424409225003286e-05, "loss": 4.5632, "step": 749200 }, { "epoch": 4.517854343290602, "grad_norm": 2.853510618209839, "learning_rate": 8.422580334184013e-05, "loss": 4.7965, "step": 749250 }, { "epoch": 4.518155835071936, "grad_norm": 2.421260118484497, "learning_rate": 8.420751564412752e-05, "loss": 5.0184, "step": 749300 }, { "epoch": 4.5184573268532695, "grad_norm": 2.9574410915374756, "learning_rate": 8.418922915723173e-05, "loss": 4.5964, "step": 749350 }, { "epoch": 4.518758818634604, "grad_norm": 3.5249290466308594, "learning_rate": 8.417094388148929e-05, "loss": 4.6126, "step": 749400 }, { "epoch": 4.519060310415938, "grad_norm": 3.9167652130126953, "learning_rate": 8.415265981723664e-05, "loss": 4.589, "step": 749450 }, { "epoch": 4.519361802197272, "grad_norm": 3.099670886993408, "learning_rate": 8.413437696481031e-05, "loss": 4.779, "step": 749500 }, { "epoch": 4.519663293978606, "grad_norm": 3.4744701385498047, "learning_rate": 8.411609532454685e-05, "loss": 5.0024, "step": 749550 }, { "epoch": 4.5199647857599405, "grad_norm": 2.4709174633026123, "learning_rate": 8.409781489678263e-05, "loss": 5.0414, "step": 749600 }, { "epoch": 4.520266277541274, "grad_norm": 2.4886481761932373, "learning_rate": 8.407953568185402e-05, "loss": 4.5263, "step": 749650 }, { "epoch": 4.520567769322608, "grad_norm": 2.105091094970703, "learning_rate": 8.406125768009748e-05, "loss": 4.5722, "step": 749700 }, { "epoch": 4.520869261103942, "grad_norm": 2.5869319438934326, "learning_rate": 8.404298089184945e-05, "loss": 5.1655, "step": 749750 }, { "epoch": 4.5211707528852765, "grad_norm": 2.254793882369995, "learning_rate": 8.402470531744618e-05, "loss": 4.7383, "step": 749800 }, { "epoch": 4.521472244666611, "grad_norm": 2.2388155460357666, "learning_rate": 8.400643095722404e-05, "loss": 4.1983, "step": 749850 }, { "epoch": 4.521773736447944, "grad_norm": 1.331539511680603, "learning_rate": 8.398815781151944e-05, "loss": 5.0739, "step": 749900 }, { "epoch": 4.522075228229278, "grad_norm": 4.972090721130371, "learning_rate": 8.396988588066851e-05, "loss": 4.936, "step": 749950 }, { "epoch": 4.522376720010612, "grad_norm": 3.4888124465942383, "learning_rate": 8.395161516500767e-05, "loss": 4.5446, "step": 750000 }, { "epoch": 4.522376720010612, "eval_loss": 5.14589786529541, "eval_runtime": 38.9071, "eval_samples_per_second": 13.16, "eval_steps_per_second": 6.58, "eval_tts_loss": 7.699731626147428, "step": 750000 }, { "epoch": 4.522678211791947, "grad_norm": 2.169943332672119, "learning_rate": 8.393334566487312e-05, "loss": 4.7543, "step": 750050 }, { "epoch": 4.522979703573281, "grad_norm": 2.4918041229248047, "learning_rate": 8.391507738060098e-05, "loss": 4.9918, "step": 750100 }, { "epoch": 4.523281195354615, "grad_norm": 0.9701102375984192, "learning_rate": 8.389681031252752e-05, "loss": 4.8564, "step": 750150 }, { "epoch": 4.523582687135948, "grad_norm": 3.6525394916534424, "learning_rate": 8.387854446098899e-05, "loss": 4.3955, "step": 750200 }, { "epoch": 4.523884178917283, "grad_norm": 3.8627634048461914, "learning_rate": 8.386027982632147e-05, "loss": 4.4367, "step": 750250 }, { "epoch": 4.524185670698617, "grad_norm": 3.7537670135498047, "learning_rate": 8.384201640886107e-05, "loss": 5.0326, "step": 750300 }, { "epoch": 4.524487162479951, "grad_norm": 3.257145404815674, "learning_rate": 8.382375420894404e-05, "loss": 4.7683, "step": 750350 }, { "epoch": 4.524788654261285, "grad_norm": 2.5936899185180664, "learning_rate": 8.380549322690634e-05, "loss": 4.7766, "step": 750400 }, { "epoch": 4.5250901460426185, "grad_norm": 2.905410051345825, "learning_rate": 8.378723346308403e-05, "loss": 4.7065, "step": 750450 }, { "epoch": 4.525391637823953, "grad_norm": 3.2929623126983643, "learning_rate": 8.376897491781326e-05, "loss": 4.3765, "step": 750500 }, { "epoch": 4.525693129605287, "grad_norm": 3.010341167449951, "learning_rate": 8.375071759142994e-05, "loss": 4.8006, "step": 750550 }, { "epoch": 4.525994621386621, "grad_norm": 3.143699884414673, "learning_rate": 8.373246148427009e-05, "loss": 4.2662, "step": 750600 }, { "epoch": 4.526296113167955, "grad_norm": 2.671441078186035, "learning_rate": 8.37142065966698e-05, "loss": 5.1194, "step": 750650 }, { "epoch": 4.5265976049492895, "grad_norm": 1.5089857578277588, "learning_rate": 8.369595292896487e-05, "loss": 4.58, "step": 750700 }, { "epoch": 4.526899096730623, "grad_norm": 4.21748685836792, "learning_rate": 8.367770048149138e-05, "loss": 4.846, "step": 750750 }, { "epoch": 4.527200588511957, "grad_norm": 2.0071463584899902, "learning_rate": 8.365944925458516e-05, "loss": 5.0207, "step": 750800 }, { "epoch": 4.527502080293291, "grad_norm": 1.208251714706421, "learning_rate": 8.364119924858205e-05, "loss": 4.6233, "step": 750850 }, { "epoch": 4.5278035720746255, "grad_norm": 3.31534481048584, "learning_rate": 8.362295046381794e-05, "loss": 4.57, "step": 750900 }, { "epoch": 4.528105063855959, "grad_norm": 3.311028480529785, "learning_rate": 8.36047029006288e-05, "loss": 4.4465, "step": 750950 }, { "epoch": 4.528406555637293, "grad_norm": 2.4256558418273926, "learning_rate": 8.358645655935026e-05, "loss": 4.8898, "step": 751000 }, { "epoch": 4.528708047418627, "grad_norm": 3.3226945400238037, "learning_rate": 8.356821144031823e-05, "loss": 4.9324, "step": 751050 }, { "epoch": 4.529009539199961, "grad_norm": 1.7070444822311401, "learning_rate": 8.354996754386852e-05, "loss": 4.3523, "step": 751100 }, { "epoch": 4.529311030981296, "grad_norm": 2.0351736545562744, "learning_rate": 8.353172487033677e-05, "loss": 4.3788, "step": 751150 }, { "epoch": 4.52961252276263, "grad_norm": 2.3917338848114014, "learning_rate": 8.351348342005882e-05, "loss": 5.0813, "step": 751200 }, { "epoch": 4.529914014543964, "grad_norm": 2.499192953109741, "learning_rate": 8.349524319337033e-05, "loss": 5.0365, "step": 751250 }, { "epoch": 4.530215506325297, "grad_norm": 1.256948709487915, "learning_rate": 8.347700419060693e-05, "loss": 4.4378, "step": 751300 }, { "epoch": 4.5305169981066316, "grad_norm": 2.3708291053771973, "learning_rate": 8.345876641210433e-05, "loss": 4.8559, "step": 751350 }, { "epoch": 4.530818489887966, "grad_norm": 3.0912582874298096, "learning_rate": 8.344052985819821e-05, "loss": 4.7568, "step": 751400 }, { "epoch": 4.5311199816693, "grad_norm": 4.384160995483398, "learning_rate": 8.342229452922412e-05, "loss": 5.2697, "step": 751450 }, { "epoch": 4.531421473450633, "grad_norm": 2.3656983375549316, "learning_rate": 8.340406042551769e-05, "loss": 4.6562, "step": 751500 }, { "epoch": 4.5317229652319675, "grad_norm": 3.6136183738708496, "learning_rate": 8.338582754741455e-05, "loss": 4.4858, "step": 751550 }, { "epoch": 4.532024457013302, "grad_norm": 4.928226470947266, "learning_rate": 8.336759589525018e-05, "loss": 4.5564, "step": 751600 }, { "epoch": 4.532325948794636, "grad_norm": 3.2792317867279053, "learning_rate": 8.334936546936006e-05, "loss": 4.8951, "step": 751650 }, { "epoch": 4.53262744057597, "grad_norm": 3.447979211807251, "learning_rate": 8.333113627007981e-05, "loss": 4.877, "step": 751700 }, { "epoch": 4.532928932357304, "grad_norm": 2.048675060272217, "learning_rate": 8.33129082977448e-05, "loss": 4.7802, "step": 751750 }, { "epoch": 4.533230424138638, "grad_norm": 3.0552937984466553, "learning_rate": 8.329468155269055e-05, "loss": 4.9082, "step": 751800 }, { "epoch": 4.533531915919972, "grad_norm": 1.6473826169967651, "learning_rate": 8.327645603525255e-05, "loss": 4.7173, "step": 751850 }, { "epoch": 4.533833407701306, "grad_norm": 2.5211668014526367, "learning_rate": 8.325823174576609e-05, "loss": 4.1951, "step": 751900 }, { "epoch": 4.53413489948264, "grad_norm": 1.3099048137664795, "learning_rate": 8.324000868456668e-05, "loss": 4.4732, "step": 751950 }, { "epoch": 4.5344363912639745, "grad_norm": 2.022212505340576, "learning_rate": 8.32217868519896e-05, "loss": 4.9133, "step": 752000 }, { "epoch": 4.534737883045308, "grad_norm": 2.56532621383667, "learning_rate": 8.32035662483703e-05, "loss": 5.0591, "step": 752050 }, { "epoch": 4.535039374826642, "grad_norm": 3.437650203704834, "learning_rate": 8.318534687404396e-05, "loss": 4.8202, "step": 752100 }, { "epoch": 4.535340866607976, "grad_norm": 2.966323137283325, "learning_rate": 8.316712872934602e-05, "loss": 4.4389, "step": 752150 }, { "epoch": 4.53564235838931, "grad_norm": 1.71957528591156, "learning_rate": 8.314891181461166e-05, "loss": 4.5976, "step": 752200 }, { "epoch": 4.535943850170645, "grad_norm": 2.9798924922943115, "learning_rate": 8.313069613017617e-05, "loss": 4.6227, "step": 752250 }, { "epoch": 4.536245341951979, "grad_norm": 3.6306235790252686, "learning_rate": 8.311248167637482e-05, "loss": 4.726, "step": 752300 }, { "epoch": 4.536546833733312, "grad_norm": 2.936676025390625, "learning_rate": 8.309426845354278e-05, "loss": 5.0005, "step": 752350 }, { "epoch": 4.536848325514646, "grad_norm": 3.8906171321868896, "learning_rate": 8.307605646201527e-05, "loss": 4.8088, "step": 752400 }, { "epoch": 4.5371498172959805, "grad_norm": 1.5421594381332397, "learning_rate": 8.305784570212744e-05, "loss": 4.3645, "step": 752450 }, { "epoch": 4.537451309077315, "grad_norm": 2.3151657581329346, "learning_rate": 8.303963617421441e-05, "loss": 4.8272, "step": 752500 }, { "epoch": 4.537752800858649, "grad_norm": 9.285358428955078, "learning_rate": 8.302142787861128e-05, "loss": 4.4488, "step": 752550 }, { "epoch": 4.538054292639982, "grad_norm": 3.851862907409668, "learning_rate": 8.300322081565326e-05, "loss": 4.7069, "step": 752600 }, { "epoch": 4.5383557844213165, "grad_norm": 3.78000807762146, "learning_rate": 8.298501498567528e-05, "loss": 4.7332, "step": 752650 }, { "epoch": 4.538657276202651, "grad_norm": 3.545736312866211, "learning_rate": 8.296681038901245e-05, "loss": 4.3802, "step": 752700 }, { "epoch": 4.538958767983985, "grad_norm": 0.8097816109657288, "learning_rate": 8.29486070259999e-05, "loss": 4.8748, "step": 752750 }, { "epoch": 4.539260259765319, "grad_norm": 2.5614473819732666, "learning_rate": 8.293040489697254e-05, "loss": 4.5675, "step": 752800 }, { "epoch": 4.539561751546653, "grad_norm": 2.672260284423828, "learning_rate": 8.291220400226531e-05, "loss": 4.6877, "step": 752850 }, { "epoch": 4.539863243327987, "grad_norm": 2.90371036529541, "learning_rate": 8.289400434221324e-05, "loss": 4.7354, "step": 752900 }, { "epoch": 4.540164735109321, "grad_norm": 3.2714600563049316, "learning_rate": 8.287580591715127e-05, "loss": 4.4051, "step": 752950 }, { "epoch": 4.540466226890655, "grad_norm": 1.3810256719589233, "learning_rate": 8.285760872741428e-05, "loss": 4.9215, "step": 753000 }, { "epoch": 4.540466226890655, "eval_loss": 5.133138656616211, "eval_runtime": 39.238, "eval_samples_per_second": 13.049, "eval_steps_per_second": 6.524, "eval_tts_loss": 7.677029042233946, "step": 753000 }, { "epoch": 4.540767718671989, "grad_norm": 3.1192197799682617, "learning_rate": 8.283941277333717e-05, "loss": 4.694, "step": 753050 }, { "epoch": 4.541069210453323, "grad_norm": 2.319196939468384, "learning_rate": 8.28212180552549e-05, "loss": 4.4078, "step": 753100 }, { "epoch": 4.541370702234657, "grad_norm": 2.488870620727539, "learning_rate": 8.280302457350215e-05, "loss": 4.6589, "step": 753150 }, { "epoch": 4.541672194015991, "grad_norm": 0.9202827215194702, "learning_rate": 8.278483232841391e-05, "loss": 4.7179, "step": 753200 }, { "epoch": 4.541973685797325, "grad_norm": 2.7468655109405518, "learning_rate": 8.276664132032492e-05, "loss": 5.1097, "step": 753250 }, { "epoch": 4.542275177578659, "grad_norm": 3.552088975906372, "learning_rate": 8.27484515495699e-05, "loss": 4.7418, "step": 753300 }, { "epoch": 4.542576669359994, "grad_norm": 3.255075216293335, "learning_rate": 8.273026301648364e-05, "loss": 4.93, "step": 753350 }, { "epoch": 4.542878161141327, "grad_norm": 1.8708552122116089, "learning_rate": 8.271207572140095e-05, "loss": 4.7795, "step": 753400 }, { "epoch": 4.543179652922661, "grad_norm": 2.1867055892944336, "learning_rate": 8.269388966465643e-05, "loss": 4.9317, "step": 753450 }, { "epoch": 4.543481144703995, "grad_norm": 2.6530394554138184, "learning_rate": 8.267570484658482e-05, "loss": 5.1432, "step": 753500 }, { "epoch": 4.5437826364853295, "grad_norm": 2.5405313968658447, "learning_rate": 8.265752126752082e-05, "loss": 4.3593, "step": 753550 }, { "epoch": 4.544084128266664, "grad_norm": 3.3462257385253906, "learning_rate": 8.263933892779907e-05, "loss": 4.5081, "step": 753600 }, { "epoch": 4.544385620047997, "grad_norm": 2.9276487827301025, "learning_rate": 8.262115782775409e-05, "loss": 4.5127, "step": 753650 }, { "epoch": 4.544687111829331, "grad_norm": 1.6315993070602417, "learning_rate": 8.260297796772059e-05, "loss": 4.848, "step": 753700 }, { "epoch": 4.5449886036106655, "grad_norm": 8.655057907104492, "learning_rate": 8.258479934803304e-05, "loss": 4.8191, "step": 753750 }, { "epoch": 4.545290095392, "grad_norm": 2.44529390335083, "learning_rate": 8.256662196902608e-05, "loss": 4.7156, "step": 753800 }, { "epoch": 4.545591587173334, "grad_norm": 3.4031715393066406, "learning_rate": 8.254844583103424e-05, "loss": 4.9453, "step": 753850 }, { "epoch": 4.545893078954668, "grad_norm": 2.6261563301086426, "learning_rate": 8.253027093439196e-05, "loss": 4.6549, "step": 753900 }, { "epoch": 4.546194570736001, "grad_norm": 2.960768461227417, "learning_rate": 8.25120972794338e-05, "loss": 4.7876, "step": 753950 }, { "epoch": 4.546496062517336, "grad_norm": 2.950161933898926, "learning_rate": 8.249392486649418e-05, "loss": 4.8245, "step": 754000 }, { "epoch": 4.54679755429867, "grad_norm": 3.6837291717529297, "learning_rate": 8.247575369590748e-05, "loss": 4.8404, "step": 754050 }, { "epoch": 4.547099046080004, "grad_norm": 2.606853485107422, "learning_rate": 8.245758376800819e-05, "loss": 4.5206, "step": 754100 }, { "epoch": 4.547400537861338, "grad_norm": 3.7066540718078613, "learning_rate": 8.243941508313071e-05, "loss": 4.4156, "step": 754150 }, { "epoch": 4.547702029642672, "grad_norm": 3.3555455207824707, "learning_rate": 8.242124764160933e-05, "loss": 4.7394, "step": 754200 }, { "epoch": 4.548003521424006, "grad_norm": 1.020814299583435, "learning_rate": 8.240308144377847e-05, "loss": 4.9533, "step": 754250 }, { "epoch": 4.54830501320534, "grad_norm": 2.5697007179260254, "learning_rate": 8.238491648997247e-05, "loss": 4.5831, "step": 754300 }, { "epoch": 4.548606504986674, "grad_norm": 3.0674779415130615, "learning_rate": 8.236675278052554e-05, "loss": 4.531, "step": 754350 }, { "epoch": 4.548907996768008, "grad_norm": 2.708871603012085, "learning_rate": 8.234859031577207e-05, "loss": 4.7317, "step": 754400 }, { "epoch": 4.549209488549343, "grad_norm": 1.0774174928665161, "learning_rate": 8.233042909604623e-05, "loss": 4.4002, "step": 754450 }, { "epoch": 4.549510980330676, "grad_norm": 2.4309678077697754, "learning_rate": 8.231226912168222e-05, "loss": 4.6886, "step": 754500 }, { "epoch": 4.54981247211201, "grad_norm": 2.73060941696167, "learning_rate": 8.229411039301432e-05, "loss": 4.8312, "step": 754550 }, { "epoch": 4.550113963893344, "grad_norm": 2.150269031524658, "learning_rate": 8.227595291037673e-05, "loss": 4.6801, "step": 754600 }, { "epoch": 4.5504154556746785, "grad_norm": 3.365316390991211, "learning_rate": 8.225779667410354e-05, "loss": 5.0287, "step": 754650 }, { "epoch": 4.550716947456012, "grad_norm": 2.787454128265381, "learning_rate": 8.223964168452893e-05, "loss": 5.0358, "step": 754700 }, { "epoch": 4.551018439237346, "grad_norm": 3.365142583847046, "learning_rate": 8.222148794198705e-05, "loss": 5.1114, "step": 754750 }, { "epoch": 4.55131993101868, "grad_norm": 2.4149603843688965, "learning_rate": 8.220333544681198e-05, "loss": 5.2382, "step": 754800 }, { "epoch": 4.5516214228000145, "grad_norm": 2.674095630645752, "learning_rate": 8.218518419933772e-05, "loss": 4.9409, "step": 754850 }, { "epoch": 4.551922914581349, "grad_norm": 2.689054012298584, "learning_rate": 8.21670341998984e-05, "loss": 4.5772, "step": 754900 }, { "epoch": 4.552224406362683, "grad_norm": 1.9776568412780762, "learning_rate": 8.214888544882799e-05, "loss": 4.4867, "step": 754950 }, { "epoch": 4.552525898144017, "grad_norm": 1.132277011871338, "learning_rate": 8.213073794646048e-05, "loss": 4.7697, "step": 755000 }, { "epoch": 4.55282738992535, "grad_norm": 2.4807164669036865, "learning_rate": 8.211259169312995e-05, "loss": 4.7363, "step": 755050 }, { "epoch": 4.553128881706685, "grad_norm": 2.683889865875244, "learning_rate": 8.209444668917023e-05, "loss": 4.6605, "step": 755100 }, { "epoch": 4.553430373488019, "grad_norm": 3.3337254524230957, "learning_rate": 8.207630293491537e-05, "loss": 4.7913, "step": 755150 }, { "epoch": 4.553731865269353, "grad_norm": 1.6197280883789062, "learning_rate": 8.205816043069925e-05, "loss": 4.7471, "step": 755200 }, { "epoch": 4.554033357050686, "grad_norm": 1.4992889165878296, "learning_rate": 8.204001917685565e-05, "loss": 4.7154, "step": 755250 }, { "epoch": 4.5543348488320206, "grad_norm": 2.2510929107666016, "learning_rate": 8.20218791737185e-05, "loss": 4.634, "step": 755300 }, { "epoch": 4.554636340613355, "grad_norm": 2.873817205429077, "learning_rate": 8.200374042162173e-05, "loss": 4.7797, "step": 755350 }, { "epoch": 4.554937832394689, "grad_norm": 1.4662096500396729, "learning_rate": 8.1985602920899e-05, "loss": 4.4235, "step": 755400 }, { "epoch": 4.555239324176023, "grad_norm": 3.883739709854126, "learning_rate": 8.196746667188421e-05, "loss": 4.827, "step": 755450 }, { "epoch": 4.555540815957357, "grad_norm": 3.0783355236053467, "learning_rate": 8.194933167491117e-05, "loss": 4.4661, "step": 755500 }, { "epoch": 4.555842307738691, "grad_norm": 3.075284719467163, "learning_rate": 8.193119793031351e-05, "loss": 4.5572, "step": 755550 }, { "epoch": 4.556143799520025, "grad_norm": 0.878574788570404, "learning_rate": 8.191306543842505e-05, "loss": 4.6585, "step": 755600 }, { "epoch": 4.556445291301359, "grad_norm": 2.0707881450653076, "learning_rate": 8.189493419957946e-05, "loss": 4.4856, "step": 755650 }, { "epoch": 4.556746783082693, "grad_norm": 2.3026340007781982, "learning_rate": 8.187680421411036e-05, "loss": 4.5718, "step": 755700 }, { "epoch": 4.5570482748640275, "grad_norm": 2.3421883583068848, "learning_rate": 8.185867548235148e-05, "loss": 4.6717, "step": 755750 }, { "epoch": 4.557349766645361, "grad_norm": 2.9034740924835205, "learning_rate": 8.184054800463648e-05, "loss": 4.7047, "step": 755800 }, { "epoch": 4.557651258426695, "grad_norm": 3.7767715454101562, "learning_rate": 8.182242178129884e-05, "loss": 4.8909, "step": 755850 }, { "epoch": 4.557952750208029, "grad_norm": 4.7004475593566895, "learning_rate": 8.180429681267228e-05, "loss": 4.9554, "step": 755900 }, { "epoch": 4.5582542419893635, "grad_norm": 2.9725944995880127, "learning_rate": 8.178617309909035e-05, "loss": 4.2608, "step": 755950 }, { "epoch": 4.558555733770698, "grad_norm": 1.3797603845596313, "learning_rate": 8.176805064088655e-05, "loss": 4.3684, "step": 756000 }, { "epoch": 4.558555733770698, "eval_loss": 5.12050724029541, "eval_runtime": 39.014, "eval_samples_per_second": 13.123, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.6885030031017045, "step": 756000 }, { "epoch": 4.558857225552032, "grad_norm": 3.366334915161133, "learning_rate": 8.174992943839435e-05, "loss": 4.6367, "step": 756050 }, { "epoch": 4.559158717333365, "grad_norm": 2.707799196243286, "learning_rate": 8.173180949194736e-05, "loss": 4.886, "step": 756100 }, { "epoch": 4.559460209114699, "grad_norm": 3.169717788696289, "learning_rate": 8.171369080187894e-05, "loss": 4.8371, "step": 756150 }, { "epoch": 4.559761700896034, "grad_norm": 3.2287325859069824, "learning_rate": 8.169557336852259e-05, "loss": 4.4114, "step": 756200 }, { "epoch": 4.560063192677368, "grad_norm": 4.699334621429443, "learning_rate": 8.167745719221174e-05, "loss": 4.9844, "step": 756250 }, { "epoch": 4.560364684458702, "grad_norm": 2.712874412536621, "learning_rate": 8.165934227327984e-05, "loss": 4.8812, "step": 756300 }, { "epoch": 4.560666176240035, "grad_norm": 2.2303600311279297, "learning_rate": 8.164122861206023e-05, "loss": 4.6027, "step": 756350 }, { "epoch": 4.5609676680213695, "grad_norm": 2.7060062885284424, "learning_rate": 8.16231162088862e-05, "loss": 4.5228, "step": 756400 }, { "epoch": 4.561269159802704, "grad_norm": 2.5827362537384033, "learning_rate": 8.160500506409118e-05, "loss": 4.7461, "step": 756450 }, { "epoch": 4.561570651584038, "grad_norm": 2.9954710006713867, "learning_rate": 8.158689517800838e-05, "loss": 4.8781, "step": 756500 }, { "epoch": 4.561872143365372, "grad_norm": 3.6022627353668213, "learning_rate": 8.156878655097114e-05, "loss": 4.7806, "step": 756550 }, { "epoch": 4.562173635146706, "grad_norm": 1.9467953443527222, "learning_rate": 8.15506791833128e-05, "loss": 4.6902, "step": 756600 }, { "epoch": 4.56247512692804, "grad_norm": 0.9170435070991516, "learning_rate": 8.153257307536645e-05, "loss": 4.4404, "step": 756650 }, { "epoch": 4.562776618709374, "grad_norm": 2.4767069816589355, "learning_rate": 8.15144682274654e-05, "loss": 4.2303, "step": 756700 }, { "epoch": 4.563078110490708, "grad_norm": 3.596562147140503, "learning_rate": 8.149636463994287e-05, "loss": 4.4131, "step": 756750 }, { "epoch": 4.563379602272042, "grad_norm": 3.085144519805908, "learning_rate": 8.147826231313201e-05, "loss": 4.7187, "step": 756800 }, { "epoch": 4.563681094053376, "grad_norm": 2.6393275260925293, "learning_rate": 8.146016124736588e-05, "loss": 4.5097, "step": 756850 }, { "epoch": 4.56398258583471, "grad_norm": 3.653576374053955, "learning_rate": 8.144206144297772e-05, "loss": 4.6125, "step": 756900 }, { "epoch": 4.564284077616044, "grad_norm": 3.6290786266326904, "learning_rate": 8.142396290030053e-05, "loss": 5.1658, "step": 756950 }, { "epoch": 4.564585569397378, "grad_norm": 2.8741066455841064, "learning_rate": 8.140586561966743e-05, "loss": 4.789, "step": 757000 }, { "epoch": 4.5648870611787125, "grad_norm": 5.155853271484375, "learning_rate": 8.138776960141157e-05, "loss": 4.8373, "step": 757050 }, { "epoch": 4.565188552960047, "grad_norm": 2.7125155925750732, "learning_rate": 8.136967484586581e-05, "loss": 4.8716, "step": 757100 }, { "epoch": 4.56549004474138, "grad_norm": 2.916935682296753, "learning_rate": 8.135158135336332e-05, "loss": 4.3822, "step": 757150 }, { "epoch": 4.565791536522714, "grad_norm": 3.1928865909576416, "learning_rate": 8.133348912423699e-05, "loss": 4.8939, "step": 757200 }, { "epoch": 4.566093028304048, "grad_norm": 2.963505268096924, "learning_rate": 8.131539815881975e-05, "loss": 4.6161, "step": 757250 }, { "epoch": 4.566394520085383, "grad_norm": 3.2612688541412354, "learning_rate": 8.12973084574446e-05, "loss": 4.7726, "step": 757300 }, { "epoch": 4.566696011866717, "grad_norm": 2.522784471511841, "learning_rate": 8.12792200204445e-05, "loss": 4.6788, "step": 757350 }, { "epoch": 4.56699750364805, "grad_norm": 2.3420984745025635, "learning_rate": 8.12611328481522e-05, "loss": 4.6807, "step": 757400 }, { "epoch": 4.567298995429384, "grad_norm": 2.0824177265167236, "learning_rate": 8.124304694090068e-05, "loss": 4.6261, "step": 757450 }, { "epoch": 4.5676004872107185, "grad_norm": 1.553987741470337, "learning_rate": 8.12249622990228e-05, "loss": 4.405, "step": 757500 }, { "epoch": 4.567901978992053, "grad_norm": 3.5074808597564697, "learning_rate": 8.120687892285131e-05, "loss": 4.9725, "step": 757550 }, { "epoch": 4.568203470773387, "grad_norm": 3.6612167358398438, "learning_rate": 8.118879681271907e-05, "loss": 4.7664, "step": 757600 }, { "epoch": 4.568504962554721, "grad_norm": 4.146574974060059, "learning_rate": 8.117071596895885e-05, "loss": 4.7152, "step": 757650 }, { "epoch": 4.5688064543360545, "grad_norm": 4.312441825866699, "learning_rate": 8.11526363919033e-05, "loss": 4.7707, "step": 757700 }, { "epoch": 4.569107946117389, "grad_norm": 2.5124669075012207, "learning_rate": 8.113455808188524e-05, "loss": 4.6354, "step": 757750 }, { "epoch": 4.569409437898723, "grad_norm": 5.426657676696777, "learning_rate": 8.111648103923742e-05, "loss": 4.8523, "step": 757800 }, { "epoch": 4.569710929680057, "grad_norm": 2.8015551567077637, "learning_rate": 8.10984052642924e-05, "loss": 4.715, "step": 757850 }, { "epoch": 4.570012421461391, "grad_norm": 3.99891996383667, "learning_rate": 8.108033075738291e-05, "loss": 4.3106, "step": 757900 }, { "epoch": 4.570313913242725, "grad_norm": 1.929419994354248, "learning_rate": 8.106225751884164e-05, "loss": 5.0969, "step": 757950 }, { "epoch": 4.570615405024059, "grad_norm": 2.080101251602173, "learning_rate": 8.104418554900114e-05, "loss": 4.6926, "step": 758000 }, { "epoch": 4.570916896805393, "grad_norm": 2.006211280822754, "learning_rate": 8.102611484819392e-05, "loss": 4.7343, "step": 758050 }, { "epoch": 4.571218388586727, "grad_norm": 2.1422736644744873, "learning_rate": 8.100804541675271e-05, "loss": 4.8992, "step": 758100 }, { "epoch": 4.5715198803680615, "grad_norm": 3.5187151432037354, "learning_rate": 8.098997725500989e-05, "loss": 4.4445, "step": 758150 }, { "epoch": 4.571821372149396, "grad_norm": 2.0692031383514404, "learning_rate": 8.097191036329807e-05, "loss": 5.0369, "step": 758200 }, { "epoch": 4.572122863930729, "grad_norm": 2.6156833171844482, "learning_rate": 8.095384474194978e-05, "loss": 4.4673, "step": 758250 }, { "epoch": 4.572424355712063, "grad_norm": 0.7935271859169006, "learning_rate": 8.093578039129738e-05, "loss": 4.8941, "step": 758300 }, { "epoch": 4.572725847493397, "grad_norm": 3.0361907482147217, "learning_rate": 8.091771731167345e-05, "loss": 4.8227, "step": 758350 }, { "epoch": 4.573027339274732, "grad_norm": 3.2149059772491455, "learning_rate": 8.089965550341035e-05, "loss": 4.6943, "step": 758400 }, { "epoch": 4.573328831056065, "grad_norm": 4.2580084800720215, "learning_rate": 8.088159496684042e-05, "loss": 4.7415, "step": 758450 }, { "epoch": 4.573630322837399, "grad_norm": 3.114405393600464, "learning_rate": 8.08635357022961e-05, "loss": 4.7009, "step": 758500 }, { "epoch": 4.573931814618733, "grad_norm": 1.2347112894058228, "learning_rate": 8.084547771010978e-05, "loss": 4.7294, "step": 758550 }, { "epoch": 4.5742333064000675, "grad_norm": 2.3981027603149414, "learning_rate": 8.082742099061374e-05, "loss": 4.8066, "step": 758600 }, { "epoch": 4.574534798181402, "grad_norm": 3.777928590774536, "learning_rate": 8.080936554414025e-05, "loss": 4.6542, "step": 758650 }, { "epoch": 4.574836289962736, "grad_norm": 4.044412612915039, "learning_rate": 8.079131137102173e-05, "loss": 4.3632, "step": 758700 }, { "epoch": 4.57513778174407, "grad_norm": 3.722835063934326, "learning_rate": 8.07732584715903e-05, "loss": 4.6404, "step": 758750 }, { "epoch": 4.5754392735254035, "grad_norm": 1.0857561826705933, "learning_rate": 8.075520684617831e-05, "loss": 4.5701, "step": 758800 }, { "epoch": 4.575740765306738, "grad_norm": 4.3326640129089355, "learning_rate": 8.073715649511794e-05, "loss": 4.3246, "step": 758850 }, { "epoch": 4.576042257088072, "grad_norm": 3.295388698577881, "learning_rate": 8.071910741874127e-05, "loss": 4.6325, "step": 758900 }, { "epoch": 4.576343748869406, "grad_norm": 2.6359689235687256, "learning_rate": 8.070105961738057e-05, "loss": 4.5054, "step": 758950 }, { "epoch": 4.576645240650739, "grad_norm": 1.85922372341156, "learning_rate": 8.068301309136804e-05, "loss": 4.6885, "step": 759000 }, { "epoch": 4.576645240650739, "eval_loss": 5.129559516906738, "eval_runtime": 38.9922, "eval_samples_per_second": 13.131, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.691297834776022, "step": 759000 }, { "epoch": 4.576946732432074, "grad_norm": 2.1751112937927246, "learning_rate": 8.066496784103568e-05, "loss": 4.88, "step": 759050 }, { "epoch": 4.577248224213408, "grad_norm": 3.2350592613220215, "learning_rate": 8.064692386671562e-05, "loss": 4.5521, "step": 759100 }, { "epoch": 4.577549715994742, "grad_norm": 4.909759044647217, "learning_rate": 8.062888116874004e-05, "loss": 4.5089, "step": 759150 }, { "epoch": 4.577851207776076, "grad_norm": 2.8204877376556396, "learning_rate": 8.061083974744087e-05, "loss": 5.0866, "step": 759200 }, { "epoch": 4.57815269955741, "grad_norm": 3.6908040046691895, "learning_rate": 8.059279960315013e-05, "loss": 4.6238, "step": 759250 }, { "epoch": 4.578454191338744, "grad_norm": 1.539870023727417, "learning_rate": 8.057476073619992e-05, "loss": 4.7569, "step": 759300 }, { "epoch": 4.578755683120078, "grad_norm": 3.4361095428466797, "learning_rate": 8.055672314692211e-05, "loss": 4.8357, "step": 759350 }, { "epoch": 4.579057174901412, "grad_norm": 2.7458457946777344, "learning_rate": 8.05386868356487e-05, "loss": 4.6615, "step": 759400 }, { "epoch": 4.579358666682746, "grad_norm": 4.038604259490967, "learning_rate": 8.05206518027117e-05, "loss": 4.7935, "step": 759450 }, { "epoch": 4.579660158464081, "grad_norm": 0.6182407736778259, "learning_rate": 8.050261804844289e-05, "loss": 4.8629, "step": 759500 }, { "epoch": 4.579961650245414, "grad_norm": 2.5326154232025146, "learning_rate": 8.048458557317428e-05, "loss": 4.666, "step": 759550 }, { "epoch": 4.580263142026748, "grad_norm": 2.6177539825439453, "learning_rate": 8.046655437723761e-05, "loss": 4.5708, "step": 759600 }, { "epoch": 4.580564633808082, "grad_norm": 4.303219795227051, "learning_rate": 8.044852446096484e-05, "loss": 4.5333, "step": 759650 }, { "epoch": 4.5808661255894165, "grad_norm": 3.138608694076538, "learning_rate": 8.043049582468764e-05, "loss": 4.688, "step": 759700 }, { "epoch": 4.581167617370751, "grad_norm": 1.7045117616653442, "learning_rate": 8.041246846873791e-05, "loss": 4.4483, "step": 759750 }, { "epoch": 4.581469109152085, "grad_norm": 1.7186307907104492, "learning_rate": 8.039444239344744e-05, "loss": 4.6212, "step": 759800 }, { "epoch": 4.581770600933418, "grad_norm": 2.4657843112945557, "learning_rate": 8.037641759914787e-05, "loss": 4.9047, "step": 759850 }, { "epoch": 4.5820720927147525, "grad_norm": 2.326476573944092, "learning_rate": 8.035839408617096e-05, "loss": 4.7034, "step": 759900 }, { "epoch": 4.582373584496087, "grad_norm": 1.3745403289794922, "learning_rate": 8.034037185484849e-05, "loss": 4.9969, "step": 759950 }, { "epoch": 4.582675076277421, "grad_norm": 3.239128589630127, "learning_rate": 8.032235090551207e-05, "loss": 5.163, "step": 760000 }, { "epoch": 4.582976568058755, "grad_norm": 3.0174429416656494, "learning_rate": 8.030433123849328e-05, "loss": 4.7679, "step": 760050 }, { "epoch": 4.583278059840088, "grad_norm": 6.125843524932861, "learning_rate": 8.02863128541239e-05, "loss": 4.4405, "step": 760100 }, { "epoch": 4.583579551621423, "grad_norm": 4.110848903656006, "learning_rate": 8.026829575273533e-05, "loss": 4.5901, "step": 760150 }, { "epoch": 4.583881043402757, "grad_norm": 3.1113781929016113, "learning_rate": 8.025027993465931e-05, "loss": 4.7231, "step": 760200 }, { "epoch": 4.584182535184091, "grad_norm": 3.616697311401367, "learning_rate": 8.023226540022739e-05, "loss": 4.2316, "step": 760250 }, { "epoch": 4.584484026965425, "grad_norm": 2.4912967681884766, "learning_rate": 8.021425214977101e-05, "loss": 4.3637, "step": 760300 }, { "epoch": 4.584785518746759, "grad_norm": 3.7503323554992676, "learning_rate": 8.019624018362179e-05, "loss": 4.724, "step": 760350 }, { "epoch": 4.585087010528093, "grad_norm": 3.132760763168335, "learning_rate": 8.017822950211115e-05, "loss": 4.4454, "step": 760400 }, { "epoch": 4.585388502309427, "grad_norm": 1.1971485614776611, "learning_rate": 8.01602201055705e-05, "loss": 4.3174, "step": 760450 }, { "epoch": 4.585689994090761, "grad_norm": 1.3901019096374512, "learning_rate": 8.014221199433133e-05, "loss": 4.7904, "step": 760500 }, { "epoch": 4.585991485872095, "grad_norm": 3.536504030227661, "learning_rate": 8.012420516872513e-05, "loss": 4.5037, "step": 760550 }, { "epoch": 4.586292977653429, "grad_norm": 2.5248870849609375, "learning_rate": 8.010619962908313e-05, "loss": 4.7597, "step": 760600 }, { "epoch": 4.586594469434763, "grad_norm": 2.419420003890991, "learning_rate": 8.008819537573678e-05, "loss": 4.8577, "step": 760650 }, { "epoch": 4.586895961216097, "grad_norm": 4.9784088134765625, "learning_rate": 8.007019240901752e-05, "loss": 4.6916, "step": 760700 }, { "epoch": 4.587197452997431, "grad_norm": 3.0361390113830566, "learning_rate": 8.005219072925654e-05, "loss": 4.4901, "step": 760750 }, { "epoch": 4.5874989447787655, "grad_norm": 2.933669328689575, "learning_rate": 8.003419033678514e-05, "loss": 4.5938, "step": 760800 }, { "epoch": 4.5878004365601, "grad_norm": 3.235280990600586, "learning_rate": 8.001619123193464e-05, "loss": 5.0228, "step": 760850 }, { "epoch": 4.588101928341433, "grad_norm": 3.9837374687194824, "learning_rate": 7.999819341503623e-05, "loss": 4.7713, "step": 760900 }, { "epoch": 4.588403420122767, "grad_norm": 1.9115294218063354, "learning_rate": 7.998019688642118e-05, "loss": 4.62, "step": 760950 }, { "epoch": 4.5887049119041015, "grad_norm": 2.863248825073242, "learning_rate": 7.996220164642072e-05, "loss": 4.4221, "step": 761000 }, { "epoch": 4.589006403685436, "grad_norm": 2.7794783115386963, "learning_rate": 7.994420769536594e-05, "loss": 4.4511, "step": 761050 }, { "epoch": 4.58930789546677, "grad_norm": 2.36889910697937, "learning_rate": 7.992621503358806e-05, "loss": 4.5303, "step": 761100 }, { "epoch": 4.589609387248103, "grad_norm": 2.4945244789123535, "learning_rate": 7.990822366141822e-05, "loss": 4.7503, "step": 761150 }, { "epoch": 4.589910879029437, "grad_norm": 2.445683479309082, "learning_rate": 7.989023357918751e-05, "loss": 4.6474, "step": 761200 }, { "epoch": 4.590212370810772, "grad_norm": 2.9447779655456543, "learning_rate": 7.987224478722696e-05, "loss": 4.8348, "step": 761250 }, { "epoch": 4.590513862592106, "grad_norm": 2.5714895725250244, "learning_rate": 7.98542572858677e-05, "loss": 4.7536, "step": 761300 }, { "epoch": 4.59081535437344, "grad_norm": 0.9975821375846863, "learning_rate": 7.98362710754407e-05, "loss": 4.6192, "step": 761350 }, { "epoch": 4.591116846154774, "grad_norm": 1.3350354433059692, "learning_rate": 7.981828615627695e-05, "loss": 4.5711, "step": 761400 }, { "epoch": 4.5914183379361075, "grad_norm": 3.3464980125427246, "learning_rate": 7.98003025287076e-05, "loss": 4.891, "step": 761450 }, { "epoch": 4.591719829717442, "grad_norm": 1.1179687976837158, "learning_rate": 7.978232019306343e-05, "loss": 4.7467, "step": 761500 }, { "epoch": 4.592021321498776, "grad_norm": 3.9684267044067383, "learning_rate": 7.976433914967551e-05, "loss": 4.4619, "step": 761550 }, { "epoch": 4.59232281328011, "grad_norm": 1.3496062755584717, "learning_rate": 7.974635939887467e-05, "loss": 4.4465, "step": 761600 }, { "epoch": 4.592624305061444, "grad_norm": 2.609550714492798, "learning_rate": 7.97283809409918e-05, "loss": 4.4305, "step": 761650 }, { "epoch": 4.592925796842778, "grad_norm": 1.7074296474456787, "learning_rate": 7.971040377635779e-05, "loss": 4.2601, "step": 761700 }, { "epoch": 4.593227288624112, "grad_norm": 3.244076728820801, "learning_rate": 7.969242790530353e-05, "loss": 4.836, "step": 761750 }, { "epoch": 4.593528780405446, "grad_norm": 2.495296001434326, "learning_rate": 7.967445332815974e-05, "loss": 4.4862, "step": 761800 }, { "epoch": 4.59383027218678, "grad_norm": 3.0693559646606445, "learning_rate": 7.96564800452573e-05, "loss": 4.678, "step": 761850 }, { "epoch": 4.5941317639681145, "grad_norm": 2.233670949935913, "learning_rate": 7.963850805692701e-05, "loss": 4.5635, "step": 761900 }, { "epoch": 4.594433255749449, "grad_norm": 2.315990686416626, "learning_rate": 7.962053736349958e-05, "loss": 4.7011, "step": 761950 }, { "epoch": 4.594734747530782, "grad_norm": 2.2530901432037354, "learning_rate": 7.960256796530565e-05, "loss": 4.8316, "step": 762000 }, { "epoch": 4.594734747530782, "eval_loss": 5.117925643920898, "eval_runtime": 38.9661, "eval_samples_per_second": 13.14, "eval_steps_per_second": 6.57, "eval_tts_loss": 7.716637347117388, "step": 762000 }, { "epoch": 4.595036239312116, "grad_norm": 2.9765520095825195, "learning_rate": 7.958459986267604e-05, "loss": 4.9567, "step": 762050 }, { "epoch": 4.5953377310934505, "grad_norm": 2.3829987049102783, "learning_rate": 7.956663305594134e-05, "loss": 4.8493, "step": 762100 }, { "epoch": 4.595639222874785, "grad_norm": 2.4268908500671387, "learning_rate": 7.954866754543224e-05, "loss": 4.5758, "step": 762150 }, { "epoch": 4.595940714656118, "grad_norm": 3.2409305572509766, "learning_rate": 7.953070333147945e-05, "loss": 4.6435, "step": 762200 }, { "epoch": 4.596242206437452, "grad_norm": 2.96848201751709, "learning_rate": 7.951274041441342e-05, "loss": 4.6077, "step": 762250 }, { "epoch": 4.596543698218786, "grad_norm": 2.5946555137634277, "learning_rate": 7.949477879456482e-05, "loss": 5.1022, "step": 762300 }, { "epoch": 4.596845190000121, "grad_norm": 2.9243123531341553, "learning_rate": 7.947681847226426e-05, "loss": 4.947, "step": 762350 }, { "epoch": 4.597146681781455, "grad_norm": 2.0949831008911133, "learning_rate": 7.94588594478422e-05, "loss": 4.6286, "step": 762400 }, { "epoch": 4.597448173562789, "grad_norm": 2.742814779281616, "learning_rate": 7.944090172162911e-05, "loss": 4.4797, "step": 762450 }, { "epoch": 4.597749665344123, "grad_norm": 2.99699068069458, "learning_rate": 7.942294529395559e-05, "loss": 4.5256, "step": 762500 }, { "epoch": 4.5980511571254565, "grad_norm": 2.7313601970672607, "learning_rate": 7.9404990165152e-05, "loss": 4.4753, "step": 762550 }, { "epoch": 4.598352648906791, "grad_norm": 2.485973596572876, "learning_rate": 7.93870363355488e-05, "loss": 4.5058, "step": 762600 }, { "epoch": 4.598654140688125, "grad_norm": 3.180973768234253, "learning_rate": 7.93690838054765e-05, "loss": 4.637, "step": 762650 }, { "epoch": 4.598955632469459, "grad_norm": 2.1854584217071533, "learning_rate": 7.935113257526535e-05, "loss": 4.6906, "step": 762700 }, { "epoch": 4.5992571242507925, "grad_norm": 2.971050977706909, "learning_rate": 7.933318264524586e-05, "loss": 4.6486, "step": 762750 }, { "epoch": 4.599558616032127, "grad_norm": 1.5614900588989258, "learning_rate": 7.931523401574828e-05, "loss": 4.4674, "step": 762800 }, { "epoch": 4.599860107813461, "grad_norm": 2.712334156036377, "learning_rate": 7.929728668710289e-05, "loss": 4.7275, "step": 762850 }, { "epoch": 4.600161599594795, "grad_norm": 2.962341070175171, "learning_rate": 7.927934065964005e-05, "loss": 4.537, "step": 762900 }, { "epoch": 4.600463091376129, "grad_norm": 4.022374153137207, "learning_rate": 7.926139593369e-05, "loss": 4.224, "step": 762950 }, { "epoch": 4.6007645831574635, "grad_norm": 3.0029256343841553, "learning_rate": 7.924345250958308e-05, "loss": 4.6179, "step": 763000 }, { "epoch": 4.601066074938797, "grad_norm": 3.757286548614502, "learning_rate": 7.922551038764937e-05, "loss": 4.4877, "step": 763050 }, { "epoch": 4.601367566720131, "grad_norm": 3.9440689086914062, "learning_rate": 7.920756956821924e-05, "loss": 4.6902, "step": 763100 }, { "epoch": 4.601669058501465, "grad_norm": 2.4688594341278076, "learning_rate": 7.918963005162267e-05, "loss": 4.4021, "step": 763150 }, { "epoch": 4.601970550282799, "grad_norm": 2.6205923557281494, "learning_rate": 7.917169183818996e-05, "loss": 4.971, "step": 763200 }, { "epoch": 4.602272042064134, "grad_norm": 3.022172451019287, "learning_rate": 7.915375492825113e-05, "loss": 4.6362, "step": 763250 }, { "epoch": 4.602573533845467, "grad_norm": 2.9968385696411133, "learning_rate": 7.91358193221364e-05, "loss": 5.1097, "step": 763300 }, { "epoch": 4.602875025626801, "grad_norm": 3.2190463542938232, "learning_rate": 7.911788502017575e-05, "loss": 4.844, "step": 763350 }, { "epoch": 4.603176517408135, "grad_norm": 3.166452407836914, "learning_rate": 7.909995202269926e-05, "loss": 4.9102, "step": 763400 }, { "epoch": 4.60347800918947, "grad_norm": 2.322270154953003, "learning_rate": 7.908202033003701e-05, "loss": 4.547, "step": 763450 }, { "epoch": 4.603779500970804, "grad_norm": 2.5380702018737793, "learning_rate": 7.906408994251894e-05, "loss": 4.3716, "step": 763500 }, { "epoch": 4.604080992752138, "grad_norm": 3.6641581058502197, "learning_rate": 7.904616086047512e-05, "loss": 4.6814, "step": 763550 }, { "epoch": 4.604382484533471, "grad_norm": 2.932267665863037, "learning_rate": 7.902823308423546e-05, "loss": 4.3849, "step": 763600 }, { "epoch": 4.6046839763148055, "grad_norm": 2.358616352081299, "learning_rate": 7.901030661412979e-05, "loss": 4.7116, "step": 763650 }, { "epoch": 4.60498546809614, "grad_norm": 3.4610350131988525, "learning_rate": 7.899238145048814e-05, "loss": 4.7458, "step": 763700 }, { "epoch": 4.605286959877474, "grad_norm": 4.164801597595215, "learning_rate": 7.897445759364042e-05, "loss": 4.4203, "step": 763750 }, { "epoch": 4.605588451658808, "grad_norm": 3.550203323364258, "learning_rate": 7.895653504391643e-05, "loss": 4.7666, "step": 763800 }, { "epoch": 4.6058899434401415, "grad_norm": 0.5163931250572205, "learning_rate": 7.893861380164601e-05, "loss": 4.6803, "step": 763850 }, { "epoch": 4.606191435221476, "grad_norm": 2.4787542819976807, "learning_rate": 7.892069386715904e-05, "loss": 4.6354, "step": 763900 }, { "epoch": 4.60649292700281, "grad_norm": 3.67620849609375, "learning_rate": 7.890277524078527e-05, "loss": 4.9765, "step": 763950 }, { "epoch": 4.606794418784144, "grad_norm": 2.5422956943511963, "learning_rate": 7.888485792285442e-05, "loss": 4.8822, "step": 764000 }, { "epoch": 4.607095910565478, "grad_norm": 2.477033853530884, "learning_rate": 7.886694191369629e-05, "loss": 4.8695, "step": 764050 }, { "epoch": 4.6073974023468125, "grad_norm": 2.6190478801727295, "learning_rate": 7.884902721364055e-05, "loss": 4.5474, "step": 764100 }, { "epoch": 4.607698894128146, "grad_norm": 2.670217275619507, "learning_rate": 7.883111382301692e-05, "loss": 4.7526, "step": 764150 }, { "epoch": 4.60800038590948, "grad_norm": 2.6068899631500244, "learning_rate": 7.881320174215515e-05, "loss": 4.7579, "step": 764200 }, { "epoch": 4.608301877690814, "grad_norm": 2.7319400310516357, "learning_rate": 7.879529097138474e-05, "loss": 4.8797, "step": 764250 }, { "epoch": 4.608603369472148, "grad_norm": 3.9926884174346924, "learning_rate": 7.877738151103539e-05, "loss": 4.8274, "step": 764300 }, { "epoch": 4.608904861253482, "grad_norm": 2.140249013900757, "learning_rate": 7.875947336143675e-05, "loss": 4.9937, "step": 764350 }, { "epoch": 4.609206353034816, "grad_norm": 3.806675672531128, "learning_rate": 7.874156652291834e-05, "loss": 4.533, "step": 764400 }, { "epoch": 4.60950784481615, "grad_norm": 2.9197499752044678, "learning_rate": 7.872366099580965e-05, "loss": 4.8107, "step": 764450 }, { "epoch": 4.609809336597484, "grad_norm": 1.455593228340149, "learning_rate": 7.870575678044032e-05, "loss": 4.0558, "step": 764500 }, { "epoch": 4.610110828378819, "grad_norm": 3.8245468139648438, "learning_rate": 7.868785387713973e-05, "loss": 4.6725, "step": 764550 }, { "epoch": 4.610412320160153, "grad_norm": 2.4058291912078857, "learning_rate": 7.866995228623742e-05, "loss": 4.7314, "step": 764600 }, { "epoch": 4.610713811941486, "grad_norm": 2.7870538234710693, "learning_rate": 7.865205200806291e-05, "loss": 4.9002, "step": 764650 }, { "epoch": 4.61101530372282, "grad_norm": 3.7666919231414795, "learning_rate": 7.86341530429455e-05, "loss": 4.6334, "step": 764700 }, { "epoch": 4.6113167955041545, "grad_norm": 2.5480494499206543, "learning_rate": 7.861625539121471e-05, "loss": 4.2864, "step": 764750 }, { "epoch": 4.611618287285489, "grad_norm": 4.082790851593018, "learning_rate": 7.859835905319989e-05, "loss": 4.7497, "step": 764800 }, { "epoch": 4.611919779066823, "grad_norm": 3.2943882942199707, "learning_rate": 7.85804640292303e-05, "loss": 4.7693, "step": 764850 }, { "epoch": 4.612221270848156, "grad_norm": 3.029055118560791, "learning_rate": 7.856257031963536e-05, "loss": 4.8085, "step": 764900 }, { "epoch": 4.6125227626294905, "grad_norm": 2.2762558460235596, "learning_rate": 7.854467792474442e-05, "loss": 4.6063, "step": 764950 }, { "epoch": 4.612824254410825, "grad_norm": 2.562747001647949, "learning_rate": 7.852678684488666e-05, "loss": 4.6784, "step": 765000 }, { "epoch": 4.612824254410825, "eval_loss": 5.129087448120117, "eval_runtime": 38.86, "eval_samples_per_second": 13.175, "eval_steps_per_second": 6.588, "eval_tts_loss": 7.644521262131139, "step": 765000 }, { "epoch": 4.613125746192159, "grad_norm": 2.821428060531616, "learning_rate": 7.850889708039138e-05, "loss": 4.5718, "step": 765050 }, { "epoch": 4.613427237973493, "grad_norm": 2.491570234298706, "learning_rate": 7.84910086315879e-05, "loss": 4.446, "step": 765100 }, { "epoch": 4.613728729754827, "grad_norm": 3.3814454078674316, "learning_rate": 7.847312149880533e-05, "loss": 4.7257, "step": 765150 }, { "epoch": 4.614030221536161, "grad_norm": 2.6119303703308105, "learning_rate": 7.845523568237284e-05, "loss": 4.6732, "step": 765200 }, { "epoch": 4.614331713317495, "grad_norm": 3.6845200061798096, "learning_rate": 7.84373511826197e-05, "loss": 5.022, "step": 765250 }, { "epoch": 4.614633205098829, "grad_norm": 1.9829941987991333, "learning_rate": 7.841946799987493e-05, "loss": 4.6281, "step": 765300 }, { "epoch": 4.614934696880163, "grad_norm": 2.8045272827148438, "learning_rate": 7.840158613446769e-05, "loss": 4.6112, "step": 765350 }, { "epoch": 4.615236188661497, "grad_norm": 3.164095401763916, "learning_rate": 7.838370558672713e-05, "loss": 4.7071, "step": 765400 }, { "epoch": 4.615537680442831, "grad_norm": 3.709913492202759, "learning_rate": 7.836582635698223e-05, "loss": 4.6011, "step": 765450 }, { "epoch": 4.615839172224165, "grad_norm": 2.123359203338623, "learning_rate": 7.834794844556207e-05, "loss": 4.8674, "step": 765500 }, { "epoch": 4.616140664005499, "grad_norm": 1.2031561136245728, "learning_rate": 7.83300718527957e-05, "loss": 4.6147, "step": 765550 }, { "epoch": 4.616442155786833, "grad_norm": 2.4913387298583984, "learning_rate": 7.83121965790121e-05, "loss": 5.0616, "step": 765600 }, { "epoch": 4.616743647568168, "grad_norm": 2.8181042671203613, "learning_rate": 7.829432262454015e-05, "loss": 4.8258, "step": 765650 }, { "epoch": 4.617045139349502, "grad_norm": 3.4773125648498535, "learning_rate": 7.827644998970893e-05, "loss": 4.6583, "step": 765700 }, { "epoch": 4.617346631130835, "grad_norm": 3.5117595195770264, "learning_rate": 7.825857867484723e-05, "loss": 4.9237, "step": 765750 }, { "epoch": 4.617648122912169, "grad_norm": 2.175703763961792, "learning_rate": 7.8240708680284e-05, "loss": 4.729, "step": 765800 }, { "epoch": 4.6179496146935035, "grad_norm": 4.211132049560547, "learning_rate": 7.82228400063482e-05, "loss": 4.8032, "step": 765850 }, { "epoch": 4.618251106474838, "grad_norm": 3.1887595653533936, "learning_rate": 7.820497265336852e-05, "loss": 4.7185, "step": 765900 }, { "epoch": 4.618552598256171, "grad_norm": 1.6684691905975342, "learning_rate": 7.818710662167393e-05, "loss": 4.1437, "step": 765950 }, { "epoch": 4.618854090037505, "grad_norm": 3.75080943107605, "learning_rate": 7.816924191159316e-05, "loss": 4.9735, "step": 766000 }, { "epoch": 4.6191555818188395, "grad_norm": 2.989879608154297, "learning_rate": 7.815137852345492e-05, "loss": 4.9189, "step": 766050 }, { "epoch": 4.619457073600174, "grad_norm": 1.2482821941375732, "learning_rate": 7.813351645758803e-05, "loss": 4.8712, "step": 766100 }, { "epoch": 4.619758565381508, "grad_norm": 1.8679636716842651, "learning_rate": 7.811565571432125e-05, "loss": 4.8474, "step": 766150 }, { "epoch": 4.620060057162842, "grad_norm": 3.1586124897003174, "learning_rate": 7.809779629398322e-05, "loss": 4.7642, "step": 766200 }, { "epoch": 4.620361548944176, "grad_norm": 2.6601507663726807, "learning_rate": 7.807993819690261e-05, "loss": 4.5496, "step": 766250 }, { "epoch": 4.62066304072551, "grad_norm": 1.2690634727478027, "learning_rate": 7.806208142340816e-05, "loss": 3.9305, "step": 766300 }, { "epoch": 4.620964532506844, "grad_norm": 2.4914755821228027, "learning_rate": 7.804422597382846e-05, "loss": 4.4749, "step": 766350 }, { "epoch": 4.621266024288178, "grad_norm": 3.54878306388855, "learning_rate": 7.802637184849204e-05, "loss": 4.9268, "step": 766400 }, { "epoch": 4.621567516069512, "grad_norm": 3.171349048614502, "learning_rate": 7.800851904772752e-05, "loss": 4.9374, "step": 766450 }, { "epoch": 4.6218690078508455, "grad_norm": 2.2436585426330566, "learning_rate": 7.799066757186351e-05, "loss": 5.0122, "step": 766500 }, { "epoch": 4.62217049963218, "grad_norm": 1.2281628847122192, "learning_rate": 7.797281742122846e-05, "loss": 4.5878, "step": 766550 }, { "epoch": 4.622471991413514, "grad_norm": 2.169161081314087, "learning_rate": 7.795496859615092e-05, "loss": 4.6974, "step": 766600 }, { "epoch": 4.622773483194848, "grad_norm": 2.2873401641845703, "learning_rate": 7.793712109695944e-05, "loss": 4.4752, "step": 766650 }, { "epoch": 4.623074974976182, "grad_norm": 2.7063345909118652, "learning_rate": 7.791927492398233e-05, "loss": 4.6679, "step": 766700 }, { "epoch": 4.623376466757517, "grad_norm": 3.3972887992858887, "learning_rate": 7.790143007754816e-05, "loss": 4.2752, "step": 766750 }, { "epoch": 4.62367795853885, "grad_norm": 3.8203392028808594, "learning_rate": 7.788358655798527e-05, "loss": 4.818, "step": 766800 }, { "epoch": 4.623979450320184, "grad_norm": 3.8035666942596436, "learning_rate": 7.786574436562201e-05, "loss": 4.7247, "step": 766850 }, { "epoch": 4.624280942101518, "grad_norm": 2.1439104080200195, "learning_rate": 7.78479035007868e-05, "loss": 4.9521, "step": 766900 }, { "epoch": 4.6245824338828525, "grad_norm": 2.5174479484558105, "learning_rate": 7.7830063963808e-05, "loss": 4.625, "step": 766950 }, { "epoch": 4.624883925664187, "grad_norm": 6.589539527893066, "learning_rate": 7.781222575501381e-05, "loss": 4.8466, "step": 767000 }, { "epoch": 4.62518541744552, "grad_norm": 3.2629079818725586, "learning_rate": 7.77943888747326e-05, "loss": 4.6342, "step": 767050 }, { "epoch": 4.625486909226854, "grad_norm": 2.7665531635284424, "learning_rate": 7.777655332329268e-05, "loss": 4.586, "step": 767100 }, { "epoch": 4.625788401008188, "grad_norm": 2.854776620864868, "learning_rate": 7.775871910102224e-05, "loss": 4.6297, "step": 767150 }, { "epoch": 4.626089892789523, "grad_norm": 3.2927238941192627, "learning_rate": 7.774088620824939e-05, "loss": 4.5445, "step": 767200 }, { "epoch": 4.626391384570857, "grad_norm": 2.2196688652038574, "learning_rate": 7.77230546453025e-05, "loss": 4.6926, "step": 767250 }, { "epoch": 4.626692876352191, "grad_norm": 2.4925081729888916, "learning_rate": 7.770522441250957e-05, "loss": 4.6441, "step": 767300 }, { "epoch": 4.626994368133524, "grad_norm": 3.120237350463867, "learning_rate": 7.76873955101988e-05, "loss": 4.7307, "step": 767350 }, { "epoch": 4.627295859914859, "grad_norm": 2.1709415912628174, "learning_rate": 7.76695679386984e-05, "loss": 4.7684, "step": 767400 }, { "epoch": 4.627597351696193, "grad_norm": 3.626418352127075, "learning_rate": 7.765174169833633e-05, "loss": 4.8392, "step": 767450 }, { "epoch": 4.627898843477527, "grad_norm": 3.379455089569092, "learning_rate": 7.763391678944076e-05, "loss": 5.0581, "step": 767500 }, { "epoch": 4.628200335258861, "grad_norm": 2.8925857543945312, "learning_rate": 7.761609321233966e-05, "loss": 4.5853, "step": 767550 }, { "epoch": 4.6285018270401945, "grad_norm": 3.545506715774536, "learning_rate": 7.759827096736101e-05, "loss": 4.6293, "step": 767600 }, { "epoch": 4.628803318821529, "grad_norm": 1.2855823040008545, "learning_rate": 7.758045005483288e-05, "loss": 4.5756, "step": 767650 }, { "epoch": 4.629104810602863, "grad_norm": 1.9374817609786987, "learning_rate": 7.756263047508326e-05, "loss": 4.5369, "step": 767700 }, { "epoch": 4.629406302384197, "grad_norm": 3.6711902618408203, "learning_rate": 7.754481222844001e-05, "loss": 4.7667, "step": 767750 }, { "epoch": 4.629707794165531, "grad_norm": 1.7964719533920288, "learning_rate": 7.752699531523109e-05, "loss": 4.7088, "step": 767800 }, { "epoch": 4.630009285946866, "grad_norm": 2.527726650238037, "learning_rate": 7.750917973578445e-05, "loss": 5.0373, "step": 767850 }, { "epoch": 4.630310777728199, "grad_norm": 0.9267692565917969, "learning_rate": 7.749136549042785e-05, "loss": 4.8287, "step": 767900 }, { "epoch": 4.630612269509533, "grad_norm": 3.814359426498413, "learning_rate": 7.747355257948924e-05, "loss": 4.7247, "step": 767950 }, { "epoch": 4.630913761290867, "grad_norm": 3.0162851810455322, "learning_rate": 7.745574100329641e-05, "loss": 4.2593, "step": 768000 }, { "epoch": 4.630913761290867, "eval_loss": 5.119594573974609, "eval_runtime": 38.842, "eval_samples_per_second": 13.182, "eval_steps_per_second": 6.591, "eval_tts_loss": 7.713245193119798, "step": 768000 }, { "epoch": 4.6312152530722015, "grad_norm": 2.493424892425537, "learning_rate": 7.743793076217707e-05, "loss": 4.5278, "step": 768050 }, { "epoch": 4.631516744853535, "grad_norm": 3.4880728721618652, "learning_rate": 7.742012185645909e-05, "loss": 4.7461, "step": 768100 }, { "epoch": 4.631818236634869, "grad_norm": 2.4106192588806152, "learning_rate": 7.740231428647022e-05, "loss": 4.7351, "step": 768150 }, { "epoch": 4.632119728416203, "grad_norm": 6.639963150024414, "learning_rate": 7.73845080525381e-05, "loss": 4.7879, "step": 768200 }, { "epoch": 4.632421220197537, "grad_norm": 2.420003890991211, "learning_rate": 7.736670315499047e-05, "loss": 4.3807, "step": 768250 }, { "epoch": 4.632722711978872, "grad_norm": 2.6471433639526367, "learning_rate": 7.734889959415509e-05, "loss": 4.6958, "step": 768300 }, { "epoch": 4.633024203760206, "grad_norm": 1.7492666244506836, "learning_rate": 7.733109737035953e-05, "loss": 4.6204, "step": 768350 }, { "epoch": 4.633325695541539, "grad_norm": 2.138167381286621, "learning_rate": 7.731329648393137e-05, "loss": 4.7221, "step": 768400 }, { "epoch": 4.633627187322873, "grad_norm": 4.066776752471924, "learning_rate": 7.72954969351983e-05, "loss": 4.9178, "step": 768450 }, { "epoch": 4.633928679104208, "grad_norm": 4.0533905029296875, "learning_rate": 7.727769872448783e-05, "loss": 4.3934, "step": 768500 }, { "epoch": 4.634230170885542, "grad_norm": 3.044264316558838, "learning_rate": 7.725990185212748e-05, "loss": 4.8362, "step": 768550 }, { "epoch": 4.634531662666876, "grad_norm": 2.304326057434082, "learning_rate": 7.724210631844494e-05, "loss": 4.8591, "step": 768600 }, { "epoch": 4.634833154448209, "grad_norm": 1.8414355516433716, "learning_rate": 7.722431212376751e-05, "loss": 4.5653, "step": 768650 }, { "epoch": 4.6351346462295435, "grad_norm": 3.409325122833252, "learning_rate": 7.720651926842284e-05, "loss": 4.8321, "step": 768700 }, { "epoch": 4.635436138010878, "grad_norm": 2.715566873550415, "learning_rate": 7.71887277527382e-05, "loss": 4.4419, "step": 768750 }, { "epoch": 4.635737629792212, "grad_norm": 2.251049041748047, "learning_rate": 7.717093757704123e-05, "loss": 4.5752, "step": 768800 }, { "epoch": 4.636039121573546, "grad_norm": 3.9598984718322754, "learning_rate": 7.715314874165914e-05, "loss": 4.8621, "step": 768850 }, { "epoch": 4.63634061335488, "grad_norm": 2.316221237182617, "learning_rate": 7.713536124691945e-05, "loss": 4.7617, "step": 768900 }, { "epoch": 4.636642105136214, "grad_norm": 2.7587344646453857, "learning_rate": 7.711757509314939e-05, "loss": 4.9133, "step": 768950 }, { "epoch": 4.636943596917548, "grad_norm": 2.6637496948242188, "learning_rate": 7.709979028067635e-05, "loss": 4.8216, "step": 769000 }, { "epoch": 4.637245088698882, "grad_norm": 3.8166327476501465, "learning_rate": 7.708200680982772e-05, "loss": 5.1305, "step": 769050 }, { "epoch": 4.637546580480216, "grad_norm": 2.7339978218078613, "learning_rate": 7.70642246809306e-05, "loss": 4.6612, "step": 769100 }, { "epoch": 4.6378480722615505, "grad_norm": 3.403041124343872, "learning_rate": 7.704644389431241e-05, "loss": 4.5959, "step": 769150 }, { "epoch": 4.638149564042884, "grad_norm": 4.395195484161377, "learning_rate": 7.702866445030032e-05, "loss": 4.9558, "step": 769200 }, { "epoch": 4.638451055824218, "grad_norm": 2.790245532989502, "learning_rate": 7.701088634922146e-05, "loss": 4.5219, "step": 769250 }, { "epoch": 4.638752547605552, "grad_norm": 2.527405023574829, "learning_rate": 7.699310959140308e-05, "loss": 4.8277, "step": 769300 }, { "epoch": 4.639054039386886, "grad_norm": 0.9901661276817322, "learning_rate": 7.697533417717237e-05, "loss": 4.546, "step": 769350 }, { "epoch": 4.639355531168221, "grad_norm": 3.2425925731658936, "learning_rate": 7.695756010685639e-05, "loss": 4.8275, "step": 769400 }, { "epoch": 4.639657022949555, "grad_norm": 3.744664192199707, "learning_rate": 7.693978738078228e-05, "loss": 4.5789, "step": 769450 }, { "epoch": 4.639958514730888, "grad_norm": 2.6834352016448975, "learning_rate": 7.692201599927716e-05, "loss": 4.4081, "step": 769500 }, { "epoch": 4.640260006512222, "grad_norm": 3.7438180446624756, "learning_rate": 7.690424596266806e-05, "loss": 4.9853, "step": 769550 }, { "epoch": 4.640561498293557, "grad_norm": 3.6968822479248047, "learning_rate": 7.688647727128194e-05, "loss": 5.1335, "step": 769600 }, { "epoch": 4.640862990074891, "grad_norm": 2.2257606983184814, "learning_rate": 7.686870992544591e-05, "loss": 4.6949, "step": 769650 }, { "epoch": 4.641164481856224, "grad_norm": 2.63655686378479, "learning_rate": 7.685094392548688e-05, "loss": 4.7546, "step": 769700 }, { "epoch": 4.641465973637558, "grad_norm": 3.852216958999634, "learning_rate": 7.683317927173182e-05, "loss": 4.3602, "step": 769750 }, { "epoch": 4.6417674654188925, "grad_norm": 3.082490921020508, "learning_rate": 7.68154159645077e-05, "loss": 4.6931, "step": 769800 }, { "epoch": 4.642068957200227, "grad_norm": 2.0822157859802246, "learning_rate": 7.679765400414143e-05, "loss": 5.0422, "step": 769850 }, { "epoch": 4.642370448981561, "grad_norm": 5.266607761383057, "learning_rate": 7.677989339095983e-05, "loss": 4.878, "step": 769900 }, { "epoch": 4.642671940762895, "grad_norm": 2.384723663330078, "learning_rate": 7.676213412528989e-05, "loss": 4.8122, "step": 769950 }, { "epoch": 4.642973432544229, "grad_norm": 3.0436646938323975, "learning_rate": 7.674437620745832e-05, "loss": 4.5709, "step": 770000 }, { "epoch": 4.643274924325563, "grad_norm": 3.051121711730957, "learning_rate": 7.672661963779192e-05, "loss": 4.2646, "step": 770050 }, { "epoch": 4.643576416106897, "grad_norm": 3.326120376586914, "learning_rate": 7.670886441661751e-05, "loss": 4.5671, "step": 770100 }, { "epoch": 4.643877907888231, "grad_norm": 3.1456198692321777, "learning_rate": 7.669111054426192e-05, "loss": 4.5181, "step": 770150 }, { "epoch": 4.644179399669565, "grad_norm": 3.462125539779663, "learning_rate": 7.667335802105177e-05, "loss": 4.5744, "step": 770200 }, { "epoch": 4.644480891450899, "grad_norm": 2.662245512008667, "learning_rate": 7.66556068473138e-05, "loss": 4.8709, "step": 770250 }, { "epoch": 4.644782383232233, "grad_norm": 8.15291976928711, "learning_rate": 7.663785702337483e-05, "loss": 4.1912, "step": 770300 }, { "epoch": 4.645083875013567, "grad_norm": 2.294635772705078, "learning_rate": 7.662010854956136e-05, "loss": 4.1164, "step": 770350 }, { "epoch": 4.645385366794901, "grad_norm": 2.274756669998169, "learning_rate": 7.660236142620003e-05, "loss": 4.5186, "step": 770400 }, { "epoch": 4.645686858576235, "grad_norm": 2.3472182750701904, "learning_rate": 7.658461565361755e-05, "loss": 5.1264, "step": 770450 }, { "epoch": 4.64598835035757, "grad_norm": 2.746598958969116, "learning_rate": 7.65668712321404e-05, "loss": 4.6145, "step": 770500 }, { "epoch": 4.646289842138903, "grad_norm": 3.1598567962646484, "learning_rate": 7.654912816209518e-05, "loss": 4.5794, "step": 770550 }, { "epoch": 4.646591333920237, "grad_norm": 5.941763877868652, "learning_rate": 7.653138644380849e-05, "loss": 4.8149, "step": 770600 }, { "epoch": 4.646892825701571, "grad_norm": 2.3971569538116455, "learning_rate": 7.651364607760674e-05, "loss": 4.8041, "step": 770650 }, { "epoch": 4.647194317482906, "grad_norm": 2.9743924140930176, "learning_rate": 7.649590706381652e-05, "loss": 4.8074, "step": 770700 }, { "epoch": 4.64749580926424, "grad_norm": 1.4887689352035522, "learning_rate": 7.647816940276423e-05, "loss": 4.8298, "step": 770750 }, { "epoch": 4.647797301045573, "grad_norm": 2.9752941131591797, "learning_rate": 7.646043309477625e-05, "loss": 4.7554, "step": 770800 }, { "epoch": 4.648098792826907, "grad_norm": 3.1485393047332764, "learning_rate": 7.644269814017906e-05, "loss": 4.7865, "step": 770850 }, { "epoch": 4.6484002846082415, "grad_norm": 2.323715925216675, "learning_rate": 7.64249645392991e-05, "loss": 4.6091, "step": 770900 }, { "epoch": 4.648701776389576, "grad_norm": 2.511767864227295, "learning_rate": 7.640723229246261e-05, "loss": 4.6933, "step": 770950 }, { "epoch": 4.64900326817091, "grad_norm": 4.027288436889648, "learning_rate": 7.638950139999599e-05, "loss": 5.1478, "step": 771000 }, { "epoch": 4.64900326817091, "eval_loss": 5.109338760375977, "eval_runtime": 39.0312, "eval_samples_per_second": 13.118, "eval_steps_per_second": 6.559, "eval_tts_loss": 7.709385835276152, "step": 771000 }, { "epoch": 4.649304759952244, "grad_norm": 3.4690475463867188, "learning_rate": 7.637177186222559e-05, "loss": 4.548, "step": 771050 }, { "epoch": 4.6496062517335774, "grad_norm": 5.0438761711120605, "learning_rate": 7.63540436794776e-05, "loss": 4.6359, "step": 771100 }, { "epoch": 4.649907743514912, "grad_norm": 2.6690680980682373, "learning_rate": 7.633631685207841e-05, "loss": 4.1209, "step": 771150 }, { "epoch": 4.650209235296246, "grad_norm": 2.32448410987854, "learning_rate": 7.631859138035419e-05, "loss": 4.4446, "step": 771200 }, { "epoch": 4.65051072707758, "grad_norm": 2.905317783355713, "learning_rate": 7.630086726463107e-05, "loss": 4.7743, "step": 771250 }, { "epoch": 4.650812218858914, "grad_norm": 3.983654499053955, "learning_rate": 7.628314450523532e-05, "loss": 4.812, "step": 771300 }, { "epoch": 4.651113710640248, "grad_norm": 2.3906309604644775, "learning_rate": 7.626542310249316e-05, "loss": 4.3944, "step": 771350 }, { "epoch": 4.651415202421582, "grad_norm": 2.442258834838867, "learning_rate": 7.624770305673061e-05, "loss": 4.7829, "step": 771400 }, { "epoch": 4.651716694202916, "grad_norm": 2.7098684310913086, "learning_rate": 7.622998436827381e-05, "loss": 3.616, "step": 771450 }, { "epoch": 4.65201818598425, "grad_norm": 0.6629458665847778, "learning_rate": 7.621226703744894e-05, "loss": 4.889, "step": 771500 }, { "epoch": 4.652319677765584, "grad_norm": 2.4627294540405273, "learning_rate": 7.619455106458201e-05, "loss": 4.9524, "step": 771550 }, { "epoch": 4.652621169546919, "grad_norm": 2.7659361362457275, "learning_rate": 7.617683644999897e-05, "loss": 4.8252, "step": 771600 }, { "epoch": 4.652922661328252, "grad_norm": 2.980112075805664, "learning_rate": 7.615912319402597e-05, "loss": 4.5277, "step": 771650 }, { "epoch": 4.653224153109586, "grad_norm": 4.160601615905762, "learning_rate": 7.614141129698886e-05, "loss": 4.5179, "step": 771700 }, { "epoch": 4.65352564489092, "grad_norm": 3.278372287750244, "learning_rate": 7.612370075921368e-05, "loss": 4.7912, "step": 771750 }, { "epoch": 4.653827136672255, "grad_norm": 1.6632400751113892, "learning_rate": 7.610599158102643e-05, "loss": 4.4504, "step": 771800 }, { "epoch": 4.654128628453588, "grad_norm": 2.664520502090454, "learning_rate": 7.608828376275287e-05, "loss": 4.7988, "step": 771850 }, { "epoch": 4.654430120234922, "grad_norm": 4.526621341705322, "learning_rate": 7.607057730471904e-05, "loss": 4.5454, "step": 771900 }, { "epoch": 4.654731612016256, "grad_norm": 2.7465622425079346, "learning_rate": 7.605287220725074e-05, "loss": 4.5078, "step": 771950 }, { "epoch": 4.6550331037975905, "grad_norm": 3.29219126701355, "learning_rate": 7.603516847067372e-05, "loss": 4.048, "step": 772000 }, { "epoch": 4.655334595578925, "grad_norm": 2.6525015830993652, "learning_rate": 7.60174660953139e-05, "loss": 4.7115, "step": 772050 }, { "epoch": 4.655636087360259, "grad_norm": 3.1806178092956543, "learning_rate": 7.599976508149709e-05, "loss": 4.5472, "step": 772100 }, { "epoch": 4.655937579141592, "grad_norm": 3.390516519546509, "learning_rate": 7.598206542954892e-05, "loss": 4.6708, "step": 772150 }, { "epoch": 4.656239070922926, "grad_norm": 2.4813435077667236, "learning_rate": 7.596436713979522e-05, "loss": 4.5585, "step": 772200 }, { "epoch": 4.656540562704261, "grad_norm": 3.651670217514038, "learning_rate": 7.594667021256174e-05, "loss": 4.6142, "step": 772250 }, { "epoch": 4.656842054485595, "grad_norm": 3.1916286945343018, "learning_rate": 7.592897464817408e-05, "loss": 4.7457, "step": 772300 }, { "epoch": 4.657143546266929, "grad_norm": 2.260434865951538, "learning_rate": 7.591128044695796e-05, "loss": 4.8654, "step": 772350 }, { "epoch": 4.657445038048262, "grad_norm": 2.4556691646575928, "learning_rate": 7.589358760923902e-05, "loss": 4.8232, "step": 772400 }, { "epoch": 4.657746529829597, "grad_norm": 3.0818347930908203, "learning_rate": 7.587589613534276e-05, "loss": 4.6092, "step": 772450 }, { "epoch": 4.658048021610931, "grad_norm": 2.9356422424316406, "learning_rate": 7.585820602559486e-05, "loss": 4.6772, "step": 772500 }, { "epoch": 4.658349513392265, "grad_norm": 3.042681932449341, "learning_rate": 7.584051728032096e-05, "loss": 4.4626, "step": 772550 }, { "epoch": 4.658651005173599, "grad_norm": 2.195882797241211, "learning_rate": 7.582282989984643e-05, "loss": 4.6593, "step": 772600 }, { "epoch": 4.658952496954933, "grad_norm": 2.5814297199249268, "learning_rate": 7.580514388449686e-05, "loss": 4.6941, "step": 772650 }, { "epoch": 4.659253988736267, "grad_norm": 2.796588182449341, "learning_rate": 7.578745923459779e-05, "loss": 4.8847, "step": 772700 }, { "epoch": 4.659555480517601, "grad_norm": 3.2807412147521973, "learning_rate": 7.576977595047461e-05, "loss": 4.4541, "step": 772750 }, { "epoch": 4.659856972298935, "grad_norm": 4.223318576812744, "learning_rate": 7.575209403245275e-05, "loss": 4.2676, "step": 772800 }, { "epoch": 4.660158464080269, "grad_norm": 1.1274832487106323, "learning_rate": 7.573441348085766e-05, "loss": 4.1922, "step": 772850 }, { "epoch": 4.6604599558616036, "grad_norm": 3.2536306381225586, "learning_rate": 7.571673429601468e-05, "loss": 4.2786, "step": 772900 }, { "epoch": 4.660761447642937, "grad_norm": 3.5972659587860107, "learning_rate": 7.569905647824918e-05, "loss": 4.9168, "step": 772950 }, { "epoch": 4.661062939424271, "grad_norm": 2.7215700149536133, "learning_rate": 7.568138002788658e-05, "loss": 4.574, "step": 773000 }, { "epoch": 4.661364431205605, "grad_norm": 3.3195102214813232, "learning_rate": 7.566370494525204e-05, "loss": 4.5851, "step": 773050 }, { "epoch": 4.6616659229869395, "grad_norm": 2.3254806995391846, "learning_rate": 7.564603123067102e-05, "loss": 4.5627, "step": 773100 }, { "epoch": 4.661967414768274, "grad_norm": 2.764617919921875, "learning_rate": 7.562835888446863e-05, "loss": 4.7218, "step": 773150 }, { "epoch": 4.662268906549608, "grad_norm": 2.53108549118042, "learning_rate": 7.561068790697019e-05, "loss": 4.7412, "step": 773200 }, { "epoch": 4.662570398330941, "grad_norm": 1.447890281677246, "learning_rate": 7.559301829850082e-05, "loss": 4.741, "step": 773250 }, { "epoch": 4.662871890112275, "grad_norm": 1.7822649478912354, "learning_rate": 7.55753500593858e-05, "loss": 4.5781, "step": 773300 }, { "epoch": 4.66317338189361, "grad_norm": 3.228071689605713, "learning_rate": 7.555768318995027e-05, "loss": 4.5576, "step": 773350 }, { "epoch": 4.663474873674944, "grad_norm": 1.1607457399368286, "learning_rate": 7.554001769051932e-05, "loss": 4.6791, "step": 773400 }, { "epoch": 4.663776365456278, "grad_norm": 4.8153977394104, "learning_rate": 7.552235356141806e-05, "loss": 4.7529, "step": 773450 }, { "epoch": 4.664077857237611, "grad_norm": 3.3063583374023438, "learning_rate": 7.550469080297167e-05, "loss": 4.6242, "step": 773500 }, { "epoch": 4.664379349018946, "grad_norm": 4.545989990234375, "learning_rate": 7.548702941550514e-05, "loss": 4.6458, "step": 773550 }, { "epoch": 4.66468084080028, "grad_norm": 1.1858524084091187, "learning_rate": 7.546936939934342e-05, "loss": 4.7043, "step": 773600 }, { "epoch": 4.664982332581614, "grad_norm": 3.9314849376678467, "learning_rate": 7.545171075481166e-05, "loss": 5.0764, "step": 773650 }, { "epoch": 4.665283824362948, "grad_norm": 3.899691581726074, "learning_rate": 7.543405348223472e-05, "loss": 4.7526, "step": 773700 }, { "epoch": 4.665585316144282, "grad_norm": 2.55222225189209, "learning_rate": 7.541639758193763e-05, "loss": 4.434, "step": 773750 }, { "epoch": 4.665886807925616, "grad_norm": 2.625379800796509, "learning_rate": 7.539874305424536e-05, "loss": 4.4634, "step": 773800 }, { "epoch": 4.66618829970695, "grad_norm": 3.2466540336608887, "learning_rate": 7.538108989948268e-05, "loss": 4.9725, "step": 773850 }, { "epoch": 4.666489791488284, "grad_norm": 1.3566282987594604, "learning_rate": 7.536343811797464e-05, "loss": 4.6146, "step": 773900 }, { "epoch": 4.666791283269618, "grad_norm": 4.080916404724121, "learning_rate": 7.5345787710046e-05, "loss": 4.4857, "step": 773950 }, { "epoch": 4.667092775050952, "grad_norm": 2.3501527309417725, "learning_rate": 7.532813867602155e-05, "loss": 4.6499, "step": 774000 }, { "epoch": 4.667092775050952, "eval_loss": 5.105499744415283, "eval_runtime": 39.0587, "eval_samples_per_second": 13.108, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.655303233315284, "step": 774000 }, { "epoch": 4.667394266832286, "grad_norm": 2.618210792541504, "learning_rate": 7.531049101622613e-05, "loss": 4.836, "step": 774050 }, { "epoch": 4.66769575861362, "grad_norm": 2.255856513977051, "learning_rate": 7.529284473098458e-05, "loss": 4.8165, "step": 774100 }, { "epoch": 4.667997250394954, "grad_norm": 1.0962400436401367, "learning_rate": 7.527519982062158e-05, "loss": 4.2848, "step": 774150 }, { "epoch": 4.6682987421762885, "grad_norm": 2.3316409587860107, "learning_rate": 7.525755628546187e-05, "loss": 4.8961, "step": 774200 }, { "epoch": 4.668600233957623, "grad_norm": 1.8572700023651123, "learning_rate": 7.523991412583024e-05, "loss": 4.5749, "step": 774250 }, { "epoch": 4.668901725738956, "grad_norm": 2.3780124187469482, "learning_rate": 7.522227334205131e-05, "loss": 4.389, "step": 774300 }, { "epoch": 4.66920321752029, "grad_norm": 2.1756696701049805, "learning_rate": 7.520463393444966e-05, "loss": 5.1052, "step": 774350 }, { "epoch": 4.669504709301624, "grad_norm": 1.9956191778182983, "learning_rate": 7.518699590335003e-05, "loss": 4.7624, "step": 774400 }, { "epoch": 4.669806201082959, "grad_norm": 3.421706438064575, "learning_rate": 7.516935924907694e-05, "loss": 4.6766, "step": 774450 }, { "epoch": 4.670107692864293, "grad_norm": 2.36698842048645, "learning_rate": 7.5151723971955e-05, "loss": 4.1374, "step": 774500 }, { "epoch": 4.670409184645626, "grad_norm": 4.814591407775879, "learning_rate": 7.513409007230881e-05, "loss": 4.6734, "step": 774550 }, { "epoch": 4.67071067642696, "grad_norm": 2.3432343006134033, "learning_rate": 7.51164575504628e-05, "loss": 5.0716, "step": 774600 }, { "epoch": 4.671012168208295, "grad_norm": 1.8098952770233154, "learning_rate": 7.509882640674155e-05, "loss": 4.6374, "step": 774650 }, { "epoch": 4.671313659989629, "grad_norm": 3.762754440307617, "learning_rate": 7.508119664146953e-05, "loss": 4.577, "step": 774700 }, { "epoch": 4.671615151770963, "grad_norm": 1.4904205799102783, "learning_rate": 7.506356825497119e-05, "loss": 4.619, "step": 774750 }, { "epoch": 4.671916643552297, "grad_norm": 1.6600010395050049, "learning_rate": 7.504594124757088e-05, "loss": 4.5985, "step": 774800 }, { "epoch": 4.6722181353336305, "grad_norm": 2.378190755844116, "learning_rate": 7.502831561959311e-05, "loss": 5.0904, "step": 774850 }, { "epoch": 4.672519627114965, "grad_norm": 2.3482160568237305, "learning_rate": 7.501069137136211e-05, "loss": 4.7848, "step": 774900 }, { "epoch": 4.672821118896299, "grad_norm": 2.680644989013672, "learning_rate": 7.499306850320237e-05, "loss": 4.7216, "step": 774950 }, { "epoch": 4.673122610677633, "grad_norm": 3.5066301822662354, "learning_rate": 7.497544701543821e-05, "loss": 4.5444, "step": 775000 }, { "epoch": 4.673424102458967, "grad_norm": 2.8990635871887207, "learning_rate": 7.49578269083938e-05, "loss": 4.1839, "step": 775050 }, { "epoch": 4.673725594240301, "grad_norm": 9.704811096191406, "learning_rate": 7.494020818239358e-05, "loss": 4.7896, "step": 775100 }, { "epoch": 4.674027086021635, "grad_norm": 2.961141347885132, "learning_rate": 7.49225908377617e-05, "loss": 4.4224, "step": 775150 }, { "epoch": 4.674328577802969, "grad_norm": 3.0461466312408447, "learning_rate": 7.490497487482235e-05, "loss": 4.7991, "step": 775200 }, { "epoch": 4.674630069584303, "grad_norm": 3.1687467098236084, "learning_rate": 7.488736029389975e-05, "loss": 4.6479, "step": 775250 }, { "epoch": 4.6749315613656375, "grad_norm": 2.9837565422058105, "learning_rate": 7.486974709531816e-05, "loss": 4.5535, "step": 775300 }, { "epoch": 4.675233053146972, "grad_norm": 2.890045166015625, "learning_rate": 7.485213527940161e-05, "loss": 4.9444, "step": 775350 }, { "epoch": 4.675534544928305, "grad_norm": 1.8415340185165405, "learning_rate": 7.483452484647425e-05, "loss": 4.9425, "step": 775400 }, { "epoch": 4.675836036709639, "grad_norm": 2.2695984840393066, "learning_rate": 7.481691579686027e-05, "loss": 4.585, "step": 775450 }, { "epoch": 4.676137528490973, "grad_norm": 4.448013782501221, "learning_rate": 7.479930813088359e-05, "loss": 4.6853, "step": 775500 }, { "epoch": 4.676439020272308, "grad_norm": 3.565423011779785, "learning_rate": 7.478170184886838e-05, "loss": 4.6481, "step": 775550 }, { "epoch": 4.676740512053641, "grad_norm": 3.6516551971435547, "learning_rate": 7.476409695113861e-05, "loss": 5.09, "step": 775600 }, { "epoch": 4.677042003834975, "grad_norm": 2.956944227218628, "learning_rate": 7.474649343801822e-05, "loss": 4.5834, "step": 775650 }, { "epoch": 4.677343495616309, "grad_norm": 2.8414292335510254, "learning_rate": 7.472889130983122e-05, "loss": 4.4912, "step": 775700 }, { "epoch": 4.677644987397644, "grad_norm": 3.4795641899108887, "learning_rate": 7.471129056690162e-05, "loss": 4.6918, "step": 775750 }, { "epoch": 4.677946479178978, "grad_norm": 2.4600558280944824, "learning_rate": 7.46936912095532e-05, "loss": 4.815, "step": 775800 }, { "epoch": 4.678247970960312, "grad_norm": 4.834419250488281, "learning_rate": 7.467609323810996e-05, "loss": 4.7518, "step": 775850 }, { "epoch": 4.678549462741645, "grad_norm": 2.5278854370117188, "learning_rate": 7.465849665289576e-05, "loss": 4.8336, "step": 775900 }, { "epoch": 4.6788509545229795, "grad_norm": 4.285449981689453, "learning_rate": 7.464090145423441e-05, "loss": 4.875, "step": 775950 }, { "epoch": 4.679152446304314, "grad_norm": 3.205531358718872, "learning_rate": 7.462330764244968e-05, "loss": 4.7067, "step": 776000 }, { "epoch": 4.679453938085648, "grad_norm": 4.263006210327148, "learning_rate": 7.460571521786545e-05, "loss": 4.8935, "step": 776050 }, { "epoch": 4.679755429866982, "grad_norm": 2.4641008377075195, "learning_rate": 7.45881241808054e-05, "loss": 4.4654, "step": 776100 }, { "epoch": 4.680056921648315, "grad_norm": 2.7122609615325928, "learning_rate": 7.457053453159328e-05, "loss": 4.9733, "step": 776150 }, { "epoch": 4.68035841342965, "grad_norm": 2.6742324829101562, "learning_rate": 7.455294627055291e-05, "loss": 5.0798, "step": 776200 }, { "epoch": 4.680659905210984, "grad_norm": 2.935106039047241, "learning_rate": 7.453535939800783e-05, "loss": 4.3407, "step": 776250 }, { "epoch": 4.680961396992318, "grad_norm": 1.6625981330871582, "learning_rate": 7.451777391428183e-05, "loss": 4.5406, "step": 776300 }, { "epoch": 4.681262888773652, "grad_norm": 3.82066011428833, "learning_rate": 7.450018981969846e-05, "loss": 4.8139, "step": 776350 }, { "epoch": 4.6815643805549865, "grad_norm": 3.8298542499542236, "learning_rate": 7.44826071145813e-05, "loss": 4.5921, "step": 776400 }, { "epoch": 4.68186587233632, "grad_norm": 3.2877449989318848, "learning_rate": 7.446502579925401e-05, "loss": 4.5568, "step": 776450 }, { "epoch": 4.682167364117654, "grad_norm": 2.1921112537384033, "learning_rate": 7.444744587404011e-05, "loss": 4.8638, "step": 776500 }, { "epoch": 4.682468855898988, "grad_norm": 3.6797380447387695, "learning_rate": 7.442986733926321e-05, "loss": 4.6587, "step": 776550 }, { "epoch": 4.682770347680322, "grad_norm": 1.402713418006897, "learning_rate": 7.441229019524669e-05, "loss": 4.6253, "step": 776600 }, { "epoch": 4.683071839461657, "grad_norm": 3.0101728439331055, "learning_rate": 7.439471444231411e-05, "loss": 5.2305, "step": 776650 }, { "epoch": 4.68337333124299, "grad_norm": 3.011589527130127, "learning_rate": 7.437714008078898e-05, "loss": 4.7671, "step": 776700 }, { "epoch": 4.683674823024324, "grad_norm": 3.1550164222717285, "learning_rate": 7.435956711099465e-05, "loss": 4.8483, "step": 776750 }, { "epoch": 4.683976314805658, "grad_norm": 4.133758068084717, "learning_rate": 7.434199553325449e-05, "loss": 4.2677, "step": 776800 }, { "epoch": 4.6842778065869926, "grad_norm": 3.4136431217193604, "learning_rate": 7.4324425347892e-05, "loss": 4.6347, "step": 776850 }, { "epoch": 4.684579298368327, "grad_norm": 2.647073268890381, "learning_rate": 7.43068565552304e-05, "loss": 4.5472, "step": 776900 }, { "epoch": 4.684880790149661, "grad_norm": 3.9018449783325195, "learning_rate": 7.428928915559308e-05, "loss": 5.0811, "step": 776950 }, { "epoch": 4.685182281930994, "grad_norm": 2.6334245204925537, "learning_rate": 7.427172314930341e-05, "loss": 4.7385, "step": 777000 }, { "epoch": 4.685182281930994, "eval_loss": 5.111871719360352, "eval_runtime": 39.1392, "eval_samples_per_second": 13.082, "eval_steps_per_second": 6.541, "eval_tts_loss": 7.72349060692557, "step": 777000 }, { "epoch": 4.6854837737123285, "grad_norm": 4.085453033447266, "learning_rate": 7.425415853668456e-05, "loss": 5.027, "step": 777050 }, { "epoch": 4.685785265493663, "grad_norm": 2.8929097652435303, "learning_rate": 7.423659531805989e-05, "loss": 4.5343, "step": 777100 }, { "epoch": 4.686086757274997, "grad_norm": 2.507927417755127, "learning_rate": 7.421903349375253e-05, "loss": 4.8106, "step": 777150 }, { "epoch": 4.686388249056331, "grad_norm": 3.4383182525634766, "learning_rate": 7.420147306408567e-05, "loss": 4.867, "step": 777200 }, { "epoch": 4.686689740837664, "grad_norm": 2.391867160797119, "learning_rate": 7.418391402938254e-05, "loss": 5.0106, "step": 777250 }, { "epoch": 4.686991232618999, "grad_norm": 2.7247889041900635, "learning_rate": 7.416635638996633e-05, "loss": 4.5445, "step": 777300 }, { "epoch": 4.687292724400333, "grad_norm": 4.378165245056152, "learning_rate": 7.414880014616006e-05, "loss": 4.5888, "step": 777350 }, { "epoch": 4.687594216181667, "grad_norm": 2.8115828037261963, "learning_rate": 7.413124529828686e-05, "loss": 4.2847, "step": 777400 }, { "epoch": 4.687895707963001, "grad_norm": 4.583302974700928, "learning_rate": 7.411369184666988e-05, "loss": 4.6269, "step": 777450 }, { "epoch": 4.6881971997443355, "grad_norm": 1.4975736141204834, "learning_rate": 7.409613979163212e-05, "loss": 4.9729, "step": 777500 }, { "epoch": 4.688498691525669, "grad_norm": 4.253728866577148, "learning_rate": 7.407858913349652e-05, "loss": 4.5494, "step": 777550 }, { "epoch": 4.688800183307003, "grad_norm": 3.1747078895568848, "learning_rate": 7.406103987258618e-05, "loss": 5.0056, "step": 777600 }, { "epoch": 4.689101675088337, "grad_norm": 2.820601224899292, "learning_rate": 7.4043492009224e-05, "loss": 4.5996, "step": 777650 }, { "epoch": 4.689403166869671, "grad_norm": 2.5319645404815674, "learning_rate": 7.402594554373295e-05, "loss": 4.8421, "step": 777700 }, { "epoch": 4.689704658651005, "grad_norm": 0.8146824240684509, "learning_rate": 7.4008400476436e-05, "loss": 4.8479, "step": 777750 }, { "epoch": 4.690006150432339, "grad_norm": 2.2687323093414307, "learning_rate": 7.399085680765594e-05, "loss": 4.5565, "step": 777800 }, { "epoch": 4.690307642213673, "grad_norm": 2.301785945892334, "learning_rate": 7.39733145377157e-05, "loss": 4.488, "step": 777850 }, { "epoch": 4.690609133995007, "grad_norm": 1.3618162870407104, "learning_rate": 7.395577366693815e-05, "loss": 4.5508, "step": 777900 }, { "epoch": 4.6909106257763415, "grad_norm": 2.55495285987854, "learning_rate": 7.393823419564607e-05, "loss": 4.5889, "step": 777950 }, { "epoch": 4.691212117557676, "grad_norm": 2.4125444889068604, "learning_rate": 7.392069612416219e-05, "loss": 4.3212, "step": 778000 }, { "epoch": 4.691513609339009, "grad_norm": 5.413912773132324, "learning_rate": 7.39031594528094e-05, "loss": 4.252, "step": 778050 }, { "epoch": 4.691815101120343, "grad_norm": 3.3010292053222656, "learning_rate": 7.388562418191027e-05, "loss": 4.8009, "step": 778100 }, { "epoch": 4.6921165929016775, "grad_norm": 3.4455740451812744, "learning_rate": 7.386809031178762e-05, "loss": 4.7607, "step": 778150 }, { "epoch": 4.692418084683012, "grad_norm": 2.577054500579834, "learning_rate": 7.385055784276419e-05, "loss": 4.7558, "step": 778200 }, { "epoch": 4.692719576464346, "grad_norm": 3.317559242248535, "learning_rate": 7.383302677516251e-05, "loss": 4.6366, "step": 778250 }, { "epoch": 4.693021068245679, "grad_norm": 2.9383766651153564, "learning_rate": 7.381549710930532e-05, "loss": 4.9653, "step": 778300 }, { "epoch": 4.693322560027013, "grad_norm": 3.3677468299865723, "learning_rate": 7.37979688455152e-05, "loss": 4.6848, "step": 778350 }, { "epoch": 4.693624051808348, "grad_norm": 3.36736798286438, "learning_rate": 7.378044198411467e-05, "loss": 5.1772, "step": 778400 }, { "epoch": 4.693925543589682, "grad_norm": 2.4883666038513184, "learning_rate": 7.376291652542632e-05, "loss": 4.5112, "step": 778450 }, { "epoch": 4.694227035371016, "grad_norm": 2.4651072025299072, "learning_rate": 7.374539246977275e-05, "loss": 4.5482, "step": 778500 }, { "epoch": 4.69452852715235, "grad_norm": 2.3848581314086914, "learning_rate": 7.372786981747635e-05, "loss": 4.5569, "step": 778550 }, { "epoch": 4.694830018933684, "grad_norm": 2.044710397720337, "learning_rate": 7.371034856885967e-05, "loss": 4.5458, "step": 778600 }, { "epoch": 4.695131510715018, "grad_norm": 4.695211887359619, "learning_rate": 7.369282872424522e-05, "loss": 4.7119, "step": 778650 }, { "epoch": 4.695433002496352, "grad_norm": 4.009008407592773, "learning_rate": 7.367531028395534e-05, "loss": 4.4956, "step": 778700 }, { "epoch": 4.695734494277686, "grad_norm": 4.368645668029785, "learning_rate": 7.36577932483124e-05, "loss": 4.8862, "step": 778750 }, { "epoch": 4.69603598605902, "grad_norm": 2.4637742042541504, "learning_rate": 7.364027761763889e-05, "loss": 4.7724, "step": 778800 }, { "epoch": 4.696337477840354, "grad_norm": 4.030968189239502, "learning_rate": 7.362276339225705e-05, "loss": 4.9174, "step": 778850 }, { "epoch": 4.696638969621688, "grad_norm": 4.310900688171387, "learning_rate": 7.360525057248924e-05, "loss": 4.6672, "step": 778900 }, { "epoch": 4.696940461403022, "grad_norm": 3.2299628257751465, "learning_rate": 7.358773915865785e-05, "loss": 4.6123, "step": 778950 }, { "epoch": 4.697241953184356, "grad_norm": 3.705335855484009, "learning_rate": 7.357022915108499e-05, "loss": 4.8344, "step": 779000 }, { "epoch": 4.6975434449656905, "grad_norm": 3.073983669281006, "learning_rate": 7.355272055009301e-05, "loss": 5.0392, "step": 779050 }, { "epoch": 4.697844936747025, "grad_norm": 2.1614396572113037, "learning_rate": 7.353521335600415e-05, "loss": 4.2097, "step": 779100 }, { "epoch": 4.698146428528358, "grad_norm": 2.6019062995910645, "learning_rate": 7.351770756914059e-05, "loss": 4.6623, "step": 779150 }, { "epoch": 4.698447920309692, "grad_norm": 4.7451581954956055, "learning_rate": 7.35002031898244e-05, "loss": 5.0642, "step": 779200 }, { "epoch": 4.6987494120910265, "grad_norm": 0.6207139492034912, "learning_rate": 7.348270021837787e-05, "loss": 4.6479, "step": 779250 }, { "epoch": 4.699050903872361, "grad_norm": 3.247370719909668, "learning_rate": 7.346519865512298e-05, "loss": 4.7917, "step": 779300 }, { "epoch": 4.699352395653694, "grad_norm": 2.9730069637298584, "learning_rate": 7.34476985003819e-05, "loss": 4.6942, "step": 779350 }, { "epoch": 4.699653887435028, "grad_norm": 4.418334484100342, "learning_rate": 7.343019975447672e-05, "loss": 4.8294, "step": 779400 }, { "epoch": 4.699955379216362, "grad_norm": 3.7414236068725586, "learning_rate": 7.341270241772942e-05, "loss": 5.2193, "step": 779450 }, { "epoch": 4.700256870997697, "grad_norm": 2.082345485687256, "learning_rate": 7.339520649046207e-05, "loss": 4.8733, "step": 779500 }, { "epoch": 4.700558362779031, "grad_norm": 3.571931838989258, "learning_rate": 7.337771197299662e-05, "loss": 4.7243, "step": 779550 }, { "epoch": 4.700859854560365, "grad_norm": 3.1632091999053955, "learning_rate": 7.336021886565499e-05, "loss": 4.6986, "step": 779600 }, { "epoch": 4.701161346341698, "grad_norm": 2.484283447265625, "learning_rate": 7.334272716875914e-05, "loss": 4.5688, "step": 779650 }, { "epoch": 4.701462838123033, "grad_norm": 4.310775279998779, "learning_rate": 7.332523688263107e-05, "loss": 4.5526, "step": 779700 }, { "epoch": 4.701764329904367, "grad_norm": 2.4133453369140625, "learning_rate": 7.330774800759255e-05, "loss": 4.701, "step": 779750 }, { "epoch": 4.702065821685701, "grad_norm": 1.9805052280426025, "learning_rate": 7.329026054396547e-05, "loss": 4.8041, "step": 779800 }, { "epoch": 4.702367313467035, "grad_norm": 2.8886072635650635, "learning_rate": 7.327277449207174e-05, "loss": 4.5928, "step": 779850 }, { "epoch": 4.7026688052483685, "grad_norm": 2.7266104221343994, "learning_rate": 7.325528985223304e-05, "loss": 5.0414, "step": 779900 }, { "epoch": 4.702970297029703, "grad_norm": 3.8286263942718506, "learning_rate": 7.323780662477124e-05, "loss": 4.8358, "step": 779950 }, { "epoch": 4.703271788811037, "grad_norm": 4.089616775512695, "learning_rate": 7.322032481000805e-05, "loss": 4.7345, "step": 780000 }, { "epoch": 4.703271788811037, "eval_loss": 5.102911472320557, "eval_runtime": 39.0923, "eval_samples_per_second": 13.097, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.705530105766228, "step": 780000 }, { "epoch": 4.703573280592371, "grad_norm": 1.1794590950012207, "learning_rate": 7.320284440826523e-05, "loss": 4.5738, "step": 780050 }, { "epoch": 4.703874772373705, "grad_norm": 3.088536262512207, "learning_rate": 7.318536541986442e-05, "loss": 4.4405, "step": 780100 }, { "epoch": 4.7041762641550395, "grad_norm": 3.592283248901367, "learning_rate": 7.316788784512736e-05, "loss": 4.6351, "step": 780150 }, { "epoch": 4.704477755936373, "grad_norm": 1.4621143341064453, "learning_rate": 7.31504116843757e-05, "loss": 4.4331, "step": 780200 }, { "epoch": 4.704779247717707, "grad_norm": 2.0063748359680176, "learning_rate": 7.313293693793103e-05, "loss": 4.7608, "step": 780250 }, { "epoch": 4.705080739499041, "grad_norm": 2.4134788513183594, "learning_rate": 7.311546360611498e-05, "loss": 4.5805, "step": 780300 }, { "epoch": 4.7053822312803755, "grad_norm": 1.1707173585891724, "learning_rate": 7.30979916892491e-05, "loss": 4.2137, "step": 780350 }, { "epoch": 4.70568372306171, "grad_norm": 1.1127430200576782, "learning_rate": 7.308052118765491e-05, "loss": 4.4597, "step": 780400 }, { "epoch": 4.705985214843043, "grad_norm": 2.3672897815704346, "learning_rate": 7.306305210165394e-05, "loss": 4.8834, "step": 780450 }, { "epoch": 4.706286706624377, "grad_norm": 3.452868938446045, "learning_rate": 7.304558443156775e-05, "loss": 4.7552, "step": 780500 }, { "epoch": 4.706588198405711, "grad_norm": 3.1451661586761475, "learning_rate": 7.302811817771773e-05, "loss": 4.5811, "step": 780550 }, { "epoch": 4.706889690187046, "grad_norm": 1.2311886548995972, "learning_rate": 7.301065334042532e-05, "loss": 4.2565, "step": 780600 }, { "epoch": 4.70719118196838, "grad_norm": 2.8111350536346436, "learning_rate": 7.299318992001203e-05, "loss": 4.4757, "step": 780650 }, { "epoch": 4.707492673749714, "grad_norm": 2.741131067276001, "learning_rate": 7.297572791679918e-05, "loss": 4.6001, "step": 780700 }, { "epoch": 4.707794165531047, "grad_norm": 2.4137940406799316, "learning_rate": 7.295826733110811e-05, "loss": 4.6139, "step": 780750 }, { "epoch": 4.7080956573123816, "grad_norm": 2.555600643157959, "learning_rate": 7.294080816326022e-05, "loss": 4.6701, "step": 780800 }, { "epoch": 4.708397149093716, "grad_norm": 2.7815122604370117, "learning_rate": 7.292335041357673e-05, "loss": 4.7133, "step": 780850 }, { "epoch": 4.70869864087505, "grad_norm": 1.5534913539886475, "learning_rate": 7.2905894082379e-05, "loss": 4.0494, "step": 780900 }, { "epoch": 4.709000132656384, "grad_norm": 1.7936104536056519, "learning_rate": 7.288843916998828e-05, "loss": 4.5947, "step": 780950 }, { "epoch": 4.7093016244377175, "grad_norm": 3.0450055599212646, "learning_rate": 7.287098567672579e-05, "loss": 4.1818, "step": 781000 }, { "epoch": 4.709603116219052, "grad_norm": 3.163341760635376, "learning_rate": 7.285353360291277e-05, "loss": 4.4317, "step": 781050 }, { "epoch": 4.709904608000386, "grad_norm": 4.798808574676514, "learning_rate": 7.283608294887031e-05, "loss": 4.5342, "step": 781100 }, { "epoch": 4.71020609978172, "grad_norm": 3.0820605754852295, "learning_rate": 7.28186337149197e-05, "loss": 4.6989, "step": 781150 }, { "epoch": 4.710507591563054, "grad_norm": 2.141223907470703, "learning_rate": 7.280118590138192e-05, "loss": 5.1216, "step": 781200 }, { "epoch": 4.7108090833443885, "grad_norm": 2.780123472213745, "learning_rate": 7.27837395085782e-05, "loss": 4.7249, "step": 781250 }, { "epoch": 4.711110575125722, "grad_norm": 2.9920809268951416, "learning_rate": 7.276629453682951e-05, "loss": 4.8021, "step": 781300 }, { "epoch": 4.711412066907056, "grad_norm": 1.5713162422180176, "learning_rate": 7.274885098645695e-05, "loss": 4.3503, "step": 781350 }, { "epoch": 4.71171355868839, "grad_norm": 6.045583724975586, "learning_rate": 7.27314088577816e-05, "loss": 4.4035, "step": 781400 }, { "epoch": 4.7120150504697245, "grad_norm": 3.7040627002716064, "learning_rate": 7.271396815112435e-05, "loss": 4.3519, "step": 781450 }, { "epoch": 4.712316542251058, "grad_norm": 3.0959582328796387, "learning_rate": 7.269652886680628e-05, "loss": 4.6007, "step": 781500 }, { "epoch": 4.712618034032392, "grad_norm": 2.9723079204559326, "learning_rate": 7.267909100514828e-05, "loss": 4.7362, "step": 781550 }, { "epoch": 4.712919525813726, "grad_norm": 3.5682265758514404, "learning_rate": 7.26616545664712e-05, "loss": 4.7147, "step": 781600 }, { "epoch": 4.71322101759506, "grad_norm": 3.6857755184173584, "learning_rate": 7.2644219551096e-05, "loss": 4.7073, "step": 781650 }, { "epoch": 4.713522509376395, "grad_norm": 1.702995777130127, "learning_rate": 7.26267859593436e-05, "loss": 4.662, "step": 781700 }, { "epoch": 4.713824001157729, "grad_norm": 2.971359968185425, "learning_rate": 7.260935379153475e-05, "loss": 4.4252, "step": 781750 }, { "epoch": 4.714125492939062, "grad_norm": 2.822981357574463, "learning_rate": 7.259192304799029e-05, "loss": 4.5119, "step": 781800 }, { "epoch": 4.714426984720396, "grad_norm": 1.7097457647323608, "learning_rate": 7.257449372903107e-05, "loss": 5.0459, "step": 781850 }, { "epoch": 4.7147284765017305, "grad_norm": 4.063759803771973, "learning_rate": 7.255706583497781e-05, "loss": 4.5965, "step": 781900 }, { "epoch": 4.715029968283065, "grad_norm": 3.1679928302764893, "learning_rate": 7.253963936615119e-05, "loss": 4.6239, "step": 781950 }, { "epoch": 4.715331460064399, "grad_norm": 3.2806472778320312, "learning_rate": 7.2522214322872e-05, "loss": 4.87, "step": 782000 }, { "epoch": 4.715632951845732, "grad_norm": 2.4690439701080322, "learning_rate": 7.250479070546083e-05, "loss": 4.8483, "step": 782050 }, { "epoch": 4.7159344436270665, "grad_norm": 3.6881279945373535, "learning_rate": 7.248736851423841e-05, "loss": 4.4347, "step": 782100 }, { "epoch": 4.716235935408401, "grad_norm": 3.455413579940796, "learning_rate": 7.246994774952541e-05, "loss": 4.7775, "step": 782150 }, { "epoch": 4.716537427189735, "grad_norm": 3.3552989959716797, "learning_rate": 7.245252841164231e-05, "loss": 4.8167, "step": 782200 }, { "epoch": 4.716838918971069, "grad_norm": 2.5669538974761963, "learning_rate": 7.243511050090978e-05, "loss": 4.5142, "step": 782250 }, { "epoch": 4.717140410752403, "grad_norm": 3.5152714252471924, "learning_rate": 7.24176940176484e-05, "loss": 4.5772, "step": 782300 }, { "epoch": 4.717441902533737, "grad_norm": 1.3282274007797241, "learning_rate": 7.240027896217863e-05, "loss": 4.7317, "step": 782350 }, { "epoch": 4.717743394315071, "grad_norm": 2.470248222351074, "learning_rate": 7.238286533482094e-05, "loss": 4.5801, "step": 782400 }, { "epoch": 4.718044886096405, "grad_norm": 2.56012225151062, "learning_rate": 7.236545313589593e-05, "loss": 4.9783, "step": 782450 }, { "epoch": 4.718346377877739, "grad_norm": 5.494017601013184, "learning_rate": 7.234804236572386e-05, "loss": 4.8431, "step": 782500 }, { "epoch": 4.7186478696590735, "grad_norm": 2.836592435836792, "learning_rate": 7.233063302462529e-05, "loss": 4.8443, "step": 782550 }, { "epoch": 4.718949361440407, "grad_norm": 3.322430372238159, "learning_rate": 7.231322511292061e-05, "loss": 4.4626, "step": 782600 }, { "epoch": 4.719250853221741, "grad_norm": 3.456676721572876, "learning_rate": 7.229581863093012e-05, "loss": 4.8684, "step": 782650 }, { "epoch": 4.719552345003075, "grad_norm": 3.1296467781066895, "learning_rate": 7.227841357897425e-05, "loss": 4.583, "step": 782700 }, { "epoch": 4.719853836784409, "grad_norm": 1.4563134908676147, "learning_rate": 7.226100995737326e-05, "loss": 4.413, "step": 782750 }, { "epoch": 4.720155328565744, "grad_norm": 2.2090649604797363, "learning_rate": 7.22436077664474e-05, "loss": 4.8412, "step": 782800 }, { "epoch": 4.720456820347078, "grad_norm": 3.365468978881836, "learning_rate": 7.222620700651697e-05, "loss": 4.4913, "step": 782850 }, { "epoch": 4.720758312128411, "grad_norm": 3.6117937564849854, "learning_rate": 7.220880767790224e-05, "loss": 4.7433, "step": 782900 }, { "epoch": 4.721059803909745, "grad_norm": 2.76021671295166, "learning_rate": 7.219140978092337e-05, "loss": 4.8247, "step": 782950 }, { "epoch": 4.7213612956910795, "grad_norm": 6.209406852722168, "learning_rate": 7.217401331590055e-05, "loss": 4.6518, "step": 783000 }, { "epoch": 4.7213612956910795, "eval_loss": 5.094202041625977, "eval_runtime": 38.9346, "eval_samples_per_second": 13.15, "eval_steps_per_second": 6.575, "eval_tts_loss": 7.674693354299619, "step": 783000 }, { "epoch": 4.721662787472414, "grad_norm": 2.7825675010681152, "learning_rate": 7.215661828315401e-05, "loss": 4.7895, "step": 783050 }, { "epoch": 4.721964279253747, "grad_norm": 3.161006450653076, "learning_rate": 7.21392246830038e-05, "loss": 4.7207, "step": 783100 }, { "epoch": 4.722265771035081, "grad_norm": 3.0478663444519043, "learning_rate": 7.212183251577002e-05, "loss": 4.8581, "step": 783150 }, { "epoch": 4.7225672628164155, "grad_norm": 4.106105804443359, "learning_rate": 7.210444178177283e-05, "loss": 4.5451, "step": 783200 }, { "epoch": 4.72286875459775, "grad_norm": 3.9906442165374756, "learning_rate": 7.208705248133214e-05, "loss": 4.4832, "step": 783250 }, { "epoch": 4.723170246379084, "grad_norm": 2.5146615505218506, "learning_rate": 7.206966461476808e-05, "loss": 4.6582, "step": 783300 }, { "epoch": 4.723471738160418, "grad_norm": 0.9474396109580994, "learning_rate": 7.205227818240062e-05, "loss": 4.9795, "step": 783350 }, { "epoch": 4.723773229941751, "grad_norm": 3.279332399368286, "learning_rate": 7.203489318454984e-05, "loss": 4.7365, "step": 783400 }, { "epoch": 4.724074721723086, "grad_norm": 1.8342711925506592, "learning_rate": 7.20175096215355e-05, "loss": 3.9854, "step": 783450 }, { "epoch": 4.72437621350442, "grad_norm": 3.020864486694336, "learning_rate": 7.200012749367766e-05, "loss": 5.2118, "step": 783500 }, { "epoch": 4.724677705285754, "grad_norm": 2.4294919967651367, "learning_rate": 7.198274680129615e-05, "loss": 4.774, "step": 783550 }, { "epoch": 4.724979197067088, "grad_norm": 2.686894416809082, "learning_rate": 7.196536754471083e-05, "loss": 4.683, "step": 783600 }, { "epoch": 4.725280688848422, "grad_norm": 2.982572078704834, "learning_rate": 7.194798972424153e-05, "loss": 5.1806, "step": 783650 }, { "epoch": 4.725582180629756, "grad_norm": 1.276607871055603, "learning_rate": 7.193061334020817e-05, "loss": 4.4936, "step": 783700 }, { "epoch": 4.72588367241109, "grad_norm": 4.58088493347168, "learning_rate": 7.191323839293038e-05, "loss": 4.4577, "step": 783750 }, { "epoch": 4.726185164192424, "grad_norm": 4.407951831817627, "learning_rate": 7.189586488272802e-05, "loss": 5.3814, "step": 783800 }, { "epoch": 4.726486655973758, "grad_norm": 3.3998568058013916, "learning_rate": 7.187849280992086e-05, "loss": 4.8953, "step": 783850 }, { "epoch": 4.726788147755093, "grad_norm": 0.8805456757545471, "learning_rate": 7.186112217482854e-05, "loss": 4.6574, "step": 783900 }, { "epoch": 4.727089639536426, "grad_norm": 1.0430279970169067, "learning_rate": 7.184375297777073e-05, "loss": 4.6743, "step": 783950 }, { "epoch": 4.72739113131776, "grad_norm": 4.256295204162598, "learning_rate": 7.182638521906715e-05, "loss": 4.7242, "step": 784000 }, { "epoch": 4.727692623099094, "grad_norm": 2.4439001083374023, "learning_rate": 7.180901889903733e-05, "loss": 4.5033, "step": 784050 }, { "epoch": 4.7279941148804285, "grad_norm": 3.3305866718292236, "learning_rate": 7.179165401800092e-05, "loss": 4.8962, "step": 784100 }, { "epoch": 4.728295606661763, "grad_norm": 2.678546190261841, "learning_rate": 7.177429057627759e-05, "loss": 4.6515, "step": 784150 }, { "epoch": 4.728597098443096, "grad_norm": 3.838531732559204, "learning_rate": 7.175692857418675e-05, "loss": 4.4161, "step": 784200 }, { "epoch": 4.72889859022443, "grad_norm": 2.893465995788574, "learning_rate": 7.173956801204803e-05, "loss": 4.7627, "step": 784250 }, { "epoch": 4.7292000820057645, "grad_norm": 3.7189197540283203, "learning_rate": 7.172220889018087e-05, "loss": 4.4725, "step": 784300 }, { "epoch": 4.729501573787099, "grad_norm": 2.393772840499878, "learning_rate": 7.170485120890469e-05, "loss": 4.7442, "step": 784350 }, { "epoch": 4.729803065568433, "grad_norm": 2.285737991333008, "learning_rate": 7.168749496853901e-05, "loss": 4.8831, "step": 784400 }, { "epoch": 4.730104557349767, "grad_norm": 1.6721405982971191, "learning_rate": 7.167014016940326e-05, "loss": 4.4302, "step": 784450 }, { "epoch": 4.7304060491311, "grad_norm": 5.069298267364502, "learning_rate": 7.165278681181676e-05, "loss": 4.6062, "step": 784500 }, { "epoch": 4.730707540912435, "grad_norm": 2.53166127204895, "learning_rate": 7.16354348960989e-05, "loss": 4.6861, "step": 784550 }, { "epoch": 4.731009032693769, "grad_norm": 2.877708911895752, "learning_rate": 7.161808442256909e-05, "loss": 4.9603, "step": 784600 }, { "epoch": 4.731310524475103, "grad_norm": 3.160712480545044, "learning_rate": 7.160073539154653e-05, "loss": 4.3114, "step": 784650 }, { "epoch": 4.731612016256437, "grad_norm": 2.250446081161499, "learning_rate": 7.158338780335058e-05, "loss": 4.5683, "step": 784700 }, { "epoch": 4.7319135080377706, "grad_norm": 3.117680072784424, "learning_rate": 7.15660416583005e-05, "loss": 4.4821, "step": 784750 }, { "epoch": 4.732214999819105, "grad_norm": 3.227055788040161, "learning_rate": 7.154869695671543e-05, "loss": 4.6686, "step": 784800 }, { "epoch": 4.732516491600439, "grad_norm": 3.4025661945343018, "learning_rate": 7.153135369891464e-05, "loss": 4.4534, "step": 784850 }, { "epoch": 4.732817983381773, "grad_norm": 3.298039674758911, "learning_rate": 7.151401188521735e-05, "loss": 4.6219, "step": 784900 }, { "epoch": 4.733119475163107, "grad_norm": 3.4837801456451416, "learning_rate": 7.149667151594262e-05, "loss": 4.8563, "step": 784950 }, { "epoch": 4.733420966944442, "grad_norm": 3.6463887691497803, "learning_rate": 7.147933259140961e-05, "loss": 4.8124, "step": 785000 }, { "epoch": 4.733722458725775, "grad_norm": 3.6644434928894043, "learning_rate": 7.14619951119375e-05, "loss": 4.4689, "step": 785050 }, { "epoch": 4.734023950507109, "grad_norm": 3.4712650775909424, "learning_rate": 7.144465907784528e-05, "loss": 4.9095, "step": 785100 }, { "epoch": 4.734325442288443, "grad_norm": 3.433875560760498, "learning_rate": 7.142732448945194e-05, "loss": 4.6562, "step": 785150 }, { "epoch": 4.7346269340697775, "grad_norm": 3.5827181339263916, "learning_rate": 7.140999134707663e-05, "loss": 4.6138, "step": 785200 }, { "epoch": 4.734928425851111, "grad_norm": 2.3686254024505615, "learning_rate": 7.139265965103822e-05, "loss": 4.3498, "step": 785250 }, { "epoch": 4.735229917632445, "grad_norm": 3.256387948989868, "learning_rate": 7.137532940165573e-05, "loss": 4.8491, "step": 785300 }, { "epoch": 4.735531409413779, "grad_norm": 2.9939863681793213, "learning_rate": 7.135800059924815e-05, "loss": 4.5748, "step": 785350 }, { "epoch": 4.7358329011951135, "grad_norm": 4.137753009796143, "learning_rate": 7.134067324413429e-05, "loss": 4.4994, "step": 785400 }, { "epoch": 4.736134392976448, "grad_norm": 3.016484498977661, "learning_rate": 7.132334733663314e-05, "loss": 4.4545, "step": 785450 }, { "epoch": 4.736435884757782, "grad_norm": 2.9898293018341064, "learning_rate": 7.130602287706346e-05, "loss": 4.7295, "step": 785500 }, { "epoch": 4.736737376539115, "grad_norm": 2.9376211166381836, "learning_rate": 7.128869986574415e-05, "loss": 4.7605, "step": 785550 }, { "epoch": 4.737038868320449, "grad_norm": 3.906667709350586, "learning_rate": 7.127137830299395e-05, "loss": 4.5672, "step": 785600 }, { "epoch": 4.737340360101784, "grad_norm": 2.617941379547119, "learning_rate": 7.125405818913177e-05, "loss": 4.995, "step": 785650 }, { "epoch": 4.737641851883118, "grad_norm": 3.5771453380584717, "learning_rate": 7.123673952447617e-05, "loss": 4.6511, "step": 785700 }, { "epoch": 4.737943343664452, "grad_norm": 3.0261974334716797, "learning_rate": 7.121942230934599e-05, "loss": 4.7077, "step": 785750 }, { "epoch": 4.738244835445785, "grad_norm": 3.0145606994628906, "learning_rate": 7.120210654406e-05, "loss": 4.9542, "step": 785800 }, { "epoch": 4.7385463272271195, "grad_norm": 3.3112435340881348, "learning_rate": 7.118479222893671e-05, "loss": 4.2279, "step": 785850 }, { "epoch": 4.738847819008454, "grad_norm": 5.064829349517822, "learning_rate": 7.11674793642949e-05, "loss": 4.3503, "step": 785900 }, { "epoch": 4.739149310789788, "grad_norm": 2.51534104347229, "learning_rate": 7.115016795045313e-05, "loss": 4.781, "step": 785950 }, { "epoch": 4.739450802571122, "grad_norm": 3.683577537536621, "learning_rate": 7.113285798772994e-05, "loss": 4.6149, "step": 786000 }, { "epoch": 4.739450802571122, "eval_loss": 5.093106269836426, "eval_runtime": 39.1189, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.657067575346278, "step": 786000 }, { "epoch": 4.739752294352456, "grad_norm": 1.6739152669906616, "learning_rate": 7.111554947644397e-05, "loss": 4.7875, "step": 786050 }, { "epoch": 4.74005378613379, "grad_norm": 3.2442543506622314, "learning_rate": 7.109824241691378e-05, "loss": 4.8172, "step": 786100 }, { "epoch": 4.740355277915124, "grad_norm": 0.484971821308136, "learning_rate": 7.108093680945777e-05, "loss": 4.5017, "step": 786150 }, { "epoch": 4.740656769696458, "grad_norm": 2.4585888385772705, "learning_rate": 7.106363265439451e-05, "loss": 4.9107, "step": 786200 }, { "epoch": 4.740958261477792, "grad_norm": 4.256406784057617, "learning_rate": 7.104632995204249e-05, "loss": 4.6593, "step": 786250 }, { "epoch": 4.7412597532591265, "grad_norm": 2.7429091930389404, "learning_rate": 7.102902870272009e-05, "loss": 5.037, "step": 786300 }, { "epoch": 4.74156124504046, "grad_norm": 1.8298680782318115, "learning_rate": 7.101172890674567e-05, "loss": 4.5672, "step": 786350 }, { "epoch": 4.741862736821794, "grad_norm": 2.353632926940918, "learning_rate": 7.099443056443772e-05, "loss": 4.7564, "step": 786400 }, { "epoch": 4.742164228603128, "grad_norm": 3.8464837074279785, "learning_rate": 7.097713367611446e-05, "loss": 4.7044, "step": 786450 }, { "epoch": 4.7424657203844625, "grad_norm": 1.8821955919265747, "learning_rate": 7.095983824209429e-05, "loss": 4.3562, "step": 786500 }, { "epoch": 4.742767212165797, "grad_norm": 3.049870729446411, "learning_rate": 7.094254426269555e-05, "loss": 4.5087, "step": 786550 }, { "epoch": 4.743068703947131, "grad_norm": 6.472546577453613, "learning_rate": 7.092525173823642e-05, "loss": 4.9465, "step": 786600 }, { "epoch": 4.743370195728464, "grad_norm": 3.1681621074676514, "learning_rate": 7.090796066903518e-05, "loss": 4.4453, "step": 786650 }, { "epoch": 4.743671687509798, "grad_norm": 5.090301990509033, "learning_rate": 7.08906710554101e-05, "loss": 4.606, "step": 786700 }, { "epoch": 4.743973179291133, "grad_norm": 1.52064049243927, "learning_rate": 7.087338289767933e-05, "loss": 4.8246, "step": 786750 }, { "epoch": 4.744274671072467, "grad_norm": 3.0943870544433594, "learning_rate": 7.085609619616098e-05, "loss": 4.6637, "step": 786800 }, { "epoch": 4.7445761628538, "grad_norm": 1.310246229171753, "learning_rate": 7.083881095117321e-05, "loss": 4.5965, "step": 786850 }, { "epoch": 4.744877654635134, "grad_norm": 5.6298346519470215, "learning_rate": 7.082152716303423e-05, "loss": 4.8085, "step": 786900 }, { "epoch": 4.7451791464164685, "grad_norm": 4.651181697845459, "learning_rate": 7.080424483206201e-05, "loss": 4.9326, "step": 786950 }, { "epoch": 4.745480638197803, "grad_norm": 2.096792221069336, "learning_rate": 7.078696395857463e-05, "loss": 4.6982, "step": 787000 }, { "epoch": 4.745782129979137, "grad_norm": 2.854816198348999, "learning_rate": 7.076968454289017e-05, "loss": 4.653, "step": 787050 }, { "epoch": 4.746083621760471, "grad_norm": 2.4116170406341553, "learning_rate": 7.075240658532663e-05, "loss": 4.7924, "step": 787100 }, { "epoch": 4.7463851135418045, "grad_norm": 3.1991124153137207, "learning_rate": 7.073513008620188e-05, "loss": 4.4592, "step": 787150 }, { "epoch": 4.746686605323139, "grad_norm": 2.7177929878234863, "learning_rate": 7.071785504583402e-05, "loss": 4.4392, "step": 787200 }, { "epoch": 4.746988097104473, "grad_norm": 1.0409237146377563, "learning_rate": 7.070058146454082e-05, "loss": 4.5155, "step": 787250 }, { "epoch": 4.747289588885807, "grad_norm": 4.537574291229248, "learning_rate": 7.068330934264025e-05, "loss": 4.9125, "step": 787300 }, { "epoch": 4.747591080667141, "grad_norm": 2.6394598484039307, "learning_rate": 7.066603868045023e-05, "loss": 5.0629, "step": 787350 }, { "epoch": 4.747892572448475, "grad_norm": 3.2113263607025146, "learning_rate": 7.064876947828853e-05, "loss": 4.6967, "step": 787400 }, { "epoch": 4.748194064229809, "grad_norm": 3.7579357624053955, "learning_rate": 7.063150173647302e-05, "loss": 4.4119, "step": 787450 }, { "epoch": 4.748495556011143, "grad_norm": 3.331106662750244, "learning_rate": 7.061423545532146e-05, "loss": 4.6727, "step": 787500 }, { "epoch": 4.748797047792477, "grad_norm": 2.3977859020233154, "learning_rate": 7.059697063515156e-05, "loss": 4.7403, "step": 787550 }, { "epoch": 4.7490985395738115, "grad_norm": 1.7794030904769897, "learning_rate": 7.057970727628108e-05, "loss": 4.8203, "step": 787600 }, { "epoch": 4.749400031355146, "grad_norm": 2.4872078895568848, "learning_rate": 7.056244537902783e-05, "loss": 4.8751, "step": 787650 }, { "epoch": 4.749701523136479, "grad_norm": 2.2591443061828613, "learning_rate": 7.054518494370934e-05, "loss": 4.6895, "step": 787700 }, { "epoch": 4.750003014917813, "grad_norm": 2.8013601303100586, "learning_rate": 7.052792597064331e-05, "loss": 4.7795, "step": 787750 }, { "epoch": 4.750304506699147, "grad_norm": 2.8987252712249756, "learning_rate": 7.051066846014748e-05, "loss": 5.0873, "step": 787800 }, { "epoch": 4.750605998480482, "grad_norm": 2.5514135360717773, "learning_rate": 7.049341241253929e-05, "loss": 4.6849, "step": 787850 }, { "epoch": 4.750907490261816, "grad_norm": 2.8483166694641113, "learning_rate": 7.047615782813645e-05, "loss": 4.101, "step": 787900 }, { "epoch": 4.751208982043149, "grad_norm": 1.8381085395812988, "learning_rate": 7.045890470725643e-05, "loss": 4.4185, "step": 787950 }, { "epoch": 4.751510473824483, "grad_norm": 2.820000648498535, "learning_rate": 7.044165305021669e-05, "loss": 4.7419, "step": 788000 }, { "epoch": 4.7518119656058175, "grad_norm": 4.1277594566345215, "learning_rate": 7.04244028573348e-05, "loss": 4.5886, "step": 788050 }, { "epoch": 4.752113457387152, "grad_norm": 2.4058828353881836, "learning_rate": 7.040715412892827e-05, "loss": 4.6412, "step": 788100 }, { "epoch": 4.752414949168486, "grad_norm": 2.5357229709625244, "learning_rate": 7.038990686531444e-05, "loss": 4.5926, "step": 788150 }, { "epoch": 4.75271644094982, "grad_norm": 2.1704347133636475, "learning_rate": 7.037266106681074e-05, "loss": 4.8358, "step": 788200 }, { "epoch": 4.7530179327311535, "grad_norm": 3.1259615421295166, "learning_rate": 7.035541673373466e-05, "loss": 4.707, "step": 788250 }, { "epoch": 4.753319424512488, "grad_norm": 2.2457892894744873, "learning_rate": 7.033817386640346e-05, "loss": 4.4888, "step": 788300 }, { "epoch": 4.753620916293822, "grad_norm": 3.6532673835754395, "learning_rate": 7.032093246513444e-05, "loss": 4.6089, "step": 788350 }, { "epoch": 4.753922408075156, "grad_norm": 2.8512539863586426, "learning_rate": 7.0303692530245e-05, "loss": 4.9259, "step": 788400 }, { "epoch": 4.75422389985649, "grad_norm": 2.933145046234131, "learning_rate": 7.028645406205229e-05, "loss": 4.8325, "step": 788450 }, { "epoch": 4.754525391637824, "grad_norm": 1.5394340753555298, "learning_rate": 7.026921706087367e-05, "loss": 4.7835, "step": 788500 }, { "epoch": 4.754826883419158, "grad_norm": 3.9173145294189453, "learning_rate": 7.025198152702637e-05, "loss": 4.8108, "step": 788550 }, { "epoch": 4.755128375200492, "grad_norm": 4.582583427429199, "learning_rate": 7.023474746082748e-05, "loss": 4.5481, "step": 788600 }, { "epoch": 4.755429866981826, "grad_norm": 2.224060535430908, "learning_rate": 7.02175148625943e-05, "loss": 4.1964, "step": 788650 }, { "epoch": 4.75573135876316, "grad_norm": 1.2999677658081055, "learning_rate": 7.02002837326439e-05, "loss": 4.0524, "step": 788700 }, { "epoch": 4.756032850544495, "grad_norm": 3.1506710052490234, "learning_rate": 7.018305407129334e-05, "loss": 4.4477, "step": 788750 }, { "epoch": 4.756334342325828, "grad_norm": 3.416304111480713, "learning_rate": 7.016582587885976e-05, "loss": 4.2342, "step": 788800 }, { "epoch": 4.756635834107162, "grad_norm": 4.001834392547607, "learning_rate": 7.014859915566029e-05, "loss": 5.0018, "step": 788850 }, { "epoch": 4.756937325888496, "grad_norm": 2.5400514602661133, "learning_rate": 7.013137390201183e-05, "loss": 4.6259, "step": 788900 }, { "epoch": 4.757238817669831, "grad_norm": 2.3205955028533936, "learning_rate": 7.011415011823147e-05, "loss": 4.983, "step": 788950 }, { "epoch": 4.757540309451164, "grad_norm": 2.9968485832214355, "learning_rate": 7.009692780463622e-05, "loss": 4.9739, "step": 789000 }, { "epoch": 4.757540309451164, "eval_loss": 5.085927963256836, "eval_runtime": 39.3262, "eval_samples_per_second": 13.019, "eval_steps_per_second": 6.51, "eval_tts_loss": 7.702500124386016, "step": 789000 }, { "epoch": 4.757841801232498, "grad_norm": 1.5637317895889282, "learning_rate": 7.007970696154294e-05, "loss": 4.7209, "step": 789050 }, { "epoch": 4.758143293013832, "grad_norm": 3.9392077922821045, "learning_rate": 7.006248758926865e-05, "loss": 4.9471, "step": 789100 }, { "epoch": 4.7584447847951665, "grad_norm": 3.615332841873169, "learning_rate": 7.00452696881302e-05, "loss": 4.936, "step": 789150 }, { "epoch": 4.758746276576501, "grad_norm": 2.987699508666992, "learning_rate": 7.00280532584444e-05, "loss": 4.8317, "step": 789200 }, { "epoch": 4.759047768357835, "grad_norm": 1.69687819480896, "learning_rate": 7.001083830052816e-05, "loss": 4.7335, "step": 789250 }, { "epoch": 4.759349260139168, "grad_norm": 1.7932487726211548, "learning_rate": 6.999362481469834e-05, "loss": 4.4494, "step": 789300 }, { "epoch": 4.7596507519205025, "grad_norm": 4.716724872589111, "learning_rate": 6.997641280127164e-05, "loss": 4.7974, "step": 789350 }, { "epoch": 4.759952243701837, "grad_norm": 1.1956546306610107, "learning_rate": 6.995920226056484e-05, "loss": 4.3887, "step": 789400 }, { "epoch": 4.760253735483171, "grad_norm": 0.6930851340293884, "learning_rate": 6.994199319289476e-05, "loss": 4.7962, "step": 789450 }, { "epoch": 4.760555227264505, "grad_norm": 3.1844022274017334, "learning_rate": 6.992478559857806e-05, "loss": 4.2974, "step": 789500 }, { "epoch": 4.760856719045838, "grad_norm": 3.333385944366455, "learning_rate": 6.990757947793134e-05, "loss": 4.4991, "step": 789550 }, { "epoch": 4.761158210827173, "grad_norm": 3.0147619247436523, "learning_rate": 6.989037483127138e-05, "loss": 4.9287, "step": 789600 }, { "epoch": 4.761459702608507, "grad_norm": 2.816877841949463, "learning_rate": 6.987317165891469e-05, "loss": 4.8253, "step": 789650 }, { "epoch": 4.761761194389841, "grad_norm": 2.605597734451294, "learning_rate": 6.985596996117793e-05, "loss": 4.2353, "step": 789700 }, { "epoch": 4.762062686171175, "grad_norm": 2.768540382385254, "learning_rate": 6.983876973837774e-05, "loss": 4.5788, "step": 789750 }, { "epoch": 4.762364177952509, "grad_norm": 5.1566362380981445, "learning_rate": 6.982157099083053e-05, "loss": 5.0515, "step": 789800 }, { "epoch": 4.762665669733843, "grad_norm": 3.844386577606201, "learning_rate": 6.980437371885293e-05, "loss": 4.539, "step": 789850 }, { "epoch": 4.762967161515177, "grad_norm": 5.00886344909668, "learning_rate": 6.978717792276139e-05, "loss": 4.5996, "step": 789900 }, { "epoch": 4.763268653296511, "grad_norm": 2.7860238552093506, "learning_rate": 6.976998360287233e-05, "loss": 4.521, "step": 789950 }, { "epoch": 4.763570145077845, "grad_norm": 0.9153858423233032, "learning_rate": 6.975279075950221e-05, "loss": 4.5505, "step": 790000 }, { "epoch": 4.76387163685918, "grad_norm": 2.690141439437866, "learning_rate": 6.973559939296747e-05, "loss": 4.401, "step": 790050 }, { "epoch": 4.764173128640513, "grad_norm": 4.070663928985596, "learning_rate": 6.971840950358454e-05, "loss": 4.0743, "step": 790100 }, { "epoch": 4.764474620421847, "grad_norm": 3.1271450519561768, "learning_rate": 6.970122109166965e-05, "loss": 4.6627, "step": 790150 }, { "epoch": 4.764776112203181, "grad_norm": 2.938354253768921, "learning_rate": 6.968403415753919e-05, "loss": 4.363, "step": 790200 }, { "epoch": 4.7650776039845155, "grad_norm": 2.5926239490509033, "learning_rate": 6.966684870150952e-05, "loss": 4.8297, "step": 790250 }, { "epoch": 4.76537909576585, "grad_norm": 3.936291217803955, "learning_rate": 6.964966472389685e-05, "loss": 4.8332, "step": 790300 }, { "epoch": 4.765680587547184, "grad_norm": 2.335827350616455, "learning_rate": 6.96324822250174e-05, "loss": 4.5021, "step": 790350 }, { "epoch": 4.765982079328517, "grad_norm": 2.9147682189941406, "learning_rate": 6.961530120518748e-05, "loss": 5.151, "step": 790400 }, { "epoch": 4.7662835711098515, "grad_norm": 2.1932766437530518, "learning_rate": 6.959812166472316e-05, "loss": 4.5605, "step": 790450 }, { "epoch": 4.766585062891186, "grad_norm": 3.5336596965789795, "learning_rate": 6.958094360394067e-05, "loss": 4.834, "step": 790500 }, { "epoch": 4.76688655467252, "grad_norm": 1.4517956972122192, "learning_rate": 6.95637670231562e-05, "loss": 4.5564, "step": 790550 }, { "epoch": 4.767188046453853, "grad_norm": 2.496809244155884, "learning_rate": 6.954659192268577e-05, "loss": 4.4625, "step": 790600 }, { "epoch": 4.767489538235187, "grad_norm": 2.426528215408325, "learning_rate": 6.952941830284553e-05, "loss": 4.5893, "step": 790650 }, { "epoch": 4.767791030016522, "grad_norm": 2.4932918548583984, "learning_rate": 6.951224616395154e-05, "loss": 4.8729, "step": 790700 }, { "epoch": 4.768092521797856, "grad_norm": 3.168842315673828, "learning_rate": 6.949507550631973e-05, "loss": 4.711, "step": 790750 }, { "epoch": 4.76839401357919, "grad_norm": 2.4457828998565674, "learning_rate": 6.947790633026617e-05, "loss": 4.8235, "step": 790800 }, { "epoch": 4.768695505360524, "grad_norm": 3.3263907432556152, "learning_rate": 6.946073863610688e-05, "loss": 4.66, "step": 790850 }, { "epoch": 4.7689969971418575, "grad_norm": 3.0428853034973145, "learning_rate": 6.94435724241577e-05, "loss": 4.5013, "step": 790900 }, { "epoch": 4.769298488923192, "grad_norm": 4.327554702758789, "learning_rate": 6.942640769473462e-05, "loss": 4.708, "step": 790950 }, { "epoch": 4.769599980704526, "grad_norm": 2.3332810401916504, "learning_rate": 6.940924444815359e-05, "loss": 4.1617, "step": 791000 }, { "epoch": 4.76990147248586, "grad_norm": 2.6886472702026367, "learning_rate": 6.939208268473039e-05, "loss": 4.7826, "step": 791050 }, { "epoch": 4.770202964267194, "grad_norm": 3.6193361282348633, "learning_rate": 6.937492240478082e-05, "loss": 4.9882, "step": 791100 }, { "epoch": 4.770504456048528, "grad_norm": 1.8525949716567993, "learning_rate": 6.93577636086208e-05, "loss": 5.056, "step": 791150 }, { "epoch": 4.770805947829862, "grad_norm": 3.252012252807617, "learning_rate": 6.934060629656601e-05, "loss": 4.3731, "step": 791200 }, { "epoch": 4.771107439611196, "grad_norm": 2.06585693359375, "learning_rate": 6.932345046893224e-05, "loss": 4.5414, "step": 791250 }, { "epoch": 4.77140893139253, "grad_norm": 3.385460138320923, "learning_rate": 6.93062961260353e-05, "loss": 4.5963, "step": 791300 }, { "epoch": 4.7717104231738645, "grad_norm": 3.1336686611175537, "learning_rate": 6.928914326819076e-05, "loss": 4.8823, "step": 791350 }, { "epoch": 4.772011914955199, "grad_norm": 4.34786319732666, "learning_rate": 6.927199189571436e-05, "loss": 4.4328, "step": 791400 }, { "epoch": 4.772313406736532, "grad_norm": 2.908748149871826, "learning_rate": 6.925484200892181e-05, "loss": 4.5587, "step": 791450 }, { "epoch": 4.772614898517866, "grad_norm": 1.6415154933929443, "learning_rate": 6.923769360812866e-05, "loss": 4.7712, "step": 791500 }, { "epoch": 4.7729163902992005, "grad_norm": 3.348573684692383, "learning_rate": 6.922054669365045e-05, "loss": 4.4079, "step": 791550 }, { "epoch": 4.773217882080535, "grad_norm": 3.0185697078704834, "learning_rate": 6.920340126580283e-05, "loss": 4.2072, "step": 791600 }, { "epoch": 4.773519373861869, "grad_norm": 1.2512952089309692, "learning_rate": 6.918625732490128e-05, "loss": 4.5777, "step": 791650 }, { "epoch": 4.773820865643202, "grad_norm": 6.0173420906066895, "learning_rate": 6.916911487126134e-05, "loss": 4.7272, "step": 791700 }, { "epoch": 4.774122357424536, "grad_norm": 1.7738533020019531, "learning_rate": 6.915197390519851e-05, "loss": 4.7569, "step": 791750 }, { "epoch": 4.774423849205871, "grad_norm": 0.9442107081413269, "learning_rate": 6.913483442702821e-05, "loss": 4.5301, "step": 791800 }, { "epoch": 4.774725340987205, "grad_norm": 3.4677631855010986, "learning_rate": 6.91176964370659e-05, "loss": 4.1763, "step": 791850 }, { "epoch": 4.775026832768539, "grad_norm": 2.86895489692688, "learning_rate": 6.9100559935627e-05, "loss": 4.7164, "step": 791900 }, { "epoch": 4.775328324549873, "grad_norm": 2.8940718173980713, "learning_rate": 6.908342492302677e-05, "loss": 4.7038, "step": 791950 }, { "epoch": 4.7756298163312065, "grad_norm": 2.335092306137085, "learning_rate": 6.906629139958062e-05, "loss": 4.7142, "step": 792000 }, { "epoch": 4.7756298163312065, "eval_loss": 5.078537940979004, "eval_runtime": 39.2012, "eval_samples_per_second": 13.061, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.773912359170927, "step": 792000 }, { "epoch": 4.775931308112541, "grad_norm": 1.5684880018234253, "learning_rate": 6.904915936560393e-05, "loss": 4.5456, "step": 792050 }, { "epoch": 4.776232799893875, "grad_norm": 1.9046084880828857, "learning_rate": 6.90320288214119e-05, "loss": 4.8879, "step": 792100 }, { "epoch": 4.776534291675209, "grad_norm": 3.577291488647461, "learning_rate": 6.901489976731983e-05, "loss": 4.5545, "step": 792150 }, { "epoch": 4.776835783456543, "grad_norm": 3.2663090229034424, "learning_rate": 6.899777220364302e-05, "loss": 4.5311, "step": 792200 }, { "epoch": 4.777137275237877, "grad_norm": 3.127249240875244, "learning_rate": 6.898064613069655e-05, "loss": 4.5961, "step": 792250 }, { "epoch": 4.777438767019211, "grad_norm": 2.4065463542938232, "learning_rate": 6.896352154879573e-05, "loss": 4.6285, "step": 792300 }, { "epoch": 4.777740258800545, "grad_norm": 2.6925086975097656, "learning_rate": 6.894639845825563e-05, "loss": 4.8669, "step": 792350 }, { "epoch": 4.778041750581879, "grad_norm": 4.019041061401367, "learning_rate": 6.892927685939137e-05, "loss": 4.6018, "step": 792400 }, { "epoch": 4.7783432423632135, "grad_norm": 3.8756017684936523, "learning_rate": 6.891215675251806e-05, "loss": 4.5457, "step": 792450 }, { "epoch": 4.778644734144548, "grad_norm": 1.592057466506958, "learning_rate": 6.889503813795084e-05, "loss": 4.5082, "step": 792500 }, { "epoch": 4.778946225925881, "grad_norm": 3.5165672302246094, "learning_rate": 6.887792101600466e-05, "loss": 4.6309, "step": 792550 }, { "epoch": 4.779247717707215, "grad_norm": 2.849358081817627, "learning_rate": 6.886080538699456e-05, "loss": 4.9087, "step": 792600 }, { "epoch": 4.779549209488549, "grad_norm": 4.199710845947266, "learning_rate": 6.88436912512356e-05, "loss": 4.3359, "step": 792650 }, { "epoch": 4.779850701269884, "grad_norm": 3.9926483631134033, "learning_rate": 6.88265786090427e-05, "loss": 4.2659, "step": 792700 }, { "epoch": 4.780152193051217, "grad_norm": 3.271254062652588, "learning_rate": 6.880946746073071e-05, "loss": 4.2705, "step": 792750 }, { "epoch": 4.780453684832551, "grad_norm": 1.768300175666809, "learning_rate": 6.879235780661467e-05, "loss": 4.5648, "step": 792800 }, { "epoch": 4.780755176613885, "grad_norm": 2.160498857498169, "learning_rate": 6.877524964700934e-05, "loss": 5.1252, "step": 792850 }, { "epoch": 4.78105666839522, "grad_norm": 4.303016662597656, "learning_rate": 6.875814298222962e-05, "loss": 4.6102, "step": 792900 }, { "epoch": 4.781358160176554, "grad_norm": 2.6264970302581787, "learning_rate": 6.87410378125904e-05, "loss": 4.8083, "step": 792950 }, { "epoch": 4.781659651957888, "grad_norm": 2.8390681743621826, "learning_rate": 6.872393413840637e-05, "loss": 4.5937, "step": 793000 }, { "epoch": 4.781961143739221, "grad_norm": 3.3131468296051025, "learning_rate": 6.870683195999237e-05, "loss": 4.6022, "step": 793050 }, { "epoch": 4.7822626355205555, "grad_norm": 2.994441509246826, "learning_rate": 6.868973127766315e-05, "loss": 4.6198, "step": 793100 }, { "epoch": 4.78256412730189, "grad_norm": 2.7031240463256836, "learning_rate": 6.867263209173332e-05, "loss": 4.7114, "step": 793150 }, { "epoch": 4.782865619083224, "grad_norm": 2.071816921234131, "learning_rate": 6.865553440251763e-05, "loss": 4.6708, "step": 793200 }, { "epoch": 4.783167110864558, "grad_norm": 2.319911003112793, "learning_rate": 6.863843821033082e-05, "loss": 4.1792, "step": 793250 }, { "epoch": 4.7834686026458915, "grad_norm": 2.4843249320983887, "learning_rate": 6.862134351548737e-05, "loss": 4.8459, "step": 793300 }, { "epoch": 4.783770094427226, "grad_norm": 0.7056493163108826, "learning_rate": 6.860425031830196e-05, "loss": 4.6634, "step": 793350 }, { "epoch": 4.78407158620856, "grad_norm": 2.32025408744812, "learning_rate": 6.858715861908918e-05, "loss": 4.3574, "step": 793400 }, { "epoch": 4.784373077989894, "grad_norm": 3.257174491882324, "learning_rate": 6.85700684181636e-05, "loss": 4.7947, "step": 793450 }, { "epoch": 4.784674569771228, "grad_norm": 2.4690744876861572, "learning_rate": 6.855297971583972e-05, "loss": 4.2562, "step": 793500 }, { "epoch": 4.7849760615525625, "grad_norm": 1.0342388153076172, "learning_rate": 6.853589251243196e-05, "loss": 4.7641, "step": 793550 }, { "epoch": 4.785277553333896, "grad_norm": 3.5129942893981934, "learning_rate": 6.851880680825489e-05, "loss": 4.7209, "step": 793600 }, { "epoch": 4.78557904511523, "grad_norm": 5.070474624633789, "learning_rate": 6.850172260362285e-05, "loss": 5.041, "step": 793650 }, { "epoch": 4.785880536896564, "grad_norm": 4.659651756286621, "learning_rate": 6.84846398988503e-05, "loss": 4.813, "step": 793700 }, { "epoch": 4.786182028677898, "grad_norm": 2.3711118698120117, "learning_rate": 6.846755869425168e-05, "loss": 4.7929, "step": 793750 }, { "epoch": 4.786483520459233, "grad_norm": 3.9411768913269043, "learning_rate": 6.845047899014123e-05, "loss": 4.7495, "step": 793800 }, { "epoch": 4.786785012240566, "grad_norm": 4.086655139923096, "learning_rate": 6.84334007868334e-05, "loss": 4.764, "step": 793850 }, { "epoch": 4.7870865040219, "grad_norm": 3.0040645599365234, "learning_rate": 6.841632408464241e-05, "loss": 4.7868, "step": 793900 }, { "epoch": 4.787387995803234, "grad_norm": 2.8730263710021973, "learning_rate": 6.839924888388252e-05, "loss": 4.8935, "step": 793950 }, { "epoch": 4.787689487584569, "grad_norm": 2.346773862838745, "learning_rate": 6.838217518486798e-05, "loss": 4.6043, "step": 794000 }, { "epoch": 4.787990979365903, "grad_norm": 3.4556758403778076, "learning_rate": 6.83651029879131e-05, "loss": 4.6709, "step": 794050 }, { "epoch": 4.788292471147237, "grad_norm": 3.6256263256073, "learning_rate": 6.834803229333194e-05, "loss": 4.6834, "step": 794100 }, { "epoch": 4.78859396292857, "grad_norm": 1.077606201171875, "learning_rate": 6.833096310143873e-05, "loss": 4.6057, "step": 794150 }, { "epoch": 4.7888954547099045, "grad_norm": 3.6551170349121094, "learning_rate": 6.831389541254764e-05, "loss": 4.8, "step": 794200 }, { "epoch": 4.789196946491239, "grad_norm": 4.504553318023682, "learning_rate": 6.829682922697274e-05, "loss": 4.5444, "step": 794250 }, { "epoch": 4.789498438272573, "grad_norm": 4.96405553817749, "learning_rate": 6.827976454502805e-05, "loss": 4.686, "step": 794300 }, { "epoch": 4.789799930053906, "grad_norm": 3.1456613540649414, "learning_rate": 6.826270136702772e-05, "loss": 4.7715, "step": 794350 }, { "epoch": 4.7901014218352405, "grad_norm": 4.52561092376709, "learning_rate": 6.824563969328567e-05, "loss": 5.0052, "step": 794400 }, { "epoch": 4.790402913616575, "grad_norm": 1.0788639783859253, "learning_rate": 6.822857952411595e-05, "loss": 4.3689, "step": 794450 }, { "epoch": 4.790704405397909, "grad_norm": 2.6805002689361572, "learning_rate": 6.82115208598326e-05, "loss": 4.4728, "step": 794500 }, { "epoch": 4.791005897179243, "grad_norm": 2.8147544860839844, "learning_rate": 6.819446370074942e-05, "loss": 4.4953, "step": 794550 }, { "epoch": 4.791307388960577, "grad_norm": 1.0992125272750854, "learning_rate": 6.817740804718039e-05, "loss": 4.823, "step": 794600 }, { "epoch": 4.791608880741911, "grad_norm": 3.645192861557007, "learning_rate": 6.816035389943946e-05, "loss": 4.5231, "step": 794650 }, { "epoch": 4.791910372523245, "grad_norm": 1.927985429763794, "learning_rate": 6.814330125784043e-05, "loss": 4.4413, "step": 794700 }, { "epoch": 4.792211864304579, "grad_norm": 2.4390928745269775, "learning_rate": 6.812625012269707e-05, "loss": 4.5099, "step": 794750 }, { "epoch": 4.792513356085913, "grad_norm": 3.554497241973877, "learning_rate": 6.810920049432327e-05, "loss": 4.8733, "step": 794800 }, { "epoch": 4.792814847867247, "grad_norm": 3.6080057621002197, "learning_rate": 6.809215237303273e-05, "loss": 4.4664, "step": 794850 }, { "epoch": 4.793116339648581, "grad_norm": 3.105762004852295, "learning_rate": 6.807510575913924e-05, "loss": 4.8637, "step": 794900 }, { "epoch": 4.793417831429915, "grad_norm": 4.717528343200684, "learning_rate": 6.805806065295657e-05, "loss": 4.6546, "step": 794950 }, { "epoch": 4.793719323211249, "grad_norm": 2.664015054702759, "learning_rate": 6.80410170547983e-05, "loss": 4.428, "step": 795000 }, { "epoch": 4.793719323211249, "eval_loss": 5.078997611999512, "eval_runtime": 39.2197, "eval_samples_per_second": 13.055, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.744014932783737, "step": 795000 }, { "epoch": 4.794020814992583, "grad_norm": 3.0211474895477295, "learning_rate": 6.80239749649782e-05, "loss": 4.6244, "step": 795050 }, { "epoch": 4.794322306773918, "grad_norm": 2.6676740646362305, "learning_rate": 6.800693438380987e-05, "loss": 4.7861, "step": 795100 }, { "epoch": 4.794623798555252, "grad_norm": 2.6674022674560547, "learning_rate": 6.798989531160682e-05, "loss": 5.0759, "step": 795150 }, { "epoch": 4.794925290336585, "grad_norm": 3.694662094116211, "learning_rate": 6.797285774868273e-05, "loss": 4.454, "step": 795200 }, { "epoch": 4.795226782117919, "grad_norm": 3.027212619781494, "learning_rate": 6.795582169535119e-05, "loss": 4.8675, "step": 795250 }, { "epoch": 4.7955282738992535, "grad_norm": 2.6730992794036865, "learning_rate": 6.79387871519256e-05, "loss": 4.7175, "step": 795300 }, { "epoch": 4.795829765680588, "grad_norm": 1.7190028429031372, "learning_rate": 6.792175411871954e-05, "loss": 4.7622, "step": 795350 }, { "epoch": 4.796131257461922, "grad_norm": 2.4411895275115967, "learning_rate": 6.790472259604652e-05, "loss": 4.6801, "step": 795400 }, { "epoch": 4.796432749243255, "grad_norm": 4.13987398147583, "learning_rate": 6.788769258421991e-05, "loss": 4.5668, "step": 795450 }, { "epoch": 4.7967342410245895, "grad_norm": 3.2518131732940674, "learning_rate": 6.78706640835531e-05, "loss": 4.5857, "step": 795500 }, { "epoch": 4.797035732805924, "grad_norm": 4.216808795928955, "learning_rate": 6.785363709435956e-05, "loss": 4.7915, "step": 795550 }, { "epoch": 4.797337224587258, "grad_norm": 2.3759536743164062, "learning_rate": 6.783661161695254e-05, "loss": 4.3545, "step": 795600 }, { "epoch": 4.797638716368592, "grad_norm": 3.3076303005218506, "learning_rate": 6.781958765164543e-05, "loss": 4.5721, "step": 795650 }, { "epoch": 4.797940208149926, "grad_norm": 2.8462088108062744, "learning_rate": 6.780256519875158e-05, "loss": 4.6433, "step": 795700 }, { "epoch": 4.79824169993126, "grad_norm": 4.399953842163086, "learning_rate": 6.778554425858417e-05, "loss": 4.6332, "step": 795750 }, { "epoch": 4.798543191712594, "grad_norm": 3.869063377380371, "learning_rate": 6.776852483145648e-05, "loss": 4.6548, "step": 795800 }, { "epoch": 4.798844683493928, "grad_norm": 4.217583179473877, "learning_rate": 6.775150691768182e-05, "loss": 4.3123, "step": 795850 }, { "epoch": 4.799146175275262, "grad_norm": 4.713292121887207, "learning_rate": 6.773449051757328e-05, "loss": 4.7347, "step": 795900 }, { "epoch": 4.799447667056596, "grad_norm": 2.500065326690674, "learning_rate": 6.771747563144401e-05, "loss": 4.7365, "step": 795950 }, { "epoch": 4.79974915883793, "grad_norm": 2.432218551635742, "learning_rate": 6.770046225960721e-05, "loss": 4.537, "step": 796000 }, { "epoch": 4.800050650619264, "grad_norm": 3.134345531463623, "learning_rate": 6.76834504023759e-05, "loss": 4.644, "step": 796050 }, { "epoch": 4.800352142400598, "grad_norm": 2.8182618618011475, "learning_rate": 6.766644006006323e-05, "loss": 4.4996, "step": 796100 }, { "epoch": 4.800653634181932, "grad_norm": 3.460371971130371, "learning_rate": 6.764943123298227e-05, "loss": 4.7324, "step": 796150 }, { "epoch": 4.800955125963267, "grad_norm": 2.833353281021118, "learning_rate": 6.763242392144597e-05, "loss": 4.4627, "step": 796200 }, { "epoch": 4.801256617744601, "grad_norm": 2.9375791549682617, "learning_rate": 6.76154181257674e-05, "loss": 4.9867, "step": 796250 }, { "epoch": 4.801558109525934, "grad_norm": 1.4654451608657837, "learning_rate": 6.75984138462595e-05, "loss": 4.7648, "step": 796300 }, { "epoch": 4.801859601307268, "grad_norm": 2.4815053939819336, "learning_rate": 6.758141108323515e-05, "loss": 4.5508, "step": 796350 }, { "epoch": 4.8021610930886025, "grad_norm": 3.6277201175689697, "learning_rate": 6.75644098370073e-05, "loss": 4.892, "step": 796400 }, { "epoch": 4.802462584869937, "grad_norm": 2.372688055038452, "learning_rate": 6.754741010788891e-05, "loss": 4.4603, "step": 796450 }, { "epoch": 4.80276407665127, "grad_norm": 2.8803534507751465, "learning_rate": 6.75304118961927e-05, "loss": 4.8853, "step": 796500 }, { "epoch": 4.803065568432604, "grad_norm": 2.0998458862304688, "learning_rate": 6.751341520223158e-05, "loss": 4.6215, "step": 796550 }, { "epoch": 4.8033670602139384, "grad_norm": 3.5528016090393066, "learning_rate": 6.74964200263184e-05, "loss": 4.8427, "step": 796600 }, { "epoch": 4.803668551995273, "grad_norm": 2.4151763916015625, "learning_rate": 6.747942636876588e-05, "loss": 4.5643, "step": 796650 }, { "epoch": 4.803970043776607, "grad_norm": 1.5655107498168945, "learning_rate": 6.746243422988667e-05, "loss": 4.3704, "step": 796700 }, { "epoch": 4.804271535557941, "grad_norm": 2.7487521171569824, "learning_rate": 6.744544360999358e-05, "loss": 4.7226, "step": 796750 }, { "epoch": 4.804573027339274, "grad_norm": 3.928347587585449, "learning_rate": 6.742845450939935e-05, "loss": 4.6099, "step": 796800 }, { "epoch": 4.804874519120609, "grad_norm": 2.912712574005127, "learning_rate": 6.741146692841651e-05, "loss": 4.4967, "step": 796850 }, { "epoch": 4.805176010901943, "grad_norm": 3.70540189743042, "learning_rate": 6.739448086735779e-05, "loss": 4.9153, "step": 796900 }, { "epoch": 4.805477502683277, "grad_norm": 3.989339590072632, "learning_rate": 6.737749632653581e-05, "loss": 4.7502, "step": 796950 }, { "epoch": 4.805778994464611, "grad_norm": 3.0553925037384033, "learning_rate": 6.736051330626303e-05, "loss": 4.2104, "step": 797000 }, { "epoch": 4.8060804862459445, "grad_norm": 2.611320972442627, "learning_rate": 6.734353180685213e-05, "loss": 4.3796, "step": 797050 }, { "epoch": 4.806381978027279, "grad_norm": 3.884072780609131, "learning_rate": 6.732655182861558e-05, "loss": 4.787, "step": 797100 }, { "epoch": 4.806683469808613, "grad_norm": 2.9745397567749023, "learning_rate": 6.73095733718658e-05, "loss": 4.6403, "step": 797150 }, { "epoch": 4.806984961589947, "grad_norm": 1.7323168516159058, "learning_rate": 6.729259643691531e-05, "loss": 4.763, "step": 797200 }, { "epoch": 4.807286453371281, "grad_norm": 2.4049971103668213, "learning_rate": 6.727562102407663e-05, "loss": 5.2529, "step": 797250 }, { "epoch": 4.807587945152616, "grad_norm": 4.13689661026001, "learning_rate": 6.725864713366203e-05, "loss": 4.7534, "step": 797300 }, { "epoch": 4.807889436933949, "grad_norm": 2.6790547370910645, "learning_rate": 6.724167476598392e-05, "loss": 4.3796, "step": 797350 }, { "epoch": 4.808190928715283, "grad_norm": 2.7920665740966797, "learning_rate": 6.722470392135476e-05, "loss": 4.8132, "step": 797400 }, { "epoch": 4.808492420496617, "grad_norm": 3.4701366424560547, "learning_rate": 6.720773460008681e-05, "loss": 4.4778, "step": 797450 }, { "epoch": 4.8087939122779515, "grad_norm": 1.6663250923156738, "learning_rate": 6.71907668024923e-05, "loss": 4.9217, "step": 797500 }, { "epoch": 4.809095404059286, "grad_norm": 1.6108503341674805, "learning_rate": 6.71738005288836e-05, "loss": 4.369, "step": 797550 }, { "epoch": 4.809396895840619, "grad_norm": 4.312994003295898, "learning_rate": 6.715683577957283e-05, "loss": 4.8384, "step": 797600 }, { "epoch": 4.809698387621953, "grad_norm": 2.4071264266967773, "learning_rate": 6.713987255487228e-05, "loss": 4.7607, "step": 797650 }, { "epoch": 4.809999879403287, "grad_norm": 2.7340235710144043, "learning_rate": 6.712291085509419e-05, "loss": 4.9612, "step": 797700 }, { "epoch": 4.810301371184622, "grad_norm": 3.3952910900115967, "learning_rate": 6.71059506805506e-05, "loss": 4.1787, "step": 797750 }, { "epoch": 4.810602862965956, "grad_norm": 2.736158847808838, "learning_rate": 6.708899203155374e-05, "loss": 4.8546, "step": 797800 }, { "epoch": 4.81090435474729, "grad_norm": 3.5587363243103027, "learning_rate": 6.70720349084156e-05, "loss": 4.516, "step": 797850 }, { "epoch": 4.811205846528623, "grad_norm": 2.3833796977996826, "learning_rate": 6.70550793114484e-05, "loss": 4.4057, "step": 797900 }, { "epoch": 4.811507338309958, "grad_norm": 4.137153625488281, "learning_rate": 6.703812524096401e-05, "loss": 4.4534, "step": 797950 }, { "epoch": 4.811808830091292, "grad_norm": 1.4906933307647705, "learning_rate": 6.702117269727459e-05, "loss": 4.6852, "step": 798000 }, { "epoch": 4.811808830091292, "eval_loss": 5.082396507263184, "eval_runtime": 39.1699, "eval_samples_per_second": 13.071, "eval_steps_per_second": 6.536, "eval_tts_loss": 7.724183138562828, "step": 798000 }, { "epoch": 4.812110321872626, "grad_norm": 3.2546024322509766, "learning_rate": 6.700422168069202e-05, "loss": 4.665, "step": 798050 }, { "epoch": 4.812411813653959, "grad_norm": 3.2487595081329346, "learning_rate": 6.698727219152832e-05, "loss": 5.0082, "step": 798100 }, { "epoch": 4.8127133054352935, "grad_norm": 5.074887275695801, "learning_rate": 6.697032423009545e-05, "loss": 4.7764, "step": 798150 }, { "epoch": 4.813014797216628, "grad_norm": 4.1371259689331055, "learning_rate": 6.695337779670522e-05, "loss": 4.782, "step": 798200 }, { "epoch": 4.813316288997962, "grad_norm": 2.92002534866333, "learning_rate": 6.693643289166961e-05, "loss": 4.531, "step": 798250 }, { "epoch": 4.813617780779296, "grad_norm": 2.8525912761688232, "learning_rate": 6.691948951530042e-05, "loss": 4.4385, "step": 798300 }, { "epoch": 4.81391927256063, "grad_norm": 2.790026903152466, "learning_rate": 6.690254766790943e-05, "loss": 4.512, "step": 798350 }, { "epoch": 4.814220764341964, "grad_norm": 1.1074554920196533, "learning_rate": 6.688560734980843e-05, "loss": 4.8267, "step": 798400 }, { "epoch": 4.814522256123298, "grad_norm": 2.3100850582122803, "learning_rate": 6.686866856130927e-05, "loss": 4.5132, "step": 798450 }, { "epoch": 4.814823747904632, "grad_norm": 2.798492193222046, "learning_rate": 6.685173130272359e-05, "loss": 4.6142, "step": 798500 }, { "epoch": 4.815125239685966, "grad_norm": 2.068213701248169, "learning_rate": 6.683479557436315e-05, "loss": 4.4874, "step": 798550 }, { "epoch": 4.8154267314673005, "grad_norm": 3.2102911472320557, "learning_rate": 6.681786137653964e-05, "loss": 4.8967, "step": 798600 }, { "epoch": 4.815728223248634, "grad_norm": 3.70621395111084, "learning_rate": 6.680092870956469e-05, "loss": 4.8512, "step": 798650 }, { "epoch": 4.816029715029968, "grad_norm": 3.238798141479492, "learning_rate": 6.678399757374987e-05, "loss": 4.5457, "step": 798700 }, { "epoch": 4.816331206811302, "grad_norm": 3.622950553894043, "learning_rate": 6.676706796940688e-05, "loss": 4.6636, "step": 798750 }, { "epoch": 4.816632698592636, "grad_norm": 3.3541905879974365, "learning_rate": 6.675013989684715e-05, "loss": 4.6921, "step": 798800 }, { "epoch": 4.816934190373971, "grad_norm": 2.634819984436035, "learning_rate": 6.673321335638233e-05, "loss": 4.4022, "step": 798850 }, { "epoch": 4.817235682155305, "grad_norm": 2.8041226863861084, "learning_rate": 6.671628834832391e-05, "loss": 4.0721, "step": 798900 }, { "epoch": 4.817537173936638, "grad_norm": 2.7435100078582764, "learning_rate": 6.66993648729833e-05, "loss": 4.0639, "step": 798950 }, { "epoch": 4.817838665717972, "grad_norm": 1.3989289999008179, "learning_rate": 6.668244293067201e-05, "loss": 4.7688, "step": 799000 }, { "epoch": 4.818140157499307, "grad_norm": 2.508556604385376, "learning_rate": 6.66655225217015e-05, "loss": 4.4125, "step": 799050 }, { "epoch": 4.818441649280641, "grad_norm": 2.834094285964966, "learning_rate": 6.664860364638315e-05, "loss": 4.5772, "step": 799100 }, { "epoch": 4.818743141061975, "grad_norm": 3.323899030685425, "learning_rate": 6.663168630502821e-05, "loss": 4.7233, "step": 799150 }, { "epoch": 4.819044632843308, "grad_norm": 1.4082106351852417, "learning_rate": 6.66147704979482e-05, "loss": 4.6458, "step": 799200 }, { "epoch": 4.8193461246246425, "grad_norm": 2.903108835220337, "learning_rate": 6.659785622545427e-05, "loss": 4.4, "step": 799250 }, { "epoch": 4.819647616405977, "grad_norm": 1.7749115228652954, "learning_rate": 6.658094348785776e-05, "loss": 4.4051, "step": 799300 }, { "epoch": 4.819949108187311, "grad_norm": 3.38643479347229, "learning_rate": 6.656403228547001e-05, "loss": 5.0464, "step": 799350 }, { "epoch": 4.820250599968645, "grad_norm": 4.903872013092041, "learning_rate": 6.654712261860213e-05, "loss": 4.8828, "step": 799400 }, { "epoch": 4.820552091749979, "grad_norm": 4.599870204925537, "learning_rate": 6.653021448756542e-05, "loss": 4.7145, "step": 799450 }, { "epoch": 4.820853583531313, "grad_norm": 3.5383591651916504, "learning_rate": 6.651330789267097e-05, "loss": 5.0828, "step": 799500 }, { "epoch": 4.821155075312647, "grad_norm": 3.6200432777404785, "learning_rate": 6.649640283422992e-05, "loss": 4.455, "step": 799550 }, { "epoch": 4.821456567093981, "grad_norm": 3.0891928672790527, "learning_rate": 6.647949931255339e-05, "loss": 4.8714, "step": 799600 }, { "epoch": 4.821758058875315, "grad_norm": 1.2611448764801025, "learning_rate": 6.646259732795255e-05, "loss": 4.9899, "step": 799650 }, { "epoch": 4.8220595506566495, "grad_norm": 2.578047037124634, "learning_rate": 6.644569688073834e-05, "loss": 4.67, "step": 799700 }, { "epoch": 4.822361042437983, "grad_norm": 3.0726993083953857, "learning_rate": 6.642879797122182e-05, "loss": 4.7678, "step": 799750 }, { "epoch": 4.822662534219317, "grad_norm": 1.5368232727050781, "learning_rate": 6.641190059971408e-05, "loss": 4.6924, "step": 799800 }, { "epoch": 4.822964026000651, "grad_norm": 2.445920705795288, "learning_rate": 6.639500476652603e-05, "loss": 5.0361, "step": 799850 }, { "epoch": 4.823265517781985, "grad_norm": 3.564973831176758, "learning_rate": 6.637811047196855e-05, "loss": 4.7302, "step": 799900 }, { "epoch": 4.82356700956332, "grad_norm": 5.685024261474609, "learning_rate": 6.636121771635266e-05, "loss": 4.5533, "step": 799950 }, { "epoch": 4.823868501344654, "grad_norm": 3.0114457607269287, "learning_rate": 6.634432649998915e-05, "loss": 4.473, "step": 800000 }, { "epoch": 4.824169993125987, "grad_norm": 3.71372652053833, "learning_rate": 6.632743682318891e-05, "loss": 4.5996, "step": 800050 }, { "epoch": 4.824471484907321, "grad_norm": 2.8761508464813232, "learning_rate": 6.631054868626288e-05, "loss": 4.557, "step": 800100 }, { "epoch": 4.824772976688656, "grad_norm": 4.202939987182617, "learning_rate": 6.629366208952169e-05, "loss": 4.5292, "step": 800150 }, { "epoch": 4.82507446846999, "grad_norm": 3.07669734954834, "learning_rate": 6.627677703327618e-05, "loss": 4.7696, "step": 800200 }, { "epoch": 4.825375960251323, "grad_norm": 3.3498988151550293, "learning_rate": 6.625989351783718e-05, "loss": 4.362, "step": 800250 }, { "epoch": 4.825677452032657, "grad_norm": 3.070852279663086, "learning_rate": 6.624301154351534e-05, "loss": 4.9166, "step": 800300 }, { "epoch": 4.8259789438139915, "grad_norm": 3.059659004211426, "learning_rate": 6.62261311106213e-05, "loss": 4.4192, "step": 800350 }, { "epoch": 4.826280435595326, "grad_norm": 3.0343923568725586, "learning_rate": 6.620925221946573e-05, "loss": 4.5667, "step": 800400 }, { "epoch": 4.82658192737666, "grad_norm": 3.4763996601104736, "learning_rate": 6.619237487035938e-05, "loss": 5.2021, "step": 800450 }, { "epoch": 4.826883419157994, "grad_norm": 1.9134899377822876, "learning_rate": 6.617549906361272e-05, "loss": 4.4352, "step": 800500 }, { "epoch": 4.8271849109393274, "grad_norm": 2.4795010089874268, "learning_rate": 6.615862479953637e-05, "loss": 4.8399, "step": 800550 }, { "epoch": 4.827486402720662, "grad_norm": 0.8629022836685181, "learning_rate": 6.614175207844092e-05, "loss": 4.9819, "step": 800600 }, { "epoch": 4.827787894501996, "grad_norm": 3.794135570526123, "learning_rate": 6.612488090063686e-05, "loss": 4.3001, "step": 800650 }, { "epoch": 4.82808938628333, "grad_norm": 3.25551438331604, "learning_rate": 6.610801126643463e-05, "loss": 4.9275, "step": 800700 }, { "epoch": 4.828390878064664, "grad_norm": 2.8161239624023438, "learning_rate": 6.609114317614477e-05, "loss": 4.7169, "step": 800750 }, { "epoch": 4.828692369845998, "grad_norm": 1.0621577501296997, "learning_rate": 6.607427663007763e-05, "loss": 4.2037, "step": 800800 }, { "epoch": 4.828993861627332, "grad_norm": 3.758314847946167, "learning_rate": 6.605741162854365e-05, "loss": 4.4095, "step": 800850 }, { "epoch": 4.829295353408666, "grad_norm": 2.820070505142212, "learning_rate": 6.604054817185327e-05, "loss": 4.3695, "step": 800900 }, { "epoch": 4.82959684519, "grad_norm": 2.6270763874053955, "learning_rate": 6.602368626031675e-05, "loss": 4.9178, "step": 800950 }, { "epoch": 4.829898336971334, "grad_norm": 5.539656162261963, "learning_rate": 6.600682589424449e-05, "loss": 4.6607, "step": 801000 }, { "epoch": 4.829898336971334, "eval_loss": 5.072183609008789, "eval_runtime": 39.191, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 7.743450705763022, "step": 801000 }, { "epoch": 4.830199828752669, "grad_norm": 1.363946557044983, "learning_rate": 6.598996707394671e-05, "loss": 4.4953, "step": 801050 }, { "epoch": 4.830501320534002, "grad_norm": 5.369067668914795, "learning_rate": 6.597310979973368e-05, "loss": 4.5154, "step": 801100 }, { "epoch": 4.830802812315336, "grad_norm": 3.1669437885284424, "learning_rate": 6.595625407191564e-05, "loss": 4.919, "step": 801150 }, { "epoch": 4.83110430409667, "grad_norm": 2.867583990097046, "learning_rate": 6.593939989080285e-05, "loss": 4.6454, "step": 801200 }, { "epoch": 4.831405795878005, "grad_norm": 3.051896333694458, "learning_rate": 6.592254725670539e-05, "loss": 4.4646, "step": 801250 }, { "epoch": 4.831707287659339, "grad_norm": 3.766162872314453, "learning_rate": 6.590569616993348e-05, "loss": 4.479, "step": 801300 }, { "epoch": 4.832008779440672, "grad_norm": 2.8188421726226807, "learning_rate": 6.588884663079725e-05, "loss": 4.5837, "step": 801350 }, { "epoch": 4.832310271222006, "grad_norm": 3.564319610595703, "learning_rate": 6.587199863960674e-05, "loss": 4.899, "step": 801400 }, { "epoch": 4.8326117630033405, "grad_norm": 3.8654119968414307, "learning_rate": 6.585515219667208e-05, "loss": 4.9875, "step": 801450 }, { "epoch": 4.832913254784675, "grad_norm": 3.0137288570404053, "learning_rate": 6.583830730230326e-05, "loss": 4.8189, "step": 801500 }, { "epoch": 4.833214746566009, "grad_norm": 0.8169920444488525, "learning_rate": 6.582146395681023e-05, "loss": 4.294, "step": 801550 }, { "epoch": 4.833516238347343, "grad_norm": 1.1260316371917725, "learning_rate": 6.580462216050303e-05, "loss": 4.6616, "step": 801600 }, { "epoch": 4.833817730128676, "grad_norm": 3.1445138454437256, "learning_rate": 6.578778191369167e-05, "loss": 4.2842, "step": 801650 }, { "epoch": 4.834119221910011, "grad_norm": 3.4181480407714844, "learning_rate": 6.577094321668595e-05, "loss": 4.4791, "step": 801700 }, { "epoch": 4.834420713691345, "grad_norm": 4.137923717498779, "learning_rate": 6.57541060697958e-05, "loss": 4.6092, "step": 801750 }, { "epoch": 4.834722205472679, "grad_norm": 2.903402090072632, "learning_rate": 6.573727047333118e-05, "loss": 4.5268, "step": 801800 }, { "epoch": 4.835023697254012, "grad_norm": 0.9482282400131226, "learning_rate": 6.572043642760184e-05, "loss": 4.1868, "step": 801850 }, { "epoch": 4.835325189035347, "grad_norm": 5.792799949645996, "learning_rate": 6.570360393291753e-05, "loss": 5.088, "step": 801900 }, { "epoch": 4.835626680816681, "grad_norm": 2.7008211612701416, "learning_rate": 6.568677298958817e-05, "loss": 4.8572, "step": 801950 }, { "epoch": 4.835928172598015, "grad_norm": 2.8216702938079834, "learning_rate": 6.566994359792335e-05, "loss": 4.9105, "step": 802000 }, { "epoch": 4.836229664379349, "grad_norm": 4.080445766448975, "learning_rate": 6.565311575823289e-05, "loss": 4.5056, "step": 802050 }, { "epoch": 4.836531156160683, "grad_norm": 2.680408239364624, "learning_rate": 6.56362894708265e-05, "loss": 4.4228, "step": 802100 }, { "epoch": 4.836832647942017, "grad_norm": 5.4332098960876465, "learning_rate": 6.561946473601377e-05, "loss": 4.7007, "step": 802150 }, { "epoch": 4.837134139723351, "grad_norm": 3.454535484313965, "learning_rate": 6.560264155410444e-05, "loss": 4.3221, "step": 802200 }, { "epoch": 4.837435631504685, "grad_norm": 3.7614636421203613, "learning_rate": 6.558581992540798e-05, "loss": 5.0007, "step": 802250 }, { "epoch": 4.837737123286019, "grad_norm": 1.0633316040039062, "learning_rate": 6.55689998502341e-05, "loss": 4.6539, "step": 802300 }, { "epoch": 4.8380386150673536, "grad_norm": 3.9141898155212402, "learning_rate": 6.555218132889223e-05, "loss": 4.9525, "step": 802350 }, { "epoch": 4.838340106848687, "grad_norm": 3.8868021965026855, "learning_rate": 6.553536436169201e-05, "loss": 4.8747, "step": 802400 }, { "epoch": 4.838641598630021, "grad_norm": 2.6907265186309814, "learning_rate": 6.55185489489428e-05, "loss": 4.8498, "step": 802450 }, { "epoch": 4.838943090411355, "grad_norm": 2.832965135574341, "learning_rate": 6.550173509095415e-05, "loss": 4.4578, "step": 802500 }, { "epoch": 4.8392445821926895, "grad_norm": 2.9713101387023926, "learning_rate": 6.548492278803552e-05, "loss": 5.3071, "step": 802550 }, { "epoch": 4.839546073974024, "grad_norm": 3.9551846981048584, "learning_rate": 6.546811204049624e-05, "loss": 4.374, "step": 802600 }, { "epoch": 4.839847565755358, "grad_norm": 2.616661310195923, "learning_rate": 6.545130284864576e-05, "loss": 5.0391, "step": 802650 }, { "epoch": 4.840149057536691, "grad_norm": 0.4981392025947571, "learning_rate": 6.543449521279342e-05, "loss": 4.4821, "step": 802700 }, { "epoch": 4.840450549318025, "grad_norm": 2.2003777027130127, "learning_rate": 6.541768913324845e-05, "loss": 4.8942, "step": 802750 }, { "epoch": 4.84075204109936, "grad_norm": 3.6424710750579834, "learning_rate": 6.54008846103202e-05, "loss": 4.422, "step": 802800 }, { "epoch": 4.841053532880694, "grad_norm": 4.64838171005249, "learning_rate": 6.538408164431798e-05, "loss": 5.0826, "step": 802850 }, { "epoch": 4.841355024662028, "grad_norm": 3.0521419048309326, "learning_rate": 6.536728023555095e-05, "loss": 4.373, "step": 802900 }, { "epoch": 4.841656516443361, "grad_norm": 3.1271073818206787, "learning_rate": 6.535048038432833e-05, "loss": 4.5084, "step": 802950 }, { "epoch": 4.841958008224696, "grad_norm": 4.261964797973633, "learning_rate": 6.533368209095939e-05, "loss": 4.4706, "step": 803000 }, { "epoch": 4.84225950000603, "grad_norm": 2.8162684440612793, "learning_rate": 6.531688535575321e-05, "loss": 4.3778, "step": 803050 }, { "epoch": 4.842560991787364, "grad_norm": 1.616203784942627, "learning_rate": 6.530009017901885e-05, "loss": 4.7935, "step": 803100 }, { "epoch": 4.842862483568698, "grad_norm": 3.461130142211914, "learning_rate": 6.52832965610655e-05, "loss": 4.6475, "step": 803150 }, { "epoch": 4.843163975350032, "grad_norm": 3.3436625003814697, "learning_rate": 6.526650450220212e-05, "loss": 4.8674, "step": 803200 }, { "epoch": 4.843465467131366, "grad_norm": 3.502488136291504, "learning_rate": 6.524971400273783e-05, "loss": 4.5636, "step": 803250 }, { "epoch": 4.8437669589127, "grad_norm": 1.4735102653503418, "learning_rate": 6.523292506298163e-05, "loss": 4.5468, "step": 803300 }, { "epoch": 4.844068450694034, "grad_norm": 2.580475091934204, "learning_rate": 6.521613768324243e-05, "loss": 4.6736, "step": 803350 }, { "epoch": 4.844369942475368, "grad_norm": 2.579448699951172, "learning_rate": 6.519935186382929e-05, "loss": 4.5607, "step": 803400 }, { "epoch": 4.8446714342567025, "grad_norm": 2.5666842460632324, "learning_rate": 6.518256760505101e-05, "loss": 4.4445, "step": 803450 }, { "epoch": 4.844972926038036, "grad_norm": 1.9064315557479858, "learning_rate": 6.516578490721658e-05, "loss": 4.7373, "step": 803500 }, { "epoch": 4.84527441781937, "grad_norm": 3.4439077377319336, "learning_rate": 6.514900377063476e-05, "loss": 4.7321, "step": 803550 }, { "epoch": 4.845575909600704, "grad_norm": 1.5302696228027344, "learning_rate": 6.513222419561444e-05, "loss": 4.5086, "step": 803600 }, { "epoch": 4.8458774013820385, "grad_norm": 2.1467723846435547, "learning_rate": 6.511544618246446e-05, "loss": 4.5141, "step": 803650 }, { "epoch": 4.846178893163373, "grad_norm": 2.7936127185821533, "learning_rate": 6.509866973149351e-05, "loss": 4.6344, "step": 803700 }, { "epoch": 4.846480384944707, "grad_norm": 3.49241304397583, "learning_rate": 6.508189484301038e-05, "loss": 4.0823, "step": 803750 }, { "epoch": 4.84678187672604, "grad_norm": 2.7626373767852783, "learning_rate": 6.506512151732386e-05, "loss": 4.8271, "step": 803800 }, { "epoch": 4.847083368507374, "grad_norm": 2.6650967597961426, "learning_rate": 6.504834975474255e-05, "loss": 4.9682, "step": 803850 }, { "epoch": 4.847384860288709, "grad_norm": 2.8757808208465576, "learning_rate": 6.503157955557507e-05, "loss": 4.6971, "step": 803900 }, { "epoch": 4.847686352070043, "grad_norm": 2.9206008911132812, "learning_rate": 6.50148109201302e-05, "loss": 4.343, "step": 803950 }, { "epoch": 4.847987843851376, "grad_norm": 4.992050647735596, "learning_rate": 6.499804384871637e-05, "loss": 5.0955, "step": 804000 }, { "epoch": 4.847987843851376, "eval_loss": 5.0724945068359375, "eval_runtime": 39.1191, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 7.720305632314427, "step": 804000 }, { "epoch": 4.84828933563271, "grad_norm": 3.4565341472625732, "learning_rate": 6.498127834164226e-05, "loss": 4.632, "step": 804050 }, { "epoch": 4.848590827414045, "grad_norm": 3.0727949142456055, "learning_rate": 6.496451439921643e-05, "loss": 4.8135, "step": 804100 }, { "epoch": 4.848892319195379, "grad_norm": 2.283923625946045, "learning_rate": 6.494775202174732e-05, "loss": 4.7457, "step": 804150 }, { "epoch": 4.849193810976713, "grad_norm": 6.801211833953857, "learning_rate": 6.493099120954351e-05, "loss": 4.2398, "step": 804200 }, { "epoch": 4.849495302758047, "grad_norm": 3.2383456230163574, "learning_rate": 6.49142319629134e-05, "loss": 4.5104, "step": 804250 }, { "epoch": 4.8497967945393805, "grad_norm": 3.205531597137451, "learning_rate": 6.489747428216538e-05, "loss": 4.7386, "step": 804300 }, { "epoch": 4.850098286320715, "grad_norm": 4.0913615226745605, "learning_rate": 6.488071816760789e-05, "loss": 4.5317, "step": 804350 }, { "epoch": 4.850399778102049, "grad_norm": 2.696943759918213, "learning_rate": 6.486396361954937e-05, "loss": 4.9347, "step": 804400 }, { "epoch": 4.850701269883383, "grad_norm": 3.065957546234131, "learning_rate": 6.484721063829805e-05, "loss": 4.7477, "step": 804450 }, { "epoch": 4.851002761664717, "grad_norm": 3.895109176635742, "learning_rate": 6.483045922416229e-05, "loss": 4.6231, "step": 804500 }, { "epoch": 4.851304253446051, "grad_norm": 6.246702194213867, "learning_rate": 6.481370937745045e-05, "loss": 4.1638, "step": 804550 }, { "epoch": 4.851605745227385, "grad_norm": 2.7059109210968018, "learning_rate": 6.479696109847067e-05, "loss": 4.6219, "step": 804600 }, { "epoch": 4.851907237008719, "grad_norm": 2.768915891647339, "learning_rate": 6.478021438753126e-05, "loss": 4.5319, "step": 804650 }, { "epoch": 4.852208728790053, "grad_norm": 4.630606174468994, "learning_rate": 6.476346924494042e-05, "loss": 4.6309, "step": 804700 }, { "epoch": 4.8525102205713875, "grad_norm": 2.4206292629241943, "learning_rate": 6.474672567100622e-05, "loss": 4.4602, "step": 804750 }, { "epoch": 4.852811712352722, "grad_norm": 3.524442195892334, "learning_rate": 6.472998366603687e-05, "loss": 4.7477, "step": 804800 }, { "epoch": 4.853113204134055, "grad_norm": 2.5268075466156006, "learning_rate": 6.471324323034055e-05, "loss": 4.3354, "step": 804850 }, { "epoch": 4.853414695915389, "grad_norm": 3.6994950771331787, "learning_rate": 6.469650436422521e-05, "loss": 5.0379, "step": 804900 }, { "epoch": 4.853716187696723, "grad_norm": 3.472665786743164, "learning_rate": 6.467976706799897e-05, "loss": 4.568, "step": 804950 }, { "epoch": 4.854017679478058, "grad_norm": 0.6321718692779541, "learning_rate": 6.466303134196993e-05, "loss": 4.4641, "step": 805000 }, { "epoch": 4.854319171259392, "grad_norm": 3.131855010986328, "learning_rate": 6.464629718644599e-05, "loss": 4.7086, "step": 805050 }, { "epoch": 4.854620663040725, "grad_norm": 4.613230228424072, "learning_rate": 6.462956460173511e-05, "loss": 4.505, "step": 805100 }, { "epoch": 4.854922154822059, "grad_norm": 5.167689323425293, "learning_rate": 6.461283358814531e-05, "loss": 4.7296, "step": 805150 }, { "epoch": 4.855223646603394, "grad_norm": 0.8075562119483948, "learning_rate": 6.45961041459844e-05, "loss": 4.1474, "step": 805200 }, { "epoch": 4.855525138384728, "grad_norm": 4.269379138946533, "learning_rate": 6.457937627556033e-05, "loss": 4.5542, "step": 805250 }, { "epoch": 4.855826630166062, "grad_norm": 3.561732053756714, "learning_rate": 6.456264997718097e-05, "loss": 4.7637, "step": 805300 }, { "epoch": 4.856128121947396, "grad_norm": 2.107436180114746, "learning_rate": 6.454592525115407e-05, "loss": 4.7687, "step": 805350 }, { "epoch": 4.8564296137287295, "grad_norm": 2.6876158714294434, "learning_rate": 6.452920209778751e-05, "loss": 4.2036, "step": 805400 }, { "epoch": 4.856731105510064, "grad_norm": 1.0533039569854736, "learning_rate": 6.451248051738903e-05, "loss": 4.5803, "step": 805450 }, { "epoch": 4.857032597291398, "grad_norm": 3.1239993572235107, "learning_rate": 6.449576051026628e-05, "loss": 4.9615, "step": 805500 }, { "epoch": 4.857334089072732, "grad_norm": 2.769984722137451, "learning_rate": 6.447904207672703e-05, "loss": 4.4482, "step": 805550 }, { "epoch": 4.857635580854065, "grad_norm": 2.5772454738616943, "learning_rate": 6.446232521707904e-05, "loss": 4.9083, "step": 805600 }, { "epoch": 4.8579370726354, "grad_norm": 2.9038918018341064, "learning_rate": 6.444560993162983e-05, "loss": 4.381, "step": 805650 }, { "epoch": 4.858238564416734, "grad_norm": 4.3792219161987305, "learning_rate": 6.442889622068706e-05, "loss": 4.6011, "step": 805700 }, { "epoch": 4.858540056198068, "grad_norm": 4.349776268005371, "learning_rate": 6.441218408455841e-05, "loss": 4.7798, "step": 805750 }, { "epoch": 4.858841547979402, "grad_norm": 3.3257577419281006, "learning_rate": 6.43954735235513e-05, "loss": 4.6629, "step": 805800 }, { "epoch": 4.8591430397607365, "grad_norm": 0.6719509959220886, "learning_rate": 6.43787645379734e-05, "loss": 5.0671, "step": 805850 }, { "epoch": 4.85944453154207, "grad_norm": 2.513599395751953, "learning_rate": 6.436205712813215e-05, "loss": 4.8787, "step": 805900 }, { "epoch": 4.859746023323404, "grad_norm": 3.6932358741760254, "learning_rate": 6.434535129433495e-05, "loss": 4.7116, "step": 805950 }, { "epoch": 4.860047515104738, "grad_norm": 3.624058246612549, "learning_rate": 6.432864703688936e-05, "loss": 4.6063, "step": 806000 }, { "epoch": 4.860349006886072, "grad_norm": 2.3892438411712646, "learning_rate": 6.43119443561028e-05, "loss": 4.7006, "step": 806050 }, { "epoch": 4.860650498667407, "grad_norm": 4.181216716766357, "learning_rate": 6.429524325228257e-05, "loss": 4.6406, "step": 806100 }, { "epoch": 4.86095199044874, "grad_norm": 3.672149419784546, "learning_rate": 6.427854372573608e-05, "loss": 4.4923, "step": 806150 }, { "epoch": 4.861253482230074, "grad_norm": 2.700127124786377, "learning_rate": 6.426184577677071e-05, "loss": 4.6651, "step": 806200 }, { "epoch": 4.861554974011408, "grad_norm": 5.235482692718506, "learning_rate": 6.424514940569371e-05, "loss": 4.4276, "step": 806250 }, { "epoch": 4.8618564657927426, "grad_norm": 2.257736921310425, "learning_rate": 6.422845461281232e-05, "loss": 4.8175, "step": 806300 }, { "epoch": 4.862157957574077, "grad_norm": 3.219653606414795, "learning_rate": 6.421176139843386e-05, "loss": 4.6912, "step": 806350 }, { "epoch": 4.862459449355411, "grad_norm": 3.791449785232544, "learning_rate": 6.419506976286545e-05, "loss": 4.9326, "step": 806400 }, { "epoch": 4.862760941136744, "grad_norm": 3.3937573432922363, "learning_rate": 6.417837970641434e-05, "loss": 4.5743, "step": 806450 }, { "epoch": 4.8630624329180785, "grad_norm": 4.473875045776367, "learning_rate": 6.416169122938772e-05, "loss": 4.0433, "step": 806500 }, { "epoch": 4.863363924699413, "grad_norm": 2.959425449371338, "learning_rate": 6.414500433209264e-05, "loss": 4.8274, "step": 806550 }, { "epoch": 4.863665416480747, "grad_norm": 3.1970977783203125, "learning_rate": 6.412831901483628e-05, "loss": 4.8031, "step": 806600 }, { "epoch": 4.863966908262081, "grad_norm": 2.527122735977173, "learning_rate": 6.411163527792565e-05, "loss": 4.6055, "step": 806650 }, { "epoch": 4.864268400043414, "grad_norm": 3.0714945793151855, "learning_rate": 6.409495312166776e-05, "loss": 4.9873, "step": 806700 }, { "epoch": 4.864569891824749, "grad_norm": 1.5959783792495728, "learning_rate": 6.407827254636966e-05, "loss": 4.5777, "step": 806750 }, { "epoch": 4.864871383606083, "grad_norm": 3.292440891265869, "learning_rate": 6.40615935523384e-05, "loss": 4.3591, "step": 806800 }, { "epoch": 4.865172875387417, "grad_norm": 2.556464672088623, "learning_rate": 6.404491613988081e-05, "loss": 4.2456, "step": 806850 }, { "epoch": 4.865474367168751, "grad_norm": 3.295017719268799, "learning_rate": 6.402824030930386e-05, "loss": 4.7602, "step": 806900 }, { "epoch": 4.8657758589500855, "grad_norm": 2.9941630363464355, "learning_rate": 6.401156606091449e-05, "loss": 4.802, "step": 806950 }, { "epoch": 4.866077350731419, "grad_norm": 4.965896129608154, "learning_rate": 6.399489339501957e-05, "loss": 4.5585, "step": 807000 }, { "epoch": 4.866077350731419, "eval_loss": 5.078523635864258, "eval_runtime": 38.8441, "eval_samples_per_second": 13.181, "eval_steps_per_second": 6.59, "eval_tts_loss": 7.7521768042723185, "step": 807000 }, { "epoch": 4.866378842512753, "grad_norm": 3.8055663108825684, "learning_rate": 6.39782223119259e-05, "loss": 4.8789, "step": 807050 }, { "epoch": 4.866680334294087, "grad_norm": 2.7853140830993652, "learning_rate": 6.396155281194022e-05, "loss": 4.5641, "step": 807100 }, { "epoch": 4.866981826075421, "grad_norm": 2.731903553009033, "learning_rate": 6.394488489536947e-05, "loss": 4.6826, "step": 807150 }, { "epoch": 4.867283317856756, "grad_norm": 2.513788938522339, "learning_rate": 6.392821856252022e-05, "loss": 4.7801, "step": 807200 }, { "epoch": 4.867584809638089, "grad_norm": 1.647912621498108, "learning_rate": 6.391155381369929e-05, "loss": 4.7464, "step": 807250 }, { "epoch": 4.867886301419423, "grad_norm": 3.3948216438293457, "learning_rate": 6.38948906492134e-05, "loss": 4.8048, "step": 807300 }, { "epoch": 4.868187793200757, "grad_norm": 2.8496875762939453, "learning_rate": 6.387822906936913e-05, "loss": 4.7987, "step": 807350 }, { "epoch": 4.8684892849820915, "grad_norm": 3.1567320823669434, "learning_rate": 6.38615690744732e-05, "loss": 5.044, "step": 807400 }, { "epoch": 4.868790776763426, "grad_norm": 2.7212908267974854, "learning_rate": 6.384491066483217e-05, "loss": 4.5281, "step": 807450 }, { "epoch": 4.86909226854476, "grad_norm": 3.1460256576538086, "learning_rate": 6.382825384075255e-05, "loss": 4.7389, "step": 807500 }, { "epoch": 4.869393760326093, "grad_norm": 4.229955196380615, "learning_rate": 6.381159860254094e-05, "loss": 4.5074, "step": 807550 }, { "epoch": 4.8696952521074275, "grad_norm": 4.012442111968994, "learning_rate": 6.379494495050392e-05, "loss": 4.1517, "step": 807600 }, { "epoch": 4.869996743888762, "grad_norm": 2.820300579071045, "learning_rate": 6.377829288494785e-05, "loss": 4.4985, "step": 807650 }, { "epoch": 4.870298235670096, "grad_norm": 3.1815719604492188, "learning_rate": 6.376164240617928e-05, "loss": 5.1051, "step": 807700 }, { "epoch": 4.870599727451429, "grad_norm": 1.4215720891952515, "learning_rate": 6.374499351450464e-05, "loss": 4.7252, "step": 807750 }, { "epoch": 4.870901219232763, "grad_norm": 3.4315319061279297, "learning_rate": 6.372834621023031e-05, "loss": 4.7333, "step": 807800 }, { "epoch": 4.871202711014098, "grad_norm": 2.7852537631988525, "learning_rate": 6.37117004936626e-05, "loss": 4.3764, "step": 807850 }, { "epoch": 4.871504202795432, "grad_norm": 5.5910325050354, "learning_rate": 6.369505636510797e-05, "loss": 4.5268, "step": 807900 }, { "epoch": 4.871805694576766, "grad_norm": 4.886362552642822, "learning_rate": 6.36784138248726e-05, "loss": 4.5113, "step": 807950 }, { "epoch": 4.8721071863581, "grad_norm": 3.7029268741607666, "learning_rate": 6.366177287326282e-05, "loss": 4.6162, "step": 808000 }, { "epoch": 4.872408678139434, "grad_norm": 2.694312334060669, "learning_rate": 6.364513351058497e-05, "loss": 4.486, "step": 808050 }, { "epoch": 4.872710169920768, "grad_norm": 2.908230781555176, "learning_rate": 6.362849573714515e-05, "loss": 4.9263, "step": 808100 }, { "epoch": 4.873011661702102, "grad_norm": 2.597663640975952, "learning_rate": 6.36118595532496e-05, "loss": 4.8409, "step": 808150 }, { "epoch": 4.873313153483436, "grad_norm": 3.0881147384643555, "learning_rate": 6.359522495920455e-05, "loss": 4.5984, "step": 808200 }, { "epoch": 4.87361464526477, "grad_norm": 2.318361520767212, "learning_rate": 6.357859195531607e-05, "loss": 4.6032, "step": 808250 }, { "epoch": 4.873916137046104, "grad_norm": 2.6375954151153564, "learning_rate": 6.356196054189023e-05, "loss": 4.9495, "step": 808300 }, { "epoch": 4.874217628827438, "grad_norm": 4.052475929260254, "learning_rate": 6.354533071923321e-05, "loss": 4.0295, "step": 808350 }, { "epoch": 4.874519120608772, "grad_norm": 3.937215805053711, "learning_rate": 6.352870248765094e-05, "loss": 4.6368, "step": 808400 }, { "epoch": 4.874820612390106, "grad_norm": 1.9053560495376587, "learning_rate": 6.35120758474495e-05, "loss": 4.5832, "step": 808450 }, { "epoch": 4.8751221041714405, "grad_norm": 3.044999122619629, "learning_rate": 6.349545079893494e-05, "loss": 4.7045, "step": 808500 }, { "epoch": 4.875423595952775, "grad_norm": 2.6277732849121094, "learning_rate": 6.34788273424131e-05, "loss": 4.8971, "step": 808550 }, { "epoch": 4.875725087734108, "grad_norm": 2.9161038398742676, "learning_rate": 6.346220547819003e-05, "loss": 4.4324, "step": 808600 }, { "epoch": 4.876026579515442, "grad_norm": 5.133168697357178, "learning_rate": 6.344558520657156e-05, "loss": 4.7943, "step": 808650 }, { "epoch": 4.8763280712967765, "grad_norm": 4.4924421310424805, "learning_rate": 6.342896652786351e-05, "loss": 4.3685, "step": 808700 }, { "epoch": 4.876629563078111, "grad_norm": 3.280210256576538, "learning_rate": 6.341234944237183e-05, "loss": 4.5468, "step": 808750 }, { "epoch": 4.876931054859445, "grad_norm": 2.9889631271362305, "learning_rate": 6.339573395040231e-05, "loss": 4.4701, "step": 808800 }, { "epoch": 4.877232546640778, "grad_norm": 2.919196128845215, "learning_rate": 6.337912005226067e-05, "loss": 4.4418, "step": 808850 }, { "epoch": 4.877534038422112, "grad_norm": 4.112672328948975, "learning_rate": 6.336250774825271e-05, "loss": 4.8947, "step": 808900 }, { "epoch": 4.877835530203447, "grad_norm": 3.6091010570526123, "learning_rate": 6.334589703868422e-05, "loss": 4.5756, "step": 808950 }, { "epoch": 4.878137021984781, "grad_norm": 2.603792428970337, "learning_rate": 6.332928792386075e-05, "loss": 4.6279, "step": 809000 }, { "epoch": 4.878438513766115, "grad_norm": 3.085418939590454, "learning_rate": 6.331268040408815e-05, "loss": 4.7153, "step": 809050 }, { "epoch": 4.878740005547449, "grad_norm": 3.214644193649292, "learning_rate": 6.329607447967193e-05, "loss": 4.83, "step": 809100 }, { "epoch": 4.879041497328783, "grad_norm": 4.308750152587891, "learning_rate": 6.327947015091768e-05, "loss": 4.9487, "step": 809150 }, { "epoch": 4.879342989110117, "grad_norm": 1.5676623582839966, "learning_rate": 6.326286741813101e-05, "loss": 4.8382, "step": 809200 }, { "epoch": 4.879644480891451, "grad_norm": 2.9074556827545166, "learning_rate": 6.324626628161755e-05, "loss": 4.5271, "step": 809250 }, { "epoch": 4.879945972672785, "grad_norm": 3.322251558303833, "learning_rate": 6.32296667416827e-05, "loss": 4.7267, "step": 809300 }, { "epoch": 4.8802474644541185, "grad_norm": 3.7634425163269043, "learning_rate": 6.3213068798632e-05, "loss": 4.7767, "step": 809350 }, { "epoch": 4.880548956235453, "grad_norm": 2.3875935077667236, "learning_rate": 6.319647245277099e-05, "loss": 4.4289, "step": 809400 }, { "epoch": 4.880850448016787, "grad_norm": 2.736386775970459, "learning_rate": 6.317987770440502e-05, "loss": 4.6548, "step": 809450 }, { "epoch": 4.881151939798121, "grad_norm": 3.529503345489502, "learning_rate": 6.316328455383944e-05, "loss": 4.7978, "step": 809500 }, { "epoch": 4.881453431579455, "grad_norm": 3.089984893798828, "learning_rate": 6.314669300137973e-05, "loss": 4.604, "step": 809550 }, { "epoch": 4.8817549233607895, "grad_norm": 4.148348331451416, "learning_rate": 6.313010304733115e-05, "loss": 4.6219, "step": 809600 }, { "epoch": 4.882056415142123, "grad_norm": 4.420154094696045, "learning_rate": 6.311351469199907e-05, "loss": 4.6335, "step": 809650 }, { "epoch": 4.882357906923457, "grad_norm": 4.892827033996582, "learning_rate": 6.309692793568879e-05, "loss": 4.4829, "step": 809700 }, { "epoch": 4.882659398704791, "grad_norm": 2.8493497371673584, "learning_rate": 6.308034277870548e-05, "loss": 4.7901, "step": 809750 }, { "epoch": 4.8829608904861255, "grad_norm": 2.6667284965515137, "learning_rate": 6.306375922135448e-05, "loss": 4.9816, "step": 809800 }, { "epoch": 4.88326238226746, "grad_norm": 3.429137706756592, "learning_rate": 6.304717726394092e-05, "loss": 4.3089, "step": 809850 }, { "epoch": 4.883563874048793, "grad_norm": 2.8382461071014404, "learning_rate": 6.303059690676993e-05, "loss": 4.6451, "step": 809900 }, { "epoch": 4.883865365830127, "grad_norm": 3.424258232116699, "learning_rate": 6.301401815014667e-05, "loss": 4.846, "step": 809950 }, { "epoch": 4.884166857611461, "grad_norm": 3.6443331241607666, "learning_rate": 6.299744099437634e-05, "loss": 4.8597, "step": 810000 }, { "epoch": 4.884166857611461, "eval_loss": 5.068855285644531, "eval_runtime": 39.0049, "eval_samples_per_second": 13.127, "eval_steps_per_second": 6.563, "eval_tts_loss": 7.745122216952296, "step": 810000 }, { "epoch": 4.884468349392796, "grad_norm": 2.336230516433716, "learning_rate": 6.29808654397639e-05, "loss": 4.5579, "step": 810050 }, { "epoch": 4.88476984117413, "grad_norm": 2.613520860671997, "learning_rate": 6.296429148661446e-05, "loss": 4.8185, "step": 810100 }, { "epoch": 4.885071332955464, "grad_norm": 3.189333200454712, "learning_rate": 6.294771913523307e-05, "loss": 4.8143, "step": 810150 }, { "epoch": 4.885372824736797, "grad_norm": 1.3550134897232056, "learning_rate": 6.293114838592461e-05, "loss": 4.458, "step": 810200 }, { "epoch": 4.8856743165181316, "grad_norm": 2.911945343017578, "learning_rate": 6.291457923899419e-05, "loss": 4.6067, "step": 810250 }, { "epoch": 4.885975808299466, "grad_norm": 2.568514823913574, "learning_rate": 6.289801169474659e-05, "loss": 4.5966, "step": 810300 }, { "epoch": 4.8862773000808, "grad_norm": 1.3556054830551147, "learning_rate": 6.288144575348684e-05, "loss": 4.16, "step": 810350 }, { "epoch": 4.886578791862134, "grad_norm": 2.3680009841918945, "learning_rate": 6.28648814155197e-05, "loss": 4.376, "step": 810400 }, { "epoch": 4.8868802836434675, "grad_norm": 2.319652795791626, "learning_rate": 6.284831868115008e-05, "loss": 4.5155, "step": 810450 }, { "epoch": 4.887181775424802, "grad_norm": 2.9542951583862305, "learning_rate": 6.283175755068282e-05, "loss": 4.3311, "step": 810500 }, { "epoch": 4.887483267206136, "grad_norm": 2.561000347137451, "learning_rate": 6.281519802442263e-05, "loss": 4.7877, "step": 810550 }, { "epoch": 4.88778475898747, "grad_norm": 2.4243059158325195, "learning_rate": 6.279864010267435e-05, "loss": 4.9472, "step": 810600 }, { "epoch": 4.888086250768804, "grad_norm": 1.46841299533844, "learning_rate": 6.278208378574265e-05, "loss": 4.8651, "step": 810650 }, { "epoch": 4.8883877425501385, "grad_norm": 2.7059810161590576, "learning_rate": 6.276552907393218e-05, "loss": 4.6609, "step": 810700 }, { "epoch": 4.888689234331472, "grad_norm": 3.6642496585845947, "learning_rate": 6.274897596754765e-05, "loss": 4.8549, "step": 810750 }, { "epoch": 4.888990726112806, "grad_norm": 3.036468029022217, "learning_rate": 6.273242446689377e-05, "loss": 4.5554, "step": 810800 }, { "epoch": 4.88929221789414, "grad_norm": 4.00445556640625, "learning_rate": 6.271587457227502e-05, "loss": 4.8991, "step": 810850 }, { "epoch": 4.8895937096754745, "grad_norm": 4.6957526206970215, "learning_rate": 6.269932628399605e-05, "loss": 4.4264, "step": 810900 }, { "epoch": 4.889895201456809, "grad_norm": 2.748530864715576, "learning_rate": 6.268277960236143e-05, "loss": 4.4575, "step": 810950 }, { "epoch": 4.890196693238142, "grad_norm": 1.7213987112045288, "learning_rate": 6.266623452767564e-05, "loss": 4.7099, "step": 811000 }, { "epoch": 4.890498185019476, "grad_norm": 3.0918338298797607, "learning_rate": 6.264969106024313e-05, "loss": 4.4792, "step": 811050 }, { "epoch": 4.89079967680081, "grad_norm": 5.998373508453369, "learning_rate": 6.263314920036846e-05, "loss": 4.5638, "step": 811100 }, { "epoch": 4.891101168582145, "grad_norm": 2.878892660140991, "learning_rate": 6.261660894835594e-05, "loss": 4.6071, "step": 811150 }, { "epoch": 4.891402660363479, "grad_norm": 2.8078503608703613, "learning_rate": 6.260007030451002e-05, "loss": 4.6933, "step": 811200 }, { "epoch": 4.891704152144813, "grad_norm": 2.4935028553009033, "learning_rate": 6.258353326913516e-05, "loss": 4.7152, "step": 811250 }, { "epoch": 4.892005643926146, "grad_norm": 2.791367292404175, "learning_rate": 6.256699784253554e-05, "loss": 4.6505, "step": 811300 }, { "epoch": 4.8923071357074805, "grad_norm": 3.8161330223083496, "learning_rate": 6.255046402501554e-05, "loss": 4.7705, "step": 811350 }, { "epoch": 4.892608627488815, "grad_norm": 2.838951826095581, "learning_rate": 6.253393181687955e-05, "loss": 4.5151, "step": 811400 }, { "epoch": 4.892910119270149, "grad_norm": 2.57362961769104, "learning_rate": 6.251740121843167e-05, "loss": 4.5604, "step": 811450 }, { "epoch": 4.893211611051482, "grad_norm": 2.9433741569519043, "learning_rate": 6.250087222997617e-05, "loss": 4.6022, "step": 811500 }, { "epoch": 4.8935131028328165, "grad_norm": 2.8525774478912354, "learning_rate": 6.248434485181729e-05, "loss": 4.8557, "step": 811550 }, { "epoch": 4.893814594614151, "grad_norm": 2.8472015857696533, "learning_rate": 6.246781908425909e-05, "loss": 4.3772, "step": 811600 }, { "epoch": 4.894116086395485, "grad_norm": 2.512392997741699, "learning_rate": 6.245129492760575e-05, "loss": 4.5735, "step": 811650 }, { "epoch": 4.894417578176819, "grad_norm": 3.59382963180542, "learning_rate": 6.243477238216146e-05, "loss": 4.6583, "step": 811700 }, { "epoch": 4.894719069958153, "grad_norm": 3.7622487545013428, "learning_rate": 6.241825144823017e-05, "loss": 4.5179, "step": 811750 }, { "epoch": 4.895020561739487, "grad_norm": 1.169249176979065, "learning_rate": 6.240173212611604e-05, "loss": 4.754, "step": 811800 }, { "epoch": 4.895322053520821, "grad_norm": 4.46796178817749, "learning_rate": 6.238521441612299e-05, "loss": 4.905, "step": 811850 }, { "epoch": 4.895623545302155, "grad_norm": 2.709019422531128, "learning_rate": 6.236869831855501e-05, "loss": 4.7697, "step": 811900 }, { "epoch": 4.895925037083489, "grad_norm": 2.497629404067993, "learning_rate": 6.235218383371607e-05, "loss": 4.712, "step": 811950 }, { "epoch": 4.8962265288648235, "grad_norm": 3.036738634109497, "learning_rate": 6.233567096191016e-05, "loss": 4.7098, "step": 812000 }, { "epoch": 4.896528020646157, "grad_norm": 4.079648971557617, "learning_rate": 6.231915970344108e-05, "loss": 4.6421, "step": 812050 }, { "epoch": 4.896829512427491, "grad_norm": 3.141226053237915, "learning_rate": 6.230265005861271e-05, "loss": 4.6095, "step": 812100 }, { "epoch": 4.897131004208825, "grad_norm": 3.0302162170410156, "learning_rate": 6.228614202772898e-05, "loss": 4.5932, "step": 812150 }, { "epoch": 4.897432495990159, "grad_norm": 3.649251699447632, "learning_rate": 6.226963561109365e-05, "loss": 5.1885, "step": 812200 }, { "epoch": 4.897733987771494, "grad_norm": 2.8555827140808105, "learning_rate": 6.225313080901041e-05, "loss": 4.6739, "step": 812250 }, { "epoch": 4.898035479552828, "grad_norm": 5.801814079284668, "learning_rate": 6.223662762178314e-05, "loss": 4.6386, "step": 812300 }, { "epoch": 4.898336971334161, "grad_norm": 2.1116862297058105, "learning_rate": 6.222012604971542e-05, "loss": 5.1002, "step": 812350 }, { "epoch": 4.898638463115495, "grad_norm": 2.899340867996216, "learning_rate": 6.220362609311103e-05, "loss": 4.709, "step": 812400 }, { "epoch": 4.8989399548968295, "grad_norm": 2.9486634731292725, "learning_rate": 6.218712775227365e-05, "loss": 4.9078, "step": 812450 }, { "epoch": 4.899241446678164, "grad_norm": 3.3914310932159424, "learning_rate": 6.217063102750683e-05, "loss": 4.5996, "step": 812500 }, { "epoch": 4.899542938459498, "grad_norm": 2.3933448791503906, "learning_rate": 6.215413591911418e-05, "loss": 4.9738, "step": 812550 }, { "epoch": 4.899844430240831, "grad_norm": 3.3707435131073, "learning_rate": 6.213764242739937e-05, "loss": 4.7498, "step": 812600 }, { "epoch": 4.9001459220221655, "grad_norm": 1.3433727025985718, "learning_rate": 6.212115055266587e-05, "loss": 4.688, "step": 812650 }, { "epoch": 4.9004474138035, "grad_norm": 2.650696039199829, "learning_rate": 6.210466029521712e-05, "loss": 4.542, "step": 812700 }, { "epoch": 4.900748905584834, "grad_norm": 6.593814849853516, "learning_rate": 6.208817165535671e-05, "loss": 4.4815, "step": 812750 }, { "epoch": 4.901050397366168, "grad_norm": 3.943861246109009, "learning_rate": 6.207168463338801e-05, "loss": 4.6025, "step": 812800 }, { "epoch": 4.901351889147502, "grad_norm": 3.2961292266845703, "learning_rate": 6.205519922961446e-05, "loss": 4.9747, "step": 812850 }, { "epoch": 4.901653380928836, "grad_norm": 5.470357894897461, "learning_rate": 6.203871544433952e-05, "loss": 4.5488, "step": 812900 }, { "epoch": 4.90195487271017, "grad_norm": 4.224762439727783, "learning_rate": 6.202223327786645e-05, "loss": 4.2763, "step": 812950 }, { "epoch": 4.902256364491504, "grad_norm": 3.3018882274627686, "learning_rate": 6.200575273049866e-05, "loss": 4.178, "step": 813000 }, { "epoch": 4.902256364491504, "eval_loss": 5.062519073486328, "eval_runtime": 38.9239, "eval_samples_per_second": 13.154, "eval_steps_per_second": 6.577, "eval_tts_loss": 7.711300264173559, "step": 813000 }, { "epoch": 4.902557856272838, "grad_norm": 4.296719551086426, "learning_rate": 6.198927380253943e-05, "loss": 4.035, "step": 813050 }, { "epoch": 4.902859348054172, "grad_norm": 2.479792356491089, "learning_rate": 6.197279649429197e-05, "loss": 4.8905, "step": 813100 }, { "epoch": 4.903160839835506, "grad_norm": 2.6728203296661377, "learning_rate": 6.195632080605956e-05, "loss": 4.6816, "step": 813150 }, { "epoch": 4.90346233161684, "grad_norm": 1.2089461088180542, "learning_rate": 6.193984673814548e-05, "loss": 4.2722, "step": 813200 }, { "epoch": 4.903763823398174, "grad_norm": 3.382545232772827, "learning_rate": 6.192337429085278e-05, "loss": 4.7204, "step": 813250 }, { "epoch": 4.904065315179508, "grad_norm": 3.802640438079834, "learning_rate": 6.19069034644847e-05, "loss": 4.3409, "step": 813300 }, { "epoch": 4.904366806960843, "grad_norm": 1.1311160326004028, "learning_rate": 6.18904342593444e-05, "loss": 5.1036, "step": 813350 }, { "epoch": 4.904668298742176, "grad_norm": 3.9670915603637695, "learning_rate": 6.187396667573492e-05, "loss": 4.5439, "step": 813400 }, { "epoch": 4.90496979052351, "grad_norm": 2.668164014816284, "learning_rate": 6.185750071395927e-05, "loss": 4.8274, "step": 813450 }, { "epoch": 4.905271282304844, "grad_norm": 1.452697515487671, "learning_rate": 6.184103637432057e-05, "loss": 4.2914, "step": 813500 }, { "epoch": 4.9055727740861785, "grad_norm": 4.223587989807129, "learning_rate": 6.182457365712175e-05, "loss": 4.942, "step": 813550 }, { "epoch": 4.905874265867513, "grad_norm": 3.274030923843384, "learning_rate": 6.18081125626658e-05, "loss": 4.9711, "step": 813600 }, { "epoch": 4.906175757648846, "grad_norm": 3.855772018432617, "learning_rate": 6.17916530912557e-05, "loss": 4.7229, "step": 813650 }, { "epoch": 4.90647724943018, "grad_norm": 4.026672840118408, "learning_rate": 6.177519524319438e-05, "loss": 4.4442, "step": 813700 }, { "epoch": 4.9067787412115145, "grad_norm": 1.5917305946350098, "learning_rate": 6.175873901878466e-05, "loss": 4.4781, "step": 813750 }, { "epoch": 4.907080232992849, "grad_norm": 4.231906890869141, "learning_rate": 6.174228441832946e-05, "loss": 4.734, "step": 813800 }, { "epoch": 4.907381724774183, "grad_norm": 2.4891648292541504, "learning_rate": 6.172583144213156e-05, "loss": 4.6839, "step": 813850 }, { "epoch": 4.907683216555517, "grad_norm": 5.825902938842773, "learning_rate": 6.170938009049372e-05, "loss": 4.6219, "step": 813900 }, { "epoch": 4.90798470833685, "grad_norm": 1.5042027235031128, "learning_rate": 6.169293036371873e-05, "loss": 4.4648, "step": 813950 }, { "epoch": 4.908286200118185, "grad_norm": 2.685546636581421, "learning_rate": 6.167648226210938e-05, "loss": 4.3822, "step": 814000 }, { "epoch": 4.908587691899519, "grad_norm": 3.6569013595581055, "learning_rate": 6.166003578596828e-05, "loss": 4.5377, "step": 814050 }, { "epoch": 4.908889183680853, "grad_norm": 2.5909931659698486, "learning_rate": 6.164359093559818e-05, "loss": 4.4707, "step": 814100 }, { "epoch": 4.909190675462187, "grad_norm": 4.405144214630127, "learning_rate": 6.162714771130172e-05, "loss": 4.6909, "step": 814150 }, { "epoch": 4.9094921672435206, "grad_norm": 1.2309811115264893, "learning_rate": 6.16107061133815e-05, "loss": 4.3655, "step": 814200 }, { "epoch": 4.909793659024855, "grad_norm": 3.5960779190063477, "learning_rate": 6.159426614214004e-05, "loss": 4.4853, "step": 814250 }, { "epoch": 4.910095150806189, "grad_norm": 4.417572021484375, "learning_rate": 6.157782779788002e-05, "loss": 4.7413, "step": 814300 }, { "epoch": 4.910396642587523, "grad_norm": 3.6589434146881104, "learning_rate": 6.156139108090383e-05, "loss": 4.7231, "step": 814350 }, { "epoch": 4.910698134368857, "grad_norm": 3.3012068271636963, "learning_rate": 6.154495599151402e-05, "loss": 4.0574, "step": 814400 }, { "epoch": 4.910999626150192, "grad_norm": 2.9079818725585938, "learning_rate": 6.152852253001313e-05, "loss": 4.9457, "step": 814450 }, { "epoch": 4.911301117931525, "grad_norm": 2.928053855895996, "learning_rate": 6.151209069670347e-05, "loss": 4.7021, "step": 814500 }, { "epoch": 4.911602609712859, "grad_norm": 2.885573387145996, "learning_rate": 6.149566049188755e-05, "loss": 4.3715, "step": 814550 }, { "epoch": 4.911904101494193, "grad_norm": 2.9970834255218506, "learning_rate": 6.147923191586764e-05, "loss": 4.6847, "step": 814600 }, { "epoch": 4.9122055932755275, "grad_norm": 2.8180935382843018, "learning_rate": 6.14628049689462e-05, "loss": 4.3982, "step": 814650 }, { "epoch": 4.912507085056862, "grad_norm": 1.088366150856018, "learning_rate": 6.144637965142544e-05, "loss": 4.9397, "step": 814700 }, { "epoch": 4.912808576838195, "grad_norm": 3.0890228748321533, "learning_rate": 6.142995596360774e-05, "loss": 4.5155, "step": 814750 }, { "epoch": 4.913110068619529, "grad_norm": 1.6458441019058228, "learning_rate": 6.141353390579524e-05, "loss": 4.0343, "step": 814800 }, { "epoch": 4.9134115604008635, "grad_norm": 3.7392561435699463, "learning_rate": 6.139711347829024e-05, "loss": 4.5976, "step": 814850 }, { "epoch": 4.913713052182198, "grad_norm": 4.429773330688477, "learning_rate": 6.138069468139497e-05, "loss": 4.622, "step": 814900 }, { "epoch": 4.914014543963532, "grad_norm": 2.5052242279052734, "learning_rate": 6.136427751541149e-05, "loss": 4.6098, "step": 814950 }, { "epoch": 4.914316035744866, "grad_norm": 3.3560056686401367, "learning_rate": 6.134786198064205e-05, "loss": 4.776, "step": 815000 }, { "epoch": 4.914617527526199, "grad_norm": 3.931778907775879, "learning_rate": 6.13314480773887e-05, "loss": 4.5903, "step": 815050 }, { "epoch": 4.914919019307534, "grad_norm": 3.776401996612549, "learning_rate": 6.131503580595345e-05, "loss": 4.4622, "step": 815100 }, { "epoch": 4.915220511088868, "grad_norm": 2.2008209228515625, "learning_rate": 6.129862516663841e-05, "loss": 4.8771, "step": 815150 }, { "epoch": 4.915522002870202, "grad_norm": 3.867872714996338, "learning_rate": 6.128221615974566e-05, "loss": 4.471, "step": 815200 }, { "epoch": 4.915823494651535, "grad_norm": 1.7036868333816528, "learning_rate": 6.126580878557703e-05, "loss": 4.5035, "step": 815250 }, { "epoch": 4.9161249864328695, "grad_norm": 3.472707986831665, "learning_rate": 6.124940304443457e-05, "loss": 4.9156, "step": 815300 }, { "epoch": 4.916426478214204, "grad_norm": 4.452794551849365, "learning_rate": 6.123299893662027e-05, "loss": 4.6169, "step": 815350 }, { "epoch": 4.916727969995538, "grad_norm": 1.716533899307251, "learning_rate": 6.121659646243592e-05, "loss": 4.5112, "step": 815400 }, { "epoch": 4.917029461776872, "grad_norm": 2.984325408935547, "learning_rate": 6.120019562218337e-05, "loss": 4.7886, "step": 815450 }, { "epoch": 4.917330953558206, "grad_norm": 4.5315327644348145, "learning_rate": 6.118379641616456e-05, "loss": 4.6803, "step": 815500 }, { "epoch": 4.91763244533954, "grad_norm": 2.617065906524658, "learning_rate": 6.116739884468115e-05, "loss": 4.485, "step": 815550 }, { "epoch": 4.917933937120874, "grad_norm": 4.090301990509033, "learning_rate": 6.1151002908035e-05, "loss": 4.4267, "step": 815600 }, { "epoch": 4.918235428902208, "grad_norm": 3.820784568786621, "learning_rate": 6.11346086065279e-05, "loss": 4.5183, "step": 815650 }, { "epoch": 4.918536920683542, "grad_norm": 3.0431771278381348, "learning_rate": 6.111821594046147e-05, "loss": 4.4773, "step": 815700 }, { "epoch": 4.9188384124648765, "grad_norm": 3.265681743621826, "learning_rate": 6.110182491013743e-05, "loss": 4.6592, "step": 815750 }, { "epoch": 4.91913990424621, "grad_norm": 4.548947334289551, "learning_rate": 6.108543551585748e-05, "loss": 5.0231, "step": 815800 }, { "epoch": 4.919441396027544, "grad_norm": 3.8417012691497803, "learning_rate": 6.106904775792322e-05, "loss": 4.4749, "step": 815850 }, { "epoch": 4.919742887808878, "grad_norm": 2.536691904067993, "learning_rate": 6.105266163663617e-05, "loss": 4.898, "step": 815900 }, { "epoch": 4.9200443795902125, "grad_norm": 4.551534652709961, "learning_rate": 6.103627715229798e-05, "loss": 4.2983, "step": 815950 }, { "epoch": 4.920345871371547, "grad_norm": 2.408308744430542, "learning_rate": 6.101989430521011e-05, "loss": 4.5712, "step": 816000 }, { "epoch": 4.920345871371547, "eval_loss": 5.06759786605835, "eval_runtime": 39.2715, "eval_samples_per_second": 13.037, "eval_steps_per_second": 6.519, "eval_tts_loss": 7.713826580781168, "step": 816000 }, { "epoch": 4.920647363152881, "grad_norm": 2.3705437183380127, "learning_rate": 6.100351309567412e-05, "loss": 4.6061, "step": 816050 }, { "epoch": 4.920948854934214, "grad_norm": 3.056262969970703, "learning_rate": 6.09871335239915e-05, "loss": 4.8763, "step": 816100 }, { "epoch": 4.921250346715548, "grad_norm": 2.6530301570892334, "learning_rate": 6.0970755590463604e-05, "loss": 4.8413, "step": 816150 }, { "epoch": 4.921551838496883, "grad_norm": 2.5618374347686768, "learning_rate": 6.095437929539196e-05, "loss": 4.6481, "step": 816200 }, { "epoch": 4.921853330278217, "grad_norm": 4.594315052032471, "learning_rate": 6.0938004639077905e-05, "loss": 4.7865, "step": 816250 }, { "epoch": 4.922154822059551, "grad_norm": 2.5935914516448975, "learning_rate": 6.09216316218227e-05, "loss": 4.9988, "step": 816300 }, { "epoch": 4.922456313840884, "grad_norm": 3.0735392570495605, "learning_rate": 6.0905260243927754e-05, "loss": 4.3905, "step": 816350 }, { "epoch": 4.9227578056222185, "grad_norm": 2.1007614135742188, "learning_rate": 6.08888905056944e-05, "loss": 4.5013, "step": 816400 }, { "epoch": 4.923059297403553, "grad_norm": 3.4487884044647217, "learning_rate": 6.087252240742379e-05, "loss": 4.0935, "step": 816450 }, { "epoch": 4.923360789184887, "grad_norm": 3.011997699737549, "learning_rate": 6.085615594941723e-05, "loss": 4.5062, "step": 816500 }, { "epoch": 4.923662280966221, "grad_norm": 2.172049045562744, "learning_rate": 6.083979113197593e-05, "loss": 4.9581, "step": 816550 }, { "epoch": 4.923963772747555, "grad_norm": 3.0019569396972656, "learning_rate": 6.082342795540106e-05, "loss": 4.3137, "step": 816600 }, { "epoch": 4.924265264528889, "grad_norm": 2.504884958267212, "learning_rate": 6.080706641999366e-05, "loss": 4.9723, "step": 816650 }, { "epoch": 4.924566756310223, "grad_norm": 2.408205032348633, "learning_rate": 6.0790706526054986e-05, "loss": 4.7726, "step": 816700 }, { "epoch": 4.924868248091557, "grad_norm": 1.2153162956237793, "learning_rate": 6.0774348273886006e-05, "loss": 4.2332, "step": 816750 }, { "epoch": 4.925169739872891, "grad_norm": 3.4208731651306152, "learning_rate": 6.0757991663787797e-05, "loss": 4.6192, "step": 816800 }, { "epoch": 4.925471231654225, "grad_norm": 3.231394052505493, "learning_rate": 6.074163669606146e-05, "loss": 4.5799, "step": 816850 }, { "epoch": 4.925772723435559, "grad_norm": 3.890352487564087, "learning_rate": 6.072528337100786e-05, "loss": 4.8172, "step": 816900 }, { "epoch": 4.926074215216893, "grad_norm": 3.988525629043579, "learning_rate": 6.0708931688928016e-05, "loss": 5.1544, "step": 816950 }, { "epoch": 4.926375706998227, "grad_norm": 3.334676504135132, "learning_rate": 6.069258165012292e-05, "loss": 4.4711, "step": 817000 }, { "epoch": 4.9266771987795615, "grad_norm": 3.189175605773926, "learning_rate": 6.0676233254893414e-05, "loss": 4.6915, "step": 817050 }, { "epoch": 4.926978690560896, "grad_norm": 2.641714334487915, "learning_rate": 6.06598865035403e-05, "loss": 4.8115, "step": 817100 }, { "epoch": 4.927280182342229, "grad_norm": 2.7925338745117188, "learning_rate": 6.064354139636449e-05, "loss": 4.9699, "step": 817150 }, { "epoch": 4.927581674123563, "grad_norm": 5.084054946899414, "learning_rate": 6.062719793366683e-05, "loss": 4.6093, "step": 817200 }, { "epoch": 4.927883165904897, "grad_norm": 2.5879228115081787, "learning_rate": 6.061085611574799e-05, "loss": 4.6058, "step": 817250 }, { "epoch": 4.928184657686232, "grad_norm": 4.080022811889648, "learning_rate": 6.059451594290879e-05, "loss": 4.6501, "step": 817300 }, { "epoch": 4.928486149467566, "grad_norm": 3.534789562225342, "learning_rate": 6.057817741544997e-05, "loss": 4.7357, "step": 817350 }, { "epoch": 4.928787641248899, "grad_norm": 1.5315696001052856, "learning_rate": 6.056184053367221e-05, "loss": 4.8289, "step": 817400 }, { "epoch": 4.929089133030233, "grad_norm": 5.080430030822754, "learning_rate": 6.054550529787606e-05, "loss": 4.5723, "step": 817450 }, { "epoch": 4.9293906248115675, "grad_norm": 3.2614471912384033, "learning_rate": 6.0529171708362294e-05, "loss": 4.4736, "step": 817500 }, { "epoch": 4.929692116592902, "grad_norm": 3.405916452407837, "learning_rate": 6.051283976543138e-05, "loss": 4.6557, "step": 817550 }, { "epoch": 4.929993608374236, "grad_norm": 4.172651767730713, "learning_rate": 6.049650946938396e-05, "loss": 4.3817, "step": 817600 }, { "epoch": 4.93029510015557, "grad_norm": 3.9028849601745605, "learning_rate": 6.04801808205206e-05, "loss": 4.658, "step": 817650 }, { "epoch": 4.9305965919369035, "grad_norm": 4.869128227233887, "learning_rate": 6.0463853819141704e-05, "loss": 4.6499, "step": 817700 }, { "epoch": 4.930898083718238, "grad_norm": 4.714982032775879, "learning_rate": 6.0447528465547865e-05, "loss": 4.4335, "step": 817750 }, { "epoch": 4.931199575499572, "grad_norm": 4.871799468994141, "learning_rate": 6.043120476003945e-05, "loss": 4.7961, "step": 817800 }, { "epoch": 4.931501067280906, "grad_norm": 2.0180065631866455, "learning_rate": 6.041488270291686e-05, "loss": 4.4341, "step": 817850 }, { "epoch": 4.93180255906224, "grad_norm": 3.562466859817505, "learning_rate": 6.039856229448048e-05, "loss": 4.931, "step": 817900 }, { "epoch": 4.932104050843574, "grad_norm": 3.1855063438415527, "learning_rate": 6.0382243535030764e-05, "loss": 4.6554, "step": 817950 }, { "epoch": 4.932405542624908, "grad_norm": 3.2895781993865967, "learning_rate": 6.036592642486791e-05, "loss": 5.1216, "step": 818000 }, { "epoch": 4.932707034406242, "grad_norm": 5.278457164764404, "learning_rate": 6.034961096429226e-05, "loss": 4.6458, "step": 818050 }, { "epoch": 4.933008526187576, "grad_norm": 3.6726272106170654, "learning_rate": 6.033329715360415e-05, "loss": 4.9544, "step": 818100 }, { "epoch": 4.93331001796891, "grad_norm": 2.7154884338378906, "learning_rate": 6.0316984993103686e-05, "loss": 4.4639, "step": 818150 }, { "epoch": 4.933611509750245, "grad_norm": 3.8090734481811523, "learning_rate": 6.030067448309118e-05, "loss": 4.5489, "step": 818200 }, { "epoch": 4.933913001531578, "grad_norm": 2.791334867477417, "learning_rate": 6.028436562386674e-05, "loss": 4.6735, "step": 818250 }, { "epoch": 4.934214493312912, "grad_norm": 0.6362251043319702, "learning_rate": 6.026805841573048e-05, "loss": 4.6676, "step": 818300 }, { "epoch": 4.934515985094246, "grad_norm": 5.326333045959473, "learning_rate": 6.025175285898255e-05, "loss": 4.5236, "step": 818350 }, { "epoch": 4.934817476875581, "grad_norm": 2.7614850997924805, "learning_rate": 6.023544895392308e-05, "loss": 4.4336, "step": 818400 }, { "epoch": 4.935118968656915, "grad_norm": 1.8018473386764526, "learning_rate": 6.0219146700852015e-05, "loss": 4.9319, "step": 818450 }, { "epoch": 4.935420460438248, "grad_norm": 3.79495906829834, "learning_rate": 6.020284610006943e-05, "loss": 4.7653, "step": 818500 }, { "epoch": 4.935721952219582, "grad_norm": 3.671557903289795, "learning_rate": 6.018654715187538e-05, "loss": 4.8405, "step": 818550 }, { "epoch": 4.9360234440009165, "grad_norm": 2.6476950645446777, "learning_rate": 6.017024985656975e-05, "loss": 4.4563, "step": 818600 }, { "epoch": 4.936324935782251, "grad_norm": 2.5041024684906006, "learning_rate": 6.015395421445242e-05, "loss": 4.8262, "step": 818650 }, { "epoch": 4.936626427563585, "grad_norm": 5.5382466316223145, "learning_rate": 6.013766022582341e-05, "loss": 4.7057, "step": 818700 }, { "epoch": 4.936927919344919, "grad_norm": 2.9110116958618164, "learning_rate": 6.012136789098247e-05, "loss": 4.7473, "step": 818750 }, { "epoch": 4.9372294111262525, "grad_norm": 2.9421064853668213, "learning_rate": 6.0105077210229476e-05, "loss": 4.5381, "step": 818800 }, { "epoch": 4.937530902907587, "grad_norm": 3.3470284938812256, "learning_rate": 6.008878818386432e-05, "loss": 4.9496, "step": 818850 }, { "epoch": 4.937832394688921, "grad_norm": 5.6279802322387695, "learning_rate": 6.007250081218664e-05, "loss": 4.5505, "step": 818900 }, { "epoch": 4.938133886470255, "grad_norm": 1.2430146932601929, "learning_rate": 6.0056215095496314e-05, "loss": 4.7818, "step": 818950 }, { "epoch": 4.938435378251588, "grad_norm": 4.21351957321167, "learning_rate": 6.0039931034092974e-05, "loss": 4.6843, "step": 819000 }, { "epoch": 4.938435378251588, "eval_loss": 5.051662921905518, "eval_runtime": 39.0122, "eval_samples_per_second": 13.124, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.727340211440747, "step": 819000 }, { "epoch": 4.938736870032923, "grad_norm": 2.7405412197113037, "learning_rate": 6.002364862827629e-05, "loss": 4.6816, "step": 819050 }, { "epoch": 4.939038361814257, "grad_norm": 2.913465976715088, "learning_rate": 6.000736787834593e-05, "loss": 4.7549, "step": 819100 }, { "epoch": 4.939339853595591, "grad_norm": 3.6203114986419678, "learning_rate": 5.999108878460161e-05, "loss": 4.6057, "step": 819150 }, { "epoch": 4.939641345376925, "grad_norm": 2.7652981281280518, "learning_rate": 5.997481134734279e-05, "loss": 4.6492, "step": 819200 }, { "epoch": 4.939942837158259, "grad_norm": 1.8757929801940918, "learning_rate": 5.995853556686909e-05, "loss": 4.5789, "step": 819250 }, { "epoch": 4.940244328939593, "grad_norm": 2.394066095352173, "learning_rate": 5.9942261443480116e-05, "loss": 4.7425, "step": 819300 }, { "epoch": 4.940545820720927, "grad_norm": 4.664588928222656, "learning_rate": 5.992598897747524e-05, "loss": 4.4802, "step": 819350 }, { "epoch": 4.940847312502261, "grad_norm": 2.7766339778900146, "learning_rate": 5.9909718169154034e-05, "loss": 4.649, "step": 819400 }, { "epoch": 4.941148804283595, "grad_norm": 3.515570640563965, "learning_rate": 5.989344901881591e-05, "loss": 4.7327, "step": 819450 }, { "epoch": 4.94145029606493, "grad_norm": 2.66856050491333, "learning_rate": 5.987718152676021e-05, "loss": 4.2878, "step": 819500 }, { "epoch": 4.941751787846263, "grad_norm": 1.3928736448287964, "learning_rate": 5.9860915693286394e-05, "loss": 4.3249, "step": 819550 }, { "epoch": 4.942053279627597, "grad_norm": 3.143374443054199, "learning_rate": 5.984465151869382e-05, "loss": 4.5499, "step": 819600 }, { "epoch": 4.942354771408931, "grad_norm": 3.1838901042938232, "learning_rate": 5.9828389003281725e-05, "loss": 4.9433, "step": 819650 }, { "epoch": 4.9426562631902655, "grad_norm": 3.1466166973114014, "learning_rate": 5.9812128147349444e-05, "loss": 4.6229, "step": 819700 }, { "epoch": 4.9429577549716, "grad_norm": 1.8136954307556152, "learning_rate": 5.97958689511963e-05, "loss": 4.4804, "step": 819750 }, { "epoch": 4.943259246752934, "grad_norm": 3.4719598293304443, "learning_rate": 5.9779611415121465e-05, "loss": 4.5539, "step": 819800 }, { "epoch": 4.943560738534267, "grad_norm": 2.2696783542633057, "learning_rate": 5.976335553942408e-05, "loss": 4.9325, "step": 819850 }, { "epoch": 4.9438622303156015, "grad_norm": 2.886171340942383, "learning_rate": 5.9747101324403406e-05, "loss": 4.5598, "step": 819900 }, { "epoch": 4.944163722096936, "grad_norm": 4.091362476348877, "learning_rate": 5.973084877035847e-05, "loss": 4.4974, "step": 819950 }, { "epoch": 4.94446521387827, "grad_norm": 2.4470784664154053, "learning_rate": 5.971459787758846e-05, "loss": 4.6513, "step": 820000 }, { "epoch": 4.944766705659604, "grad_norm": 3.0475564002990723, "learning_rate": 5.9698348646392474e-05, "loss": 4.7735, "step": 820050 }, { "epoch": 4.945068197440937, "grad_norm": 2.4876091480255127, "learning_rate": 5.968210107706948e-05, "loss": 4.4491, "step": 820100 }, { "epoch": 4.945369689222272, "grad_norm": 3.982128143310547, "learning_rate": 5.966585516991856e-05, "loss": 4.6903, "step": 820150 }, { "epoch": 4.945671181003606, "grad_norm": 1.9613721370697021, "learning_rate": 5.964961092523861e-05, "loss": 4.7692, "step": 820200 }, { "epoch": 4.94597267278494, "grad_norm": 2.682622194290161, "learning_rate": 5.963336834332868e-05, "loss": 4.8173, "step": 820250 }, { "epoch": 4.946274164566274, "grad_norm": 3.279262065887451, "learning_rate": 5.9617127424487617e-05, "loss": 4.8497, "step": 820300 }, { "epoch": 4.946575656347608, "grad_norm": 3.5724270343780518, "learning_rate": 5.9600888169014374e-05, "loss": 4.4701, "step": 820350 }, { "epoch": 4.946877148128942, "grad_norm": 0.7573471665382385, "learning_rate": 5.9584650577207724e-05, "loss": 4.3818, "step": 820400 }, { "epoch": 4.947178639910276, "grad_norm": 3.9276721477508545, "learning_rate": 5.9568414649366556e-05, "loss": 4.9281, "step": 820450 }, { "epoch": 4.94748013169161, "grad_norm": 3.175497055053711, "learning_rate": 5.955218038578967e-05, "loss": 4.451, "step": 820500 }, { "epoch": 4.947781623472944, "grad_norm": 3.768054246902466, "learning_rate": 5.953594778677586e-05, "loss": 4.9099, "step": 820550 }, { "epoch": 4.948083115254278, "grad_norm": 1.446075439453125, "learning_rate": 5.9519716852623855e-05, "loss": 4.3962, "step": 820600 }, { "epoch": 4.948384607035612, "grad_norm": 3.3916709423065186, "learning_rate": 5.950348758363228e-05, "loss": 4.7181, "step": 820650 }, { "epoch": 4.948686098816946, "grad_norm": 3.4185893535614014, "learning_rate": 5.9487259980099916e-05, "loss": 4.4083, "step": 820700 }, { "epoch": 4.94898759059828, "grad_norm": 3.2675580978393555, "learning_rate": 5.947103404232532e-05, "loss": 4.156, "step": 820750 }, { "epoch": 4.9492890823796145, "grad_norm": 4.935215950012207, "learning_rate": 5.945480977060714e-05, "loss": 4.7157, "step": 820800 }, { "epoch": 4.949590574160949, "grad_norm": 1.3968627452850342, "learning_rate": 5.943858716524404e-05, "loss": 4.8817, "step": 820850 }, { "epoch": 4.949892065942282, "grad_norm": 2.7621302604675293, "learning_rate": 5.942236622653444e-05, "loss": 4.8661, "step": 820900 }, { "epoch": 4.950193557723616, "grad_norm": 3.494206190109253, "learning_rate": 5.940614695477698e-05, "loss": 4.513, "step": 820950 }, { "epoch": 4.9504950495049505, "grad_norm": 2.4052937030792236, "learning_rate": 5.93899293502701e-05, "loss": 4.5164, "step": 821000 }, { "epoch": 4.950796541286285, "grad_norm": 4.367593288421631, "learning_rate": 5.937371341331221e-05, "loss": 4.6915, "step": 821050 }, { "epoch": 4.951098033067619, "grad_norm": 3.1812267303466797, "learning_rate": 5.93574991442018e-05, "loss": 4.7502, "step": 821100 }, { "epoch": 4.951399524848952, "grad_norm": 2.1634366512298584, "learning_rate": 5.934128654323731e-05, "loss": 4.6987, "step": 821150 }, { "epoch": 4.951701016630286, "grad_norm": 5.222354412078857, "learning_rate": 5.932507561071703e-05, "loss": 4.6023, "step": 821200 }, { "epoch": 4.952002508411621, "grad_norm": 3.422132730484009, "learning_rate": 5.930886634693931e-05, "loss": 4.8615, "step": 821250 }, { "epoch": 4.952304000192955, "grad_norm": 3.216545343399048, "learning_rate": 5.929265875220256e-05, "loss": 4.7192, "step": 821300 }, { "epoch": 4.952605491974289, "grad_norm": 4.875896453857422, "learning_rate": 5.92764528268049e-05, "loss": 4.4166, "step": 821350 }, { "epoch": 4.952906983755623, "grad_norm": 3.2424678802490234, "learning_rate": 5.926024857104473e-05, "loss": 4.4532, "step": 821400 }, { "epoch": 4.9532084755369565, "grad_norm": 3.627156972885132, "learning_rate": 5.924404598522019e-05, "loss": 4.6254, "step": 821450 }, { "epoch": 4.953509967318291, "grad_norm": 3.8563432693481445, "learning_rate": 5.922784506962941e-05, "loss": 4.5425, "step": 821500 }, { "epoch": 4.953811459099625, "grad_norm": 1.9995068311691284, "learning_rate": 5.92116458245706e-05, "loss": 4.9262, "step": 821550 }, { "epoch": 4.954112950880959, "grad_norm": 2.6421568393707275, "learning_rate": 5.919544825034193e-05, "loss": 4.6654, "step": 821600 }, { "epoch": 4.954414442662293, "grad_norm": 3.6409127712249756, "learning_rate": 5.91792523472414e-05, "loss": 4.5833, "step": 821650 }, { "epoch": 4.954715934443627, "grad_norm": 2.7036380767822266, "learning_rate": 5.916305811556714e-05, "loss": 4.9457, "step": 821700 }, { "epoch": 4.955017426224961, "grad_norm": 3.981766700744629, "learning_rate": 5.914686555561721e-05, "loss": 4.3912, "step": 821750 }, { "epoch": 4.955318918006295, "grad_norm": 2.3524675369262695, "learning_rate": 5.913067466768956e-05, "loss": 5.2878, "step": 821800 }, { "epoch": 4.955620409787629, "grad_norm": 2.724458932876587, "learning_rate": 5.911448545208211e-05, "loss": 4.6232, "step": 821850 }, { "epoch": 4.9559219015689635, "grad_norm": 3.7501885890960693, "learning_rate": 5.909829790909291e-05, "loss": 4.3827, "step": 821900 }, { "epoch": 4.956223393350298, "grad_norm": 2.6148698329925537, "learning_rate": 5.908211203901976e-05, "loss": 4.3418, "step": 821950 }, { "epoch": 4.956524885131631, "grad_norm": 1.9695461988449097, "learning_rate": 5.9065927842160604e-05, "loss": 4.6222, "step": 822000 }, { "epoch": 4.956524885131631, "eval_loss": 5.047754287719727, "eval_runtime": 39.1255, "eval_samples_per_second": 13.086, "eval_steps_per_second": 6.543, "eval_tts_loss": 7.736879994252867, "step": 822000 }, { "epoch": 4.956826376912965, "grad_norm": 1.8586466312408447, "learning_rate": 5.9049745318813316e-05, "loss": 4.3742, "step": 822050 }, { "epoch": 4.957127868694299, "grad_norm": 3.813685417175293, "learning_rate": 5.903356446927561e-05, "loss": 4.613, "step": 822100 }, { "epoch": 4.957429360475634, "grad_norm": 3.263587236404419, "learning_rate": 5.9017385293845413e-05, "loss": 5.1111, "step": 822150 }, { "epoch": 4.957730852256968, "grad_norm": 3.2022712230682373, "learning_rate": 5.9001207792820385e-05, "loss": 4.1364, "step": 822200 }, { "epoch": 4.958032344038301, "grad_norm": 3.395913600921631, "learning_rate": 5.8985031966498206e-05, "loss": 4.495, "step": 822250 }, { "epoch": 4.958333835819635, "grad_norm": 4.142703533172607, "learning_rate": 5.896885781517664e-05, "loss": 4.5065, "step": 822300 }, { "epoch": 4.95863532760097, "grad_norm": 4.053567409515381, "learning_rate": 5.895268533915339e-05, "loss": 4.6418, "step": 822350 }, { "epoch": 4.958936819382304, "grad_norm": 3.0642411708831787, "learning_rate": 5.8936514538725984e-05, "loss": 4.8294, "step": 822400 }, { "epoch": 4.959238311163638, "grad_norm": 2.3721871376037598, "learning_rate": 5.892034541419205e-05, "loss": 4.8164, "step": 822450 }, { "epoch": 4.959539802944972, "grad_norm": 2.685171127319336, "learning_rate": 5.8904177965849255e-05, "loss": 4.6547, "step": 822500 }, { "epoch": 4.9598412947263055, "grad_norm": 2.9329850673675537, "learning_rate": 5.8888012193995e-05, "loss": 4.8327, "step": 822550 }, { "epoch": 4.96014278650764, "grad_norm": 3.4072577953338623, "learning_rate": 5.8871848098926935e-05, "loss": 4.3428, "step": 822600 }, { "epoch": 4.960444278288974, "grad_norm": 3.6315503120422363, "learning_rate": 5.885568568094243e-05, "loss": 4.3467, "step": 822650 }, { "epoch": 4.960745770070308, "grad_norm": 1.2494531869888306, "learning_rate": 5.8839524940338916e-05, "loss": 4.7456, "step": 822700 }, { "epoch": 4.9610472618516415, "grad_norm": 3.538111448287964, "learning_rate": 5.8823365877413865e-05, "loss": 4.4346, "step": 822750 }, { "epoch": 4.961348753632976, "grad_norm": 2.4690845012664795, "learning_rate": 5.880720849246469e-05, "loss": 4.4214, "step": 822800 }, { "epoch": 4.96165024541431, "grad_norm": 3.5635032653808594, "learning_rate": 5.8791052785788646e-05, "loss": 4.8525, "step": 822850 }, { "epoch": 4.961951737195644, "grad_norm": 5.061214923858643, "learning_rate": 5.877489875768312e-05, "loss": 4.5147, "step": 822900 }, { "epoch": 4.962253228976978, "grad_norm": 2.11179518699646, "learning_rate": 5.8758746408445456e-05, "loss": 4.1797, "step": 822950 }, { "epoch": 4.9625547207583125, "grad_norm": 5.655355453491211, "learning_rate": 5.874259573837285e-05, "loss": 4.4482, "step": 823000 }, { "epoch": 4.962856212539646, "grad_norm": 2.71657133102417, "learning_rate": 5.872644674776248e-05, "loss": 4.1987, "step": 823050 }, { "epoch": 4.96315770432098, "grad_norm": 2.7614798545837402, "learning_rate": 5.871029943691166e-05, "loss": 4.7562, "step": 823100 }, { "epoch": 4.963459196102314, "grad_norm": 4.320824146270752, "learning_rate": 5.8694153806117464e-05, "loss": 4.7311, "step": 823150 }, { "epoch": 4.963760687883648, "grad_norm": 3.1726338863372803, "learning_rate": 5.867800985567704e-05, "loss": 5.0182, "step": 823200 }, { "epoch": 4.964062179664983, "grad_norm": 3.6868717670440674, "learning_rate": 5.8661867585887585e-05, "loss": 4.4911, "step": 823250 }, { "epoch": 4.964363671446316, "grad_norm": 7.286736965179443, "learning_rate": 5.864572699704606e-05, "loss": 5.0913, "step": 823300 }, { "epoch": 4.96466516322765, "grad_norm": 5.967582702636719, "learning_rate": 5.862958808944961e-05, "loss": 4.2686, "step": 823350 }, { "epoch": 4.964966655008984, "grad_norm": 3.314368486404419, "learning_rate": 5.8613450863395205e-05, "loss": 4.7957, "step": 823400 }, { "epoch": 4.965268146790319, "grad_norm": 2.8355486392974854, "learning_rate": 5.8597315319179765e-05, "loss": 4.5417, "step": 823450 }, { "epoch": 4.965569638571653, "grad_norm": 4.262061595916748, "learning_rate": 5.858118145710029e-05, "loss": 4.6778, "step": 823500 }, { "epoch": 4.965871130352987, "grad_norm": 2.2026188373565674, "learning_rate": 5.856504927745377e-05, "loss": 4.0812, "step": 823550 }, { "epoch": 4.96617262213432, "grad_norm": 3.1251091957092285, "learning_rate": 5.854891878053698e-05, "loss": 4.9077, "step": 823600 }, { "epoch": 4.9664741139156545, "grad_norm": 2.7578682899475098, "learning_rate": 5.853278996664684e-05, "loss": 4.6483, "step": 823650 }, { "epoch": 4.966775605696989, "grad_norm": 4.452394485473633, "learning_rate": 5.8516662836080206e-05, "loss": 4.6316, "step": 823700 }, { "epoch": 4.967077097478323, "grad_norm": 3.5812079906463623, "learning_rate": 5.850053738913381e-05, "loss": 4.8683, "step": 823750 }, { "epoch": 4.967378589259657, "grad_norm": 2.7036495208740234, "learning_rate": 5.8484413626104475e-05, "loss": 4.8935, "step": 823800 }, { "epoch": 4.9676800810409905, "grad_norm": 2.9871320724487305, "learning_rate": 5.8468291547288864e-05, "loss": 4.5824, "step": 823850 }, { "epoch": 4.967981572822325, "grad_norm": 3.515507221221924, "learning_rate": 5.845217115298377e-05, "loss": 4.2648, "step": 823900 }, { "epoch": 4.968283064603659, "grad_norm": 3.805798053741455, "learning_rate": 5.843605244348578e-05, "loss": 4.4176, "step": 823950 }, { "epoch": 4.968584556384993, "grad_norm": 2.1230452060699463, "learning_rate": 5.841993541909157e-05, "loss": 4.0217, "step": 824000 }, { "epoch": 4.968886048166327, "grad_norm": 3.8410377502441406, "learning_rate": 5.840382008009781e-05, "loss": 4.3779, "step": 824050 }, { "epoch": 4.9691875399476615, "grad_norm": 3.5135931968688965, "learning_rate": 5.838770642680096e-05, "loss": 4.7481, "step": 824100 }, { "epoch": 4.969489031728995, "grad_norm": 3.5033299922943115, "learning_rate": 5.8371594459497714e-05, "loss": 4.4719, "step": 824150 }, { "epoch": 4.969790523510329, "grad_norm": 4.896235942840576, "learning_rate": 5.835548417848448e-05, "loss": 4.4364, "step": 824200 }, { "epoch": 4.970092015291663, "grad_norm": 3.162250518798828, "learning_rate": 5.833937558405774e-05, "loss": 4.0381, "step": 824250 }, { "epoch": 4.970393507072997, "grad_norm": 3.670900821685791, "learning_rate": 5.8323268676514004e-05, "loss": 4.7103, "step": 824300 }, { "epoch": 4.970694998854331, "grad_norm": 1.9523160457611084, "learning_rate": 5.8307163456149714e-05, "loss": 4.3008, "step": 824350 }, { "epoch": 4.970996490635665, "grad_norm": 1.6510097980499268, "learning_rate": 5.829105992326118e-05, "loss": 4.5499, "step": 824400 }, { "epoch": 4.971297982416999, "grad_norm": 2.369189739227295, "learning_rate": 5.82749580781448e-05, "loss": 4.6773, "step": 824450 }, { "epoch": 4.971599474198333, "grad_norm": 2.0022823810577393, "learning_rate": 5.8258857921096974e-05, "loss": 4.701, "step": 824500 }, { "epoch": 4.971900965979668, "grad_norm": 2.9186415672302246, "learning_rate": 5.8242759452413963e-05, "loss": 4.3136, "step": 824550 }, { "epoch": 4.972202457761002, "grad_norm": 3.7969329357147217, "learning_rate": 5.8226662672391944e-05, "loss": 4.7045, "step": 824600 }, { "epoch": 4.972503949542335, "grad_norm": 2.777181386947632, "learning_rate": 5.82105675813273e-05, "loss": 4.8098, "step": 824650 }, { "epoch": 4.972805441323669, "grad_norm": 3.2215404510498047, "learning_rate": 5.819447417951611e-05, "loss": 4.515, "step": 824700 }, { "epoch": 4.9731069331050035, "grad_norm": 3.574263095855713, "learning_rate": 5.81783824672546e-05, "loss": 4.5245, "step": 824750 }, { "epoch": 4.973408424886338, "grad_norm": 2.629957437515259, "learning_rate": 5.816229244483899e-05, "loss": 4.446, "step": 824800 }, { "epoch": 4.973709916667672, "grad_norm": 3.925319194793701, "learning_rate": 5.814620411256528e-05, "loss": 4.942, "step": 824850 }, { "epoch": 4.974011408449005, "grad_norm": 2.7368173599243164, "learning_rate": 5.8130117470729596e-05, "loss": 4.2197, "step": 824900 }, { "epoch": 4.9743129002303395, "grad_norm": 5.918226718902588, "learning_rate": 5.811403251962804e-05, "loss": 4.8771, "step": 824950 }, { "epoch": 4.974614392011674, "grad_norm": 3.20747971534729, "learning_rate": 5.8097949259556596e-05, "loss": 4.44, "step": 825000 }, { "epoch": 4.974614392011674, "eval_loss": 5.041656970977783, "eval_runtime": 39.0572, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.773269058970161, "step": 825000 }, { "epoch": 4.974915883793008, "grad_norm": 2.9623687267303467, "learning_rate": 5.8081867690811185e-05, "loss": 5.0925, "step": 825050 }, { "epoch": 4.975217375574342, "grad_norm": 4.810176372528076, "learning_rate": 5.806578781368788e-05, "loss": 4.6164, "step": 825100 }, { "epoch": 4.975518867355676, "grad_norm": 2.4508156776428223, "learning_rate": 5.80497096284825e-05, "loss": 4.5704, "step": 825150 }, { "epoch": 4.97582035913701, "grad_norm": 2.9884917736053467, "learning_rate": 5.8033633135491e-05, "loss": 4.5056, "step": 825200 }, { "epoch": 4.976121850918344, "grad_norm": 3.139240264892578, "learning_rate": 5.801755833500929e-05, "loss": 4.7745, "step": 825250 }, { "epoch": 4.976423342699678, "grad_norm": 3.8979992866516113, "learning_rate": 5.800148522733309e-05, "loss": 4.5313, "step": 825300 }, { "epoch": 4.976724834481012, "grad_norm": 1.7083021402359009, "learning_rate": 5.798541381275833e-05, "loss": 4.4468, "step": 825350 }, { "epoch": 4.977026326262346, "grad_norm": 2.4144861698150635, "learning_rate": 5.796934409158071e-05, "loss": 4.8244, "step": 825400 }, { "epoch": 4.97732781804368, "grad_norm": 2.656522750854492, "learning_rate": 5.7953276064095926e-05, "loss": 4.7384, "step": 825450 }, { "epoch": 4.977629309825014, "grad_norm": 6.485114574432373, "learning_rate": 5.793720973059974e-05, "loss": 4.633, "step": 825500 }, { "epoch": 4.977930801606348, "grad_norm": 2.957383632659912, "learning_rate": 5.792114509138787e-05, "loss": 4.9019, "step": 825550 }, { "epoch": 4.978232293387682, "grad_norm": 3.2355895042419434, "learning_rate": 5.790508214675588e-05, "loss": 4.7896, "step": 825600 }, { "epoch": 4.978533785169017, "grad_norm": 3.19390606880188, "learning_rate": 5.788902089699942e-05, "loss": 4.4851, "step": 825650 }, { "epoch": 4.978835276950351, "grad_norm": 2.704618215560913, "learning_rate": 5.787296134241415e-05, "loss": 4.6505, "step": 825700 }, { "epoch": 4.979136768731684, "grad_norm": 1.2214146852493286, "learning_rate": 5.7856903483295485e-05, "loss": 4.6551, "step": 825750 }, { "epoch": 4.979438260513018, "grad_norm": 2.0731215476989746, "learning_rate": 5.784084731993908e-05, "loss": 4.5695, "step": 825800 }, { "epoch": 4.9797397522943525, "grad_norm": 3.223304510116577, "learning_rate": 5.782479285264035e-05, "loss": 4.9111, "step": 825850 }, { "epoch": 4.980041244075687, "grad_norm": 3.473008871078491, "learning_rate": 5.780874008169472e-05, "loss": 4.6582, "step": 825900 }, { "epoch": 4.980342735857021, "grad_norm": 3.203653573989868, "learning_rate": 5.779268900739766e-05, "loss": 4.7167, "step": 825950 }, { "epoch": 4.980644227638354, "grad_norm": 2.607508659362793, "learning_rate": 5.777663963004463e-05, "loss": 4.5301, "step": 826000 }, { "epoch": 4.9809457194196884, "grad_norm": 1.2605828046798706, "learning_rate": 5.776059194993088e-05, "loss": 4.4418, "step": 826050 }, { "epoch": 4.981247211201023, "grad_norm": 2.9798593521118164, "learning_rate": 5.774454596735182e-05, "loss": 4.3314, "step": 826100 }, { "epoch": 4.981548702982357, "grad_norm": 4.670090675354004, "learning_rate": 5.772850168260276e-05, "loss": 4.6244, "step": 826150 }, { "epoch": 4.981850194763691, "grad_norm": 3.1630561351776123, "learning_rate": 5.7712459095978965e-05, "loss": 4.642, "step": 826200 }, { "epoch": 4.982151686545025, "grad_norm": 2.3505001068115234, "learning_rate": 5.7696418207775604e-05, "loss": 4.9162, "step": 826250 }, { "epoch": 4.982453178326359, "grad_norm": 2.702009439468384, "learning_rate": 5.768037901828799e-05, "loss": 4.5853, "step": 826300 }, { "epoch": 4.982754670107693, "grad_norm": 3.5574004650115967, "learning_rate": 5.7664341527811215e-05, "loss": 4.5042, "step": 826350 }, { "epoch": 4.983056161889027, "grad_norm": 3.8235397338867188, "learning_rate": 5.7648305736640446e-05, "loss": 4.5732, "step": 826400 }, { "epoch": 4.983357653670361, "grad_norm": 2.3422834873199463, "learning_rate": 5.763227164507088e-05, "loss": 4.4634, "step": 826450 }, { "epoch": 4.9836591454516945, "grad_norm": 4.3415093421936035, "learning_rate": 5.7616239253397495e-05, "loss": 4.5219, "step": 826500 }, { "epoch": 4.983960637233029, "grad_norm": 3.287628412246704, "learning_rate": 5.7600208561915416e-05, "loss": 4.6377, "step": 826550 }, { "epoch": 4.984262129014363, "grad_norm": 5.4814372062683105, "learning_rate": 5.758417957091966e-05, "loss": 4.9407, "step": 826600 }, { "epoch": 4.984563620795697, "grad_norm": 2.823676586151123, "learning_rate": 5.756815228070513e-05, "loss": 4.8554, "step": 826650 }, { "epoch": 4.984865112577031, "grad_norm": 3.3450989723205566, "learning_rate": 5.755212669156685e-05, "loss": 4.428, "step": 826700 }, { "epoch": 4.985166604358366, "grad_norm": 5.303634166717529, "learning_rate": 5.75361028037998e-05, "loss": 4.5034, "step": 826750 }, { "epoch": 4.985468096139699, "grad_norm": 4.133828163146973, "learning_rate": 5.7520080617698774e-05, "loss": 4.5154, "step": 826800 }, { "epoch": 4.985769587921033, "grad_norm": 3.9803314208984375, "learning_rate": 5.7504060133558686e-05, "loss": 4.3835, "step": 826850 }, { "epoch": 4.986071079702367, "grad_norm": 3.8843348026275635, "learning_rate": 5.748804135167443e-05, "loss": 4.7744, "step": 826900 }, { "epoch": 4.9863725714837015, "grad_norm": 3.6225528717041016, "learning_rate": 5.7472024272340697e-05, "loss": 4.8863, "step": 826950 }, { "epoch": 4.986674063265036, "grad_norm": 2.8371036052703857, "learning_rate": 5.745600889585236e-05, "loss": 4.7978, "step": 827000 }, { "epoch": 4.986975555046369, "grad_norm": 2.903904914855957, "learning_rate": 5.7439995222504115e-05, "loss": 4.3345, "step": 827050 }, { "epoch": 4.987277046827703, "grad_norm": 2.550220012664795, "learning_rate": 5.742398325259062e-05, "loss": 4.6051, "step": 827100 }, { "epoch": 4.987578538609037, "grad_norm": 3.4969379901885986, "learning_rate": 5.7407972986406594e-05, "loss": 4.4282, "step": 827150 }, { "epoch": 4.987880030390372, "grad_norm": 1.7166733741760254, "learning_rate": 5.739196442424671e-05, "loss": 4.7654, "step": 827200 }, { "epoch": 4.988181522171706, "grad_norm": 3.02825665473938, "learning_rate": 5.737595756640559e-05, "loss": 4.5424, "step": 827250 }, { "epoch": 4.98848301395304, "grad_norm": 3.0483815670013428, "learning_rate": 5.735995241317774e-05, "loss": 4.2957, "step": 827300 }, { "epoch": 4.988784505734373, "grad_norm": 3.063249349594116, "learning_rate": 5.734394896485782e-05, "loss": 4.858, "step": 827350 }, { "epoch": 4.989085997515708, "grad_norm": 4.540853977203369, "learning_rate": 5.732794722174028e-05, "loss": 4.6545, "step": 827400 }, { "epoch": 4.989387489297042, "grad_norm": 3.607239246368408, "learning_rate": 5.731194718411959e-05, "loss": 4.3443, "step": 827450 }, { "epoch": 4.989688981078376, "grad_norm": 2.544323444366455, "learning_rate": 5.7295948852290216e-05, "loss": 4.766, "step": 827500 }, { "epoch": 4.98999047285971, "grad_norm": 3.417189359664917, "learning_rate": 5.727995222654667e-05, "loss": 4.7468, "step": 827550 }, { "epoch": 4.9902919646410435, "grad_norm": 2.903336524963379, "learning_rate": 5.726395730718324e-05, "loss": 5.0054, "step": 827600 }, { "epoch": 4.990593456422378, "grad_norm": 5.059940814971924, "learning_rate": 5.724796409449434e-05, "loss": 4.5444, "step": 827650 }, { "epoch": 4.990894948203712, "grad_norm": 2.1844985485076904, "learning_rate": 5.723197258877433e-05, "loss": 4.3044, "step": 827700 }, { "epoch": 4.991196439985046, "grad_norm": 2.581958532333374, "learning_rate": 5.7215982790317484e-05, "loss": 4.268, "step": 827750 }, { "epoch": 4.99149793176638, "grad_norm": 3.137983798980713, "learning_rate": 5.719999469941802e-05, "loss": 5.0263, "step": 827800 }, { "epoch": 4.9917994235477146, "grad_norm": 3.8991315364837646, "learning_rate": 5.7184008316370264e-05, "loss": 4.5032, "step": 827850 }, { "epoch": 4.992100915329048, "grad_norm": 1.9569554328918457, "learning_rate": 5.716802364146831e-05, "loss": 4.8578, "step": 827900 }, { "epoch": 4.992402407110382, "grad_norm": 3.0314781665802, "learning_rate": 5.715204067500644e-05, "loss": 4.548, "step": 827950 }, { "epoch": 4.992703898891716, "grad_norm": 3.612100839614868, "learning_rate": 5.7136059417278785e-05, "loss": 4.5448, "step": 828000 }, { "epoch": 4.992703898891716, "eval_loss": 5.047117233276367, "eval_runtime": 38.9727, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.569, "eval_tts_loss": 7.737744487475332, "step": 828000 }, { "epoch": 4.9930053906730505, "grad_norm": 3.8017733097076416, "learning_rate": 5.71200798685794e-05, "loss": 4.2705, "step": 828050 }, { "epoch": 4.993306882454384, "grad_norm": 4.201447010040283, "learning_rate": 5.710410202920238e-05, "loss": 4.8558, "step": 828100 }, { "epoch": 4.993608374235718, "grad_norm": 2.920846462249756, "learning_rate": 5.7088125899441845e-05, "loss": 4.8115, "step": 828150 }, { "epoch": 4.993909866017052, "grad_norm": 3.8851826190948486, "learning_rate": 5.7072151479591774e-05, "loss": 4.3437, "step": 828200 }, { "epoch": 4.994211357798386, "grad_norm": 3.5571846961975098, "learning_rate": 5.7056178769946075e-05, "loss": 5.0037, "step": 828250 }, { "epoch": 4.994512849579721, "grad_norm": 2.925109386444092, "learning_rate": 5.704020777079884e-05, "loss": 4.8132, "step": 828300 }, { "epoch": 4.994814341361055, "grad_norm": 3.0835437774658203, "learning_rate": 5.7024238482443843e-05, "loss": 4.949, "step": 828350 }, { "epoch": 4.995115833142388, "grad_norm": 2.6202232837677, "learning_rate": 5.7008270905175077e-05, "loss": 5.1624, "step": 828400 }, { "epoch": 4.995417324923722, "grad_norm": 3.15885066986084, "learning_rate": 5.6992305039286405e-05, "loss": 4.738, "step": 828450 }, { "epoch": 4.995718816705057, "grad_norm": 2.584901809692383, "learning_rate": 5.697634088507159e-05, "loss": 4.7914, "step": 828500 }, { "epoch": 4.996020308486391, "grad_norm": 4.361555576324463, "learning_rate": 5.6960378442824526e-05, "loss": 4.4429, "step": 828550 }, { "epoch": 4.996321800267725, "grad_norm": 2.468512773513794, "learning_rate": 5.694441771283893e-05, "loss": 4.8561, "step": 828600 }, { "epoch": 4.996623292049058, "grad_norm": 3.942110538482666, "learning_rate": 5.692845869540846e-05, "loss": 5.0603, "step": 828650 }, { "epoch": 4.9969247838303925, "grad_norm": 3.049010992050171, "learning_rate": 5.6912501390826894e-05, "loss": 4.6473, "step": 828700 }, { "epoch": 4.997226275611727, "grad_norm": 4.233275890350342, "learning_rate": 5.689654579938795e-05, "loss": 4.6455, "step": 828750 }, { "epoch": 4.997527767393061, "grad_norm": 3.1091108322143555, "learning_rate": 5.688059192138515e-05, "loss": 4.796, "step": 828800 }, { "epoch": 4.997829259174395, "grad_norm": 3.0354740619659424, "learning_rate": 5.686463975711217e-05, "loss": 4.6618, "step": 828850 }, { "epoch": 4.998130750955729, "grad_norm": 3.2858974933624268, "learning_rate": 5.6848689306862656e-05, "loss": 3.9991, "step": 828900 }, { "epoch": 4.998432242737063, "grad_norm": 3.092846155166626, "learning_rate": 5.6832740570930074e-05, "loss": 5.0619, "step": 828950 }, { "epoch": 4.998733734518397, "grad_norm": 3.536365509033203, "learning_rate": 5.681679354960788e-05, "loss": 4.6594, "step": 829000 }, { "epoch": 4.999035226299731, "grad_norm": 4.868853569030762, "learning_rate": 5.680084824318969e-05, "loss": 4.7451, "step": 829050 }, { "epoch": 4.999336718081065, "grad_norm": 3.596736431121826, "learning_rate": 5.678490465196882e-05, "loss": 4.2276, "step": 829100 }, { "epoch": 4.9996382098623995, "grad_norm": 3.5632197856903076, "learning_rate": 5.676896277623876e-05, "loss": 4.4473, "step": 829150 }, { "epoch": 4.999939701643733, "grad_norm": 3.6511027812957764, "learning_rate": 5.675302261629295e-05, "loss": 4.7571, "step": 829200 }, { "epoch": 5.000241193425067, "grad_norm": 3.5106945037841797, "learning_rate": 5.673708417242462e-05, "loss": 4.5444, "step": 829250 }, { "epoch": 5.000542685206401, "grad_norm": 4.114288330078125, "learning_rate": 5.6721147444927166e-05, "loss": 4.368, "step": 829300 }, { "epoch": 5.000844176987735, "grad_norm": 3.912811279296875, "learning_rate": 5.670521243409394e-05, "loss": 4.5986, "step": 829350 }, { "epoch": 5.00114566876907, "grad_norm": 4.75300931930542, "learning_rate": 5.668927914021813e-05, "loss": 4.4194, "step": 829400 }, { "epoch": 5.001447160550404, "grad_norm": 3.1038830280303955, "learning_rate": 5.667334756359294e-05, "loss": 5.0081, "step": 829450 }, { "epoch": 5.001748652331737, "grad_norm": 2.6719634532928467, "learning_rate": 5.665741770451164e-05, "loss": 4.3755, "step": 829500 }, { "epoch": 5.002050144113071, "grad_norm": 4.175896167755127, "learning_rate": 5.6641489563267314e-05, "loss": 4.8635, "step": 829550 }, { "epoch": 5.002351635894406, "grad_norm": 2.1869919300079346, "learning_rate": 5.662556314015315e-05, "loss": 4.2785, "step": 829600 }, { "epoch": 5.00265312767574, "grad_norm": 3.2784225940704346, "learning_rate": 5.660963843546228e-05, "loss": 4.4849, "step": 829650 }, { "epoch": 5.002954619457074, "grad_norm": 7.053240776062012, "learning_rate": 5.65937154494877e-05, "loss": 4.7276, "step": 829700 }, { "epoch": 5.003256111238407, "grad_norm": 2.358682870864868, "learning_rate": 5.657779418252253e-05, "loss": 4.5744, "step": 829750 }, { "epoch": 5.0035576030197415, "grad_norm": 4.12529182434082, "learning_rate": 5.656187463485975e-05, "loss": 4.5645, "step": 829800 }, { "epoch": 5.003859094801076, "grad_norm": 2.6466352939605713, "learning_rate": 5.654595680679226e-05, "loss": 4.8422, "step": 829850 }, { "epoch": 5.00416058658241, "grad_norm": 3.531601667404175, "learning_rate": 5.6530040698613075e-05, "loss": 4.1702, "step": 829900 }, { "epoch": 5.004462078363744, "grad_norm": 2.606318473815918, "learning_rate": 5.651412631061516e-05, "loss": 4.3691, "step": 829950 }, { "epoch": 5.0047635701450774, "grad_norm": 3.627708673477173, "learning_rate": 5.649821364309129e-05, "loss": 4.8051, "step": 830000 }, { "epoch": 5.005065061926412, "grad_norm": 2.577983856201172, "learning_rate": 5.6482302696334346e-05, "loss": 4.4839, "step": 830050 }, { "epoch": 5.005366553707746, "grad_norm": 2.607814311981201, "learning_rate": 5.6466393470637234e-05, "loss": 4.2479, "step": 830100 }, { "epoch": 5.00566804548908, "grad_norm": 2.2872421741485596, "learning_rate": 5.645048596629265e-05, "loss": 4.844, "step": 830150 }, { "epoch": 5.005969537270414, "grad_norm": 4.646861553192139, "learning_rate": 5.6434580183593336e-05, "loss": 4.4332, "step": 830200 }, { "epoch": 5.0062710290517485, "grad_norm": 4.835063934326172, "learning_rate": 5.6418676122832087e-05, "loss": 4.673, "step": 830250 }, { "epoch": 5.006572520833082, "grad_norm": 2.855091094970703, "learning_rate": 5.6402773784301505e-05, "loss": 4.1271, "step": 830300 }, { "epoch": 5.006874012614416, "grad_norm": 2.46256160736084, "learning_rate": 5.638687316829432e-05, "loss": 4.789, "step": 830350 }, { "epoch": 5.00717550439575, "grad_norm": 3.5099871158599854, "learning_rate": 5.637097427510318e-05, "loss": 4.3278, "step": 830400 }, { "epoch": 5.007476996177084, "grad_norm": 2.371729612350464, "learning_rate": 5.635507710502059e-05, "loss": 4.6155, "step": 830450 }, { "epoch": 5.007778487958419, "grad_norm": 3.586376190185547, "learning_rate": 5.6339181658339166e-05, "loss": 4.0918, "step": 830500 }, { "epoch": 5.008079979739752, "grad_norm": 1.31264066696167, "learning_rate": 5.63232879353515e-05, "loss": 4.3618, "step": 830550 }, { "epoch": 5.008381471521086, "grad_norm": 3.094433307647705, "learning_rate": 5.6307395936350026e-05, "loss": 4.4966, "step": 830600 }, { "epoch": 5.00868296330242, "grad_norm": 3.9483394622802734, "learning_rate": 5.629150566162718e-05, "loss": 4.8751, "step": 830650 }, { "epoch": 5.008984455083755, "grad_norm": 4.611080169677734, "learning_rate": 5.6275617111475445e-05, "loss": 4.2734, "step": 830700 }, { "epoch": 5.009285946865089, "grad_norm": 2.671893358230591, "learning_rate": 5.625973028618726e-05, "loss": 4.2134, "step": 830750 }, { "epoch": 5.009587438646422, "grad_norm": 2.665576934814453, "learning_rate": 5.624384518605493e-05, "loss": 4.4778, "step": 830800 }, { "epoch": 5.009888930427756, "grad_norm": 2.8681752681732178, "learning_rate": 5.6227961811370825e-05, "loss": 4.6356, "step": 830850 }, { "epoch": 5.0101904222090905, "grad_norm": 3.317854166030884, "learning_rate": 5.6212080162427325e-05, "loss": 4.4639, "step": 830900 }, { "epoch": 5.010491913990425, "grad_norm": 2.9209542274475098, "learning_rate": 5.619620023951664e-05, "loss": 4.395, "step": 830950 }, { "epoch": 5.010793405771759, "grad_norm": 1.6458510160446167, "learning_rate": 5.618032204293097e-05, "loss": 4.1589, "step": 831000 }, { "epoch": 5.010793405771759, "eval_loss": 5.039275169372559, "eval_runtime": 39.0155, "eval_samples_per_second": 13.123, "eval_steps_per_second": 6.561, "eval_tts_loss": 7.787231666074408, "step": 831000 }, { "epoch": 5.011094897553093, "grad_norm": 4.1650471687316895, "learning_rate": 5.616444557296264e-05, "loss": 4.3053, "step": 831050 }, { "epoch": 5.011396389334426, "grad_norm": 2.7942662239074707, "learning_rate": 5.614857082990372e-05, "loss": 4.1161, "step": 831100 }, { "epoch": 5.011697881115761, "grad_norm": 4.351217269897461, "learning_rate": 5.613269781404643e-05, "loss": 4.6672, "step": 831150 }, { "epoch": 5.011999372897095, "grad_norm": 4.4865875244140625, "learning_rate": 5.611682652568293e-05, "loss": 4.4833, "step": 831200 }, { "epoch": 5.012300864678429, "grad_norm": 3.5928609371185303, "learning_rate": 5.61009569651052e-05, "loss": 4.4812, "step": 831250 }, { "epoch": 5.012602356459763, "grad_norm": 0.48387640714645386, "learning_rate": 5.6085089132605426e-05, "loss": 4.8084, "step": 831300 }, { "epoch": 5.012903848241097, "grad_norm": 3.6415631771087646, "learning_rate": 5.60692230284755e-05, "loss": 5.063, "step": 831350 }, { "epoch": 5.013205340022431, "grad_norm": 2.831094741821289, "learning_rate": 5.605335865300753e-05, "loss": 4.8652, "step": 831400 }, { "epoch": 5.013506831803765, "grad_norm": 4.470372676849365, "learning_rate": 5.603749600649336e-05, "loss": 4.4554, "step": 831450 }, { "epoch": 5.013808323585099, "grad_norm": 3.9512619972229004, "learning_rate": 5.6021635089225056e-05, "loss": 4.2718, "step": 831500 }, { "epoch": 5.014109815366433, "grad_norm": 2.9764974117279053, "learning_rate": 5.600577590149438e-05, "loss": 4.6937, "step": 831550 }, { "epoch": 5.014411307147767, "grad_norm": 2.476335048675537, "learning_rate": 5.598991844359326e-05, "loss": 4.471, "step": 831600 }, { "epoch": 5.014712798929101, "grad_norm": 2.1546740531921387, "learning_rate": 5.59740627158136e-05, "loss": 4.3549, "step": 831650 }, { "epoch": 5.015014290710435, "grad_norm": 3.280500650405884, "learning_rate": 5.5958208718447075e-05, "loss": 4.6735, "step": 831700 }, { "epoch": 5.015315782491769, "grad_norm": 3.4592161178588867, "learning_rate": 5.594235645178555e-05, "loss": 4.7636, "step": 831750 }, { "epoch": 5.0156172742731036, "grad_norm": 2.545860767364502, "learning_rate": 5.5926505916120737e-05, "loss": 4.745, "step": 831800 }, { "epoch": 5.015918766054438, "grad_norm": 2.986208915710449, "learning_rate": 5.591065711174428e-05, "loss": 4.7557, "step": 831850 }, { "epoch": 5.016220257835771, "grad_norm": 3.524904251098633, "learning_rate": 5.589481003894791e-05, "loss": 4.63, "step": 831900 }, { "epoch": 5.016521749617105, "grad_norm": 0.9759164452552795, "learning_rate": 5.587896469802332e-05, "loss": 4.5908, "step": 831950 }, { "epoch": 5.0168232413984395, "grad_norm": 3.684720754623413, "learning_rate": 5.586312108926201e-05, "loss": 4.6307, "step": 832000 }, { "epoch": 5.017124733179774, "grad_norm": 2.5134732723236084, "learning_rate": 5.584727921295561e-05, "loss": 4.7527, "step": 832050 }, { "epoch": 5.017426224961108, "grad_norm": 2.946331262588501, "learning_rate": 5.583143906939574e-05, "loss": 4.3268, "step": 832100 }, { "epoch": 5.017727716742441, "grad_norm": 1.259240984916687, "learning_rate": 5.5815600658873836e-05, "loss": 4.4285, "step": 832150 }, { "epoch": 5.018029208523775, "grad_norm": 4.084100246429443, "learning_rate": 5.5799763981681336e-05, "loss": 4.6279, "step": 832200 }, { "epoch": 5.01833070030511, "grad_norm": 4.001684665679932, "learning_rate": 5.578392903810983e-05, "loss": 4.908, "step": 832250 }, { "epoch": 5.018632192086444, "grad_norm": 3.2032668590545654, "learning_rate": 5.5768095828450574e-05, "loss": 4.4242, "step": 832300 }, { "epoch": 5.018933683867778, "grad_norm": 3.7336294651031494, "learning_rate": 5.575226435299504e-05, "loss": 4.5765, "step": 832350 }, { "epoch": 5.019235175649111, "grad_norm": 3.454178810119629, "learning_rate": 5.573643461203465e-05, "loss": 4.7772, "step": 832400 }, { "epoch": 5.019536667430446, "grad_norm": 2.570815086364746, "learning_rate": 5.5720606605860596e-05, "loss": 4.4156, "step": 832450 }, { "epoch": 5.01983815921178, "grad_norm": 3.4190380573272705, "learning_rate": 5.570478033476424e-05, "loss": 4.808, "step": 832500 }, { "epoch": 5.020139650993114, "grad_norm": 3.0790677070617676, "learning_rate": 5.5688955799036894e-05, "loss": 4.7935, "step": 832550 }, { "epoch": 5.020441142774448, "grad_norm": 4.031071186065674, "learning_rate": 5.5673132998969714e-05, "loss": 4.5007, "step": 832600 }, { "epoch": 5.020742634555782, "grad_norm": 0.8970481157302856, "learning_rate": 5.5657311934853855e-05, "loss": 4.9371, "step": 832650 }, { "epoch": 5.021044126337116, "grad_norm": 5.1500959396362305, "learning_rate": 5.5641492606980595e-05, "loss": 4.5161, "step": 832700 }, { "epoch": 5.02134561811845, "grad_norm": 1.3713879585266113, "learning_rate": 5.562567501564096e-05, "loss": 4.5294, "step": 832750 }, { "epoch": 5.021647109899784, "grad_norm": 3.7153613567352295, "learning_rate": 5.560985916112609e-05, "loss": 4.4631, "step": 832800 }, { "epoch": 5.021948601681118, "grad_norm": 3.6239328384399414, "learning_rate": 5.559404504372711e-05, "loss": 4.4845, "step": 832850 }, { "epoch": 5.0222500934624525, "grad_norm": 3.240245819091797, "learning_rate": 5.557823266373497e-05, "loss": 4.8817, "step": 832900 }, { "epoch": 5.022551585243786, "grad_norm": 3.465485095977783, "learning_rate": 5.5562422021440744e-05, "loss": 4.6333, "step": 832950 }, { "epoch": 5.02285307702512, "grad_norm": 2.5928683280944824, "learning_rate": 5.5546613117135376e-05, "loss": 4.3022, "step": 833000 }, { "epoch": 5.023154568806454, "grad_norm": 4.413182735443115, "learning_rate": 5.5530805951109756e-05, "loss": 4.6635, "step": 833050 }, { "epoch": 5.0234560605877885, "grad_norm": 1.7434709072113037, "learning_rate": 5.551500052365483e-05, "loss": 4.4111, "step": 833100 }, { "epoch": 5.023757552369123, "grad_norm": 2.2130484580993652, "learning_rate": 5.549919683506154e-05, "loss": 4.6986, "step": 833150 }, { "epoch": 5.024059044150457, "grad_norm": 3.954529285430908, "learning_rate": 5.548339488562062e-05, "loss": 4.4177, "step": 833200 }, { "epoch": 5.02436053593179, "grad_norm": 2.979449987411499, "learning_rate": 5.546759467562292e-05, "loss": 4.607, "step": 833250 }, { "epoch": 5.024662027713124, "grad_norm": 2.8270468711853027, "learning_rate": 5.54517962053593e-05, "loss": 4.7843, "step": 833300 }, { "epoch": 5.024963519494459, "grad_norm": 2.7525391578674316, "learning_rate": 5.543599947512045e-05, "loss": 4.6947, "step": 833350 }, { "epoch": 5.025265011275793, "grad_norm": 4.8561906814575195, "learning_rate": 5.542020448519701e-05, "loss": 4.4112, "step": 833400 }, { "epoch": 5.025566503057127, "grad_norm": 3.5338258743286133, "learning_rate": 5.5404411235879795e-05, "loss": 4.8861, "step": 833450 }, { "epoch": 5.02586799483846, "grad_norm": 3.3080849647521973, "learning_rate": 5.5388619727459345e-05, "loss": 4.333, "step": 833500 }, { "epoch": 5.026169486619795, "grad_norm": 2.8361806869506836, "learning_rate": 5.5372829960226336e-05, "loss": 4.4238, "step": 833550 }, { "epoch": 5.026470978401129, "grad_norm": 3.248467206954956, "learning_rate": 5.5357041934471404e-05, "loss": 4.6714, "step": 833600 }, { "epoch": 5.026772470182463, "grad_norm": 6.211472034454346, "learning_rate": 5.5341255650484985e-05, "loss": 4.5073, "step": 833650 }, { "epoch": 5.027073961963797, "grad_norm": 4.3372602462768555, "learning_rate": 5.532547110855768e-05, "loss": 4.5435, "step": 833700 }, { "epoch": 5.0273754537451305, "grad_norm": 5.416399002075195, "learning_rate": 5.530968830898002e-05, "loss": 4.2522, "step": 833750 }, { "epoch": 5.027676945526465, "grad_norm": 2.9822285175323486, "learning_rate": 5.529390725204243e-05, "loss": 4.5634, "step": 833800 }, { "epoch": 5.027978437307799, "grad_norm": 3.055920124053955, "learning_rate": 5.5278127938035256e-05, "loss": 4.7527, "step": 833850 }, { "epoch": 5.028279929089133, "grad_norm": 4.712186813354492, "learning_rate": 5.526235036724898e-05, "loss": 4.2207, "step": 833900 }, { "epoch": 5.028581420870467, "grad_norm": 3.9649624824523926, "learning_rate": 5.524657453997397e-05, "loss": 4.8186, "step": 833950 }, { "epoch": 5.0288829126518015, "grad_norm": 4.2561798095703125, "learning_rate": 5.5230800456500506e-05, "loss": 4.5795, "step": 834000 }, { "epoch": 5.0288829126518015, "eval_loss": 5.034439563751221, "eval_runtime": 39.092, "eval_samples_per_second": 13.097, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.837477411372098, "step": 834000 }, { "epoch": 5.029184404433135, "grad_norm": 3.6890199184417725, "learning_rate": 5.521502811711892e-05, "loss": 4.4534, "step": 834050 }, { "epoch": 5.029485896214469, "grad_norm": 4.89108419418335, "learning_rate": 5.519925752211952e-05, "loss": 4.7457, "step": 834100 }, { "epoch": 5.029787387995803, "grad_norm": 5.041118144989014, "learning_rate": 5.5183488671792495e-05, "loss": 4.4651, "step": 834150 }, { "epoch": 5.0300888797771375, "grad_norm": 3.100306272506714, "learning_rate": 5.5167721566428e-05, "loss": 4.6573, "step": 834200 }, { "epoch": 5.030390371558472, "grad_norm": 3.6446690559387207, "learning_rate": 5.515195620631632e-05, "loss": 3.9708, "step": 834250 }, { "epoch": 5.030691863339805, "grad_norm": 2.5047879219055176, "learning_rate": 5.5136192591747456e-05, "loss": 4.8526, "step": 834300 }, { "epoch": 5.030993355121139, "grad_norm": 3.7997047901153564, "learning_rate": 5.5120430723011613e-05, "loss": 4.6619, "step": 834350 }, { "epoch": 5.031294846902473, "grad_norm": 3.1643428802490234, "learning_rate": 5.510467060039887e-05, "loss": 4.563, "step": 834400 }, { "epoch": 5.031596338683808, "grad_norm": 4.296616077423096, "learning_rate": 5.5088912224199206e-05, "loss": 4.7932, "step": 834450 }, { "epoch": 5.031897830465142, "grad_norm": 3.4514222145080566, "learning_rate": 5.507315559470272e-05, "loss": 4.5131, "step": 834500 }, { "epoch": 5.032199322246475, "grad_norm": 3.0994720458984375, "learning_rate": 5.505740071219933e-05, "loss": 4.8733, "step": 834550 }, { "epoch": 5.032500814027809, "grad_norm": 2.735811471939087, "learning_rate": 5.504164757697894e-05, "loss": 4.4766, "step": 834600 }, { "epoch": 5.032802305809144, "grad_norm": 3.2951529026031494, "learning_rate": 5.502589618933151e-05, "loss": 4.7217, "step": 834650 }, { "epoch": 5.033103797590478, "grad_norm": 3.293856620788574, "learning_rate": 5.5010146549546976e-05, "loss": 4.5812, "step": 834700 }, { "epoch": 5.033405289371812, "grad_norm": 3.153035879135132, "learning_rate": 5.499439865791509e-05, "loss": 4.7285, "step": 834750 }, { "epoch": 5.033706781153146, "grad_norm": 2.7289788722991943, "learning_rate": 5.4978652514725725e-05, "loss": 4.2288, "step": 834800 }, { "epoch": 5.0340082729344795, "grad_norm": 2.940601348876953, "learning_rate": 5.49629081202687e-05, "loss": 4.5394, "step": 834850 }, { "epoch": 5.034309764715814, "grad_norm": 3.0846526622772217, "learning_rate": 5.494716547483367e-05, "loss": 4.3847, "step": 834900 }, { "epoch": 5.034611256497148, "grad_norm": 1.5112268924713135, "learning_rate": 5.493142457871047e-05, "loss": 4.385, "step": 834950 }, { "epoch": 5.034912748278482, "grad_norm": 2.6352975368499756, "learning_rate": 5.4915685432188735e-05, "loss": 4.48, "step": 835000 }, { "epoch": 5.035214240059816, "grad_norm": 3.7660858631134033, "learning_rate": 5.489994803555806e-05, "loss": 4.3628, "step": 835050 }, { "epoch": 5.03551573184115, "grad_norm": 2.7034506797790527, "learning_rate": 5.488421238910812e-05, "loss": 4.6218, "step": 835100 }, { "epoch": 5.035817223622484, "grad_norm": 4.1365790367126465, "learning_rate": 5.4868478493128573e-05, "loss": 4.7042, "step": 835150 }, { "epoch": 5.036118715403818, "grad_norm": 3.638810157775879, "learning_rate": 5.485274634790885e-05, "loss": 4.5192, "step": 835200 }, { "epoch": 5.036420207185152, "grad_norm": 2.7426230907440186, "learning_rate": 5.483701595373858e-05, "loss": 4.6036, "step": 835250 }, { "epoch": 5.0367216989664865, "grad_norm": 2.8771653175354004, "learning_rate": 5.482128731090726e-05, "loss": 4.4523, "step": 835300 }, { "epoch": 5.03702319074782, "grad_norm": 4.774962902069092, "learning_rate": 5.480556041970432e-05, "loss": 4.52, "step": 835350 }, { "epoch": 5.037324682529154, "grad_norm": 3.7582485675811768, "learning_rate": 5.4789835280419145e-05, "loss": 4.7808, "step": 835400 }, { "epoch": 5.037626174310488, "grad_norm": 2.438110828399658, "learning_rate": 5.4774111893341226e-05, "loss": 4.7977, "step": 835450 }, { "epoch": 5.037927666091822, "grad_norm": 2.8038673400878906, "learning_rate": 5.4758390258759824e-05, "loss": 4.5099, "step": 835500 }, { "epoch": 5.038229157873157, "grad_norm": 3.8212435245513916, "learning_rate": 5.474267037696435e-05, "loss": 4.9122, "step": 835550 }, { "epoch": 5.038530649654491, "grad_norm": 5.9871907234191895, "learning_rate": 5.472695224824413e-05, "loss": 4.2039, "step": 835600 }, { "epoch": 5.038832141435824, "grad_norm": 3.8822290897369385, "learning_rate": 5.4711235872888324e-05, "loss": 4.9217, "step": 835650 }, { "epoch": 5.039133633217158, "grad_norm": 4.359978199005127, "learning_rate": 5.4695521251186315e-05, "loss": 4.7524, "step": 835700 }, { "epoch": 5.0394351249984926, "grad_norm": 2.018404960632324, "learning_rate": 5.4679808383427224e-05, "loss": 4.4607, "step": 835750 }, { "epoch": 5.039736616779827, "grad_norm": 3.9344289302825928, "learning_rate": 5.4664097269900166e-05, "loss": 4.8995, "step": 835800 }, { "epoch": 5.040038108561161, "grad_norm": 3.4835352897644043, "learning_rate": 5.464838791089437e-05, "loss": 4.5079, "step": 835850 }, { "epoch": 5.040339600342494, "grad_norm": 3.469341516494751, "learning_rate": 5.463268030669894e-05, "loss": 4.2647, "step": 835900 }, { "epoch": 5.0406410921238285, "grad_norm": 4.248530864715576, "learning_rate": 5.461697445760291e-05, "loss": 4.2392, "step": 835950 }, { "epoch": 5.040942583905163, "grad_norm": 2.834102153778076, "learning_rate": 5.4601270363895335e-05, "loss": 4.5196, "step": 836000 }, { "epoch": 5.041244075686497, "grad_norm": 3.4417884349823, "learning_rate": 5.458556802586529e-05, "loss": 4.444, "step": 836050 }, { "epoch": 5.041545567467831, "grad_norm": 2.7319135665893555, "learning_rate": 5.456986744380164e-05, "loss": 4.4131, "step": 836100 }, { "epoch": 5.041847059249164, "grad_norm": 2.963083028793335, "learning_rate": 5.455416861799346e-05, "loss": 4.1652, "step": 836150 }, { "epoch": 5.042148551030499, "grad_norm": 3.101302146911621, "learning_rate": 5.45384715487296e-05, "loss": 4.3953, "step": 836200 }, { "epoch": 5.042450042811833, "grad_norm": 3.4427335262298584, "learning_rate": 5.452277623629887e-05, "loss": 4.7621, "step": 836250 }, { "epoch": 5.042751534593167, "grad_norm": 3.0536062717437744, "learning_rate": 5.4507082680990206e-05, "loss": 4.4976, "step": 836300 }, { "epoch": 5.043053026374501, "grad_norm": 2.120368719100952, "learning_rate": 5.449139088309246e-05, "loss": 4.5759, "step": 836350 }, { "epoch": 5.0433545181558355, "grad_norm": 3.0717499256134033, "learning_rate": 5.447570084289431e-05, "loss": 4.3443, "step": 836400 }, { "epoch": 5.043656009937169, "grad_norm": 3.317685842514038, "learning_rate": 5.4460012560684565e-05, "loss": 4.2338, "step": 836450 }, { "epoch": 5.043957501718503, "grad_norm": 2.9831533432006836, "learning_rate": 5.4444326036752e-05, "loss": 4.8267, "step": 836500 }, { "epoch": 5.044258993499837, "grad_norm": 2.253767490386963, "learning_rate": 5.4428641271385245e-05, "loss": 4.2223, "step": 836550 }, { "epoch": 5.044560485281171, "grad_norm": 4.682546615600586, "learning_rate": 5.4412958264872916e-05, "loss": 4.5834, "step": 836600 }, { "epoch": 5.044861977062506, "grad_norm": 3.8458242416381836, "learning_rate": 5.439727701750373e-05, "loss": 4.4962, "step": 836650 }, { "epoch": 5.045163468843839, "grad_norm": 3.064101219177246, "learning_rate": 5.438159752956617e-05, "loss": 4.0075, "step": 836700 }, { "epoch": 5.045464960625173, "grad_norm": 2.674828290939331, "learning_rate": 5.4365919801348855e-05, "loss": 4.5798, "step": 836750 }, { "epoch": 5.045766452406507, "grad_norm": 3.119628667831421, "learning_rate": 5.4350243833140365e-05, "loss": 4.9839, "step": 836800 }, { "epoch": 5.0460679441878415, "grad_norm": 2.953488349914551, "learning_rate": 5.4334569625229084e-05, "loss": 4.4489, "step": 836850 }, { "epoch": 5.046369435969176, "grad_norm": 1.611297607421875, "learning_rate": 5.431889717790358e-05, "loss": 4.4774, "step": 836900 }, { "epoch": 5.04667092775051, "grad_norm": 5.515221118927002, "learning_rate": 5.430322649145217e-05, "loss": 4.8305, "step": 836950 }, { "epoch": 5.046972419531843, "grad_norm": 4.821357727050781, "learning_rate": 5.428755756616336e-05, "loss": 4.4195, "step": 837000 }, { "epoch": 5.046972419531843, "eval_loss": 5.02211332321167, "eval_runtime": 39.1822, "eval_samples_per_second": 13.067, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.8723726387379775, "step": 837000 }, { "epoch": 5.0472739113131775, "grad_norm": 3.608955144882202, "learning_rate": 5.427189040232542e-05, "loss": 4.629, "step": 837050 }, { "epoch": 5.047575403094512, "grad_norm": 2.7994396686553955, "learning_rate": 5.425622500022679e-05, "loss": 4.4334, "step": 837100 }, { "epoch": 5.047876894875846, "grad_norm": 4.225261688232422, "learning_rate": 5.424056136015564e-05, "loss": 4.427, "step": 837150 }, { "epoch": 5.04817838665718, "grad_norm": 2.620293378829956, "learning_rate": 5.422489948240031e-05, "loss": 4.1362, "step": 837200 }, { "epoch": 5.048479878438513, "grad_norm": 3.0225982666015625, "learning_rate": 5.420923936724907e-05, "loss": 4.7393, "step": 837250 }, { "epoch": 5.048781370219848, "grad_norm": 2.848809242248535, "learning_rate": 5.419358101499003e-05, "loss": 4.3934, "step": 837300 }, { "epoch": 5.049082862001182, "grad_norm": 2.3036718368530273, "learning_rate": 5.417792442591144e-05, "loss": 4.4069, "step": 837350 }, { "epoch": 5.049384353782516, "grad_norm": 3.1168227195739746, "learning_rate": 5.4162269600301374e-05, "loss": 4.8982, "step": 837400 }, { "epoch": 5.04968584556385, "grad_norm": 4.182187080383301, "learning_rate": 5.414661653844803e-05, "loss": 4.8033, "step": 837450 }, { "epoch": 5.049987337345184, "grad_norm": 3.242894172668457, "learning_rate": 5.413096524063935e-05, "loss": 4.3431, "step": 837500 }, { "epoch": 5.050288829126518, "grad_norm": 3.564124345779419, "learning_rate": 5.411531570716344e-05, "loss": 4.5577, "step": 837550 }, { "epoch": 5.050590320907852, "grad_norm": 4.895617961883545, "learning_rate": 5.4099667938308373e-05, "loss": 4.5842, "step": 837600 }, { "epoch": 5.050891812689186, "grad_norm": 5.9477996826171875, "learning_rate": 5.408402193436201e-05, "loss": 4.4915, "step": 837650 }, { "epoch": 5.05119330447052, "grad_norm": 2.901010274887085, "learning_rate": 5.406837769561237e-05, "loss": 4.2163, "step": 837700 }, { "epoch": 5.051494796251855, "grad_norm": 3.057107925415039, "learning_rate": 5.405273522234737e-05, "loss": 4.8834, "step": 837750 }, { "epoch": 5.051796288033188, "grad_norm": 4.5669074058532715, "learning_rate": 5.4037094514854776e-05, "loss": 3.7377, "step": 837800 }, { "epoch": 5.052097779814522, "grad_norm": 3.799950361251831, "learning_rate": 5.4021455573422536e-05, "loss": 4.6172, "step": 837850 }, { "epoch": 5.052399271595856, "grad_norm": 2.6766200065612793, "learning_rate": 5.400581839833847e-05, "loss": 4.6845, "step": 837900 }, { "epoch": 5.0527007633771905, "grad_norm": 2.854377508163452, "learning_rate": 5.3990182989890286e-05, "loss": 4.7058, "step": 837950 }, { "epoch": 5.053002255158525, "grad_norm": 4.289015293121338, "learning_rate": 5.397454934836577e-05, "loss": 4.5641, "step": 838000 }, { "epoch": 5.053303746939858, "grad_norm": 4.248662948608398, "learning_rate": 5.3958917474052684e-05, "loss": 4.9184, "step": 838050 }, { "epoch": 5.053605238721192, "grad_norm": 1.9767751693725586, "learning_rate": 5.394328736723861e-05, "loss": 4.3996, "step": 838100 }, { "epoch": 5.0539067305025265, "grad_norm": 4.538000106811523, "learning_rate": 5.39276590282113e-05, "loss": 3.9487, "step": 838150 }, { "epoch": 5.054208222283861, "grad_norm": 4.703390598297119, "learning_rate": 5.391203245725833e-05, "loss": 4.3487, "step": 838200 }, { "epoch": 5.054509714065195, "grad_norm": 3.3309409618377686, "learning_rate": 5.389640765466723e-05, "loss": 4.6913, "step": 838250 }, { "epoch": 5.054811205846528, "grad_norm": 2.8367061614990234, "learning_rate": 5.3880784620725595e-05, "loss": 4.2616, "step": 838300 }, { "epoch": 5.055112697627862, "grad_norm": 3.0710644721984863, "learning_rate": 5.386516335572099e-05, "loss": 4.6511, "step": 838350 }, { "epoch": 5.055414189409197, "grad_norm": 4.2205963134765625, "learning_rate": 5.384954385994083e-05, "loss": 4.8615, "step": 838400 }, { "epoch": 5.055715681190531, "grad_norm": 2.7294721603393555, "learning_rate": 5.3833926133672584e-05, "loss": 4.6145, "step": 838450 }, { "epoch": 5.056017172971865, "grad_norm": 2.7152936458587646, "learning_rate": 5.381831017720376e-05, "loss": 4.9865, "step": 838500 }, { "epoch": 5.056318664753199, "grad_norm": 2.79290509223938, "learning_rate": 5.380269599082167e-05, "loss": 4.3264, "step": 838550 }, { "epoch": 5.056620156534533, "grad_norm": 3.7500734329223633, "learning_rate": 5.378708357481362e-05, "loss": 4.4735, "step": 838600 }, { "epoch": 5.056921648315867, "grad_norm": 4.524226188659668, "learning_rate": 5.377147292946706e-05, "loss": 4.3902, "step": 838650 }, { "epoch": 5.057223140097201, "grad_norm": 4.437538146972656, "learning_rate": 5.375586405506915e-05, "loss": 4.7752, "step": 838700 }, { "epoch": 5.057524631878535, "grad_norm": 4.9388532638549805, "learning_rate": 5.3740256951907224e-05, "loss": 4.873, "step": 838750 }, { "epoch": 5.057826123659869, "grad_norm": 4.8580756187438965, "learning_rate": 5.372465162026854e-05, "loss": 4.5645, "step": 838800 }, { "epoch": 5.058127615441203, "grad_norm": 3.8445377349853516, "learning_rate": 5.3709048060440214e-05, "loss": 4.4105, "step": 838850 }, { "epoch": 5.058429107222537, "grad_norm": 3.271343231201172, "learning_rate": 5.369344627270948e-05, "loss": 4.5754, "step": 838900 }, { "epoch": 5.058730599003871, "grad_norm": 4.293556213378906, "learning_rate": 5.367784625736344e-05, "loss": 4.4392, "step": 838950 }, { "epoch": 5.059032090785205, "grad_norm": 4.325765609741211, "learning_rate": 5.366224801468913e-05, "loss": 4.7057, "step": 839000 }, { "epoch": 5.0593335825665395, "grad_norm": 4.64601993560791, "learning_rate": 5.364665154497366e-05, "loss": 4.4481, "step": 839050 }, { "epoch": 5.059635074347873, "grad_norm": 4.214298248291016, "learning_rate": 5.363105684850411e-05, "loss": 4.7395, "step": 839100 }, { "epoch": 5.059936566129207, "grad_norm": 2.667541980743408, "learning_rate": 5.3615463925567397e-05, "loss": 4.8151, "step": 839150 }, { "epoch": 5.060238057910541, "grad_norm": 4.998114585876465, "learning_rate": 5.359987277645051e-05, "loss": 5.0053, "step": 839200 }, { "epoch": 5.0605395496918755, "grad_norm": 1.703263759613037, "learning_rate": 5.3584283401440445e-05, "loss": 3.8869, "step": 839250 }, { "epoch": 5.06084104147321, "grad_norm": 4.028327941894531, "learning_rate": 5.356869580082402e-05, "loss": 4.6575, "step": 839300 }, { "epoch": 5.061142533254544, "grad_norm": 3.2384917736053467, "learning_rate": 5.355310997488816e-05, "loss": 4.4154, "step": 839350 }, { "epoch": 5.061444025035877, "grad_norm": 4.49855375289917, "learning_rate": 5.353752592391969e-05, "loss": 4.5672, "step": 839400 }, { "epoch": 5.061745516817211, "grad_norm": 4.505562782287598, "learning_rate": 5.352194364820533e-05, "loss": 4.0008, "step": 839450 }, { "epoch": 5.062047008598546, "grad_norm": 4.193091869354248, "learning_rate": 5.3506363148031935e-05, "loss": 4.6893, "step": 839500 }, { "epoch": 5.06234850037988, "grad_norm": 4.451463222503662, "learning_rate": 5.3490784423686255e-05, "loss": 4.5044, "step": 839550 }, { "epoch": 5.062649992161214, "grad_norm": 1.7180653810501099, "learning_rate": 5.347520747545492e-05, "loss": 4.4256, "step": 839600 }, { "epoch": 5.062951483942547, "grad_norm": 4.377798557281494, "learning_rate": 5.345963230362464e-05, "loss": 4.7156, "step": 839650 }, { "epoch": 5.0632529757238816, "grad_norm": 3.3388335704803467, "learning_rate": 5.344405890848211e-05, "loss": 4.6042, "step": 839700 }, { "epoch": 5.063554467505216, "grad_norm": 5.139294624328613, "learning_rate": 5.342848729031388e-05, "loss": 4.3843, "step": 839750 }, { "epoch": 5.06385595928655, "grad_norm": 2.1088273525238037, "learning_rate": 5.341291744940648e-05, "loss": 4.9426, "step": 839800 }, { "epoch": 5.064157451067884, "grad_norm": 3.3945930004119873, "learning_rate": 5.3397349386046553e-05, "loss": 4.6444, "step": 839850 }, { "epoch": 5.0644589428492175, "grad_norm": 2.4159443378448486, "learning_rate": 5.338178310052048e-05, "loss": 4.4405, "step": 839900 }, { "epoch": 5.064760434630552, "grad_norm": 3.361703395843506, "learning_rate": 5.336621859311483e-05, "loss": 4.6997, "step": 839950 }, { "epoch": 5.065061926411886, "grad_norm": 3.2836265563964844, "learning_rate": 5.3350655864116064e-05, "loss": 4.2111, "step": 840000 }, { "epoch": 5.065061926411886, "eval_loss": 5.029217720031738, "eval_runtime": 39.0185, "eval_samples_per_second": 13.122, "eval_steps_per_second": 6.561, "eval_tts_loss": 7.834626726628794, "step": 840000 }, { "epoch": 5.06536341819322, "grad_norm": 3.304859161376953, "learning_rate": 5.333509491381049e-05, "loss": 4.8079, "step": 840050 }, { "epoch": 5.065664909974554, "grad_norm": 3.223388671875, "learning_rate": 5.3319535742484605e-05, "loss": 4.3625, "step": 840100 }, { "epoch": 5.0659664017558885, "grad_norm": 2.6749935150146484, "learning_rate": 5.3303978350424696e-05, "loss": 4.3516, "step": 840150 }, { "epoch": 5.066267893537222, "grad_norm": 3.6796507835388184, "learning_rate": 5.328842273791703e-05, "loss": 4.3641, "step": 840200 }, { "epoch": 5.066569385318556, "grad_norm": 2.7527482509613037, "learning_rate": 5.3272868905247924e-05, "loss": 4.5711, "step": 840250 }, { "epoch": 5.06687087709989, "grad_norm": 3.475473403930664, "learning_rate": 5.3257316852703677e-05, "loss": 4.6177, "step": 840300 }, { "epoch": 5.0671723688812245, "grad_norm": 4.5638203620910645, "learning_rate": 5.3241766580570405e-05, "loss": 4.2498, "step": 840350 }, { "epoch": 5.067473860662559, "grad_norm": 4.053529739379883, "learning_rate": 5.322621808913432e-05, "loss": 4.4323, "step": 840400 }, { "epoch": 5.067775352443892, "grad_norm": 3.706913948059082, "learning_rate": 5.321067137868166e-05, "loss": 4.2937, "step": 840450 }, { "epoch": 5.068076844225226, "grad_norm": 2.293999671936035, "learning_rate": 5.31951264494984e-05, "loss": 4.5607, "step": 840500 }, { "epoch": 5.06837833600656, "grad_norm": 0.7767123579978943, "learning_rate": 5.317958330187077e-05, "loss": 4.713, "step": 840550 }, { "epoch": 5.068679827787895, "grad_norm": 3.6026978492736816, "learning_rate": 5.316404193608472e-05, "loss": 4.3779, "step": 840600 }, { "epoch": 5.068981319569229, "grad_norm": 3.8042426109313965, "learning_rate": 5.314850235242623e-05, "loss": 4.1549, "step": 840650 }, { "epoch": 5.069282811350563, "grad_norm": 8.084146499633789, "learning_rate": 5.313296455118134e-05, "loss": 4.519, "step": 840700 }, { "epoch": 5.069584303131896, "grad_norm": 3.078885316848755, "learning_rate": 5.311742853263601e-05, "loss": 4.1995, "step": 840750 }, { "epoch": 5.0698857949132305, "grad_norm": 3.6529033184051514, "learning_rate": 5.310189429707619e-05, "loss": 4.5578, "step": 840800 }, { "epoch": 5.070187286694565, "grad_norm": 4.209873199462891, "learning_rate": 5.308636184478769e-05, "loss": 4.499, "step": 840850 }, { "epoch": 5.070488778475899, "grad_norm": 4.140028476715088, "learning_rate": 5.3070831176056427e-05, "loss": 4.3353, "step": 840900 }, { "epoch": 5.070790270257233, "grad_norm": 3.1450209617614746, "learning_rate": 5.305530229116821e-05, "loss": 4.8695, "step": 840950 }, { "epoch": 5.0710917620385665, "grad_norm": 4.207912445068359, "learning_rate": 5.303977519040874e-05, "loss": 4.6565, "step": 841000 }, { "epoch": 5.071393253819901, "grad_norm": 2.8104708194732666, "learning_rate": 5.302424987406385e-05, "loss": 4.7279, "step": 841050 }, { "epoch": 5.071694745601235, "grad_norm": 3.9296505451202393, "learning_rate": 5.300872634241928e-05, "loss": 4.3734, "step": 841100 }, { "epoch": 5.071996237382569, "grad_norm": 5.253769397735596, "learning_rate": 5.2993204595760654e-05, "loss": 4.4871, "step": 841150 }, { "epoch": 5.072297729163903, "grad_norm": 6.029629230499268, "learning_rate": 5.297768463437366e-05, "loss": 4.6968, "step": 841200 }, { "epoch": 5.072599220945237, "grad_norm": 3.8931336402893066, "learning_rate": 5.296216645854397e-05, "loss": 4.8023, "step": 841250 }, { "epoch": 5.072900712726571, "grad_norm": 3.469928741455078, "learning_rate": 5.2946650068557135e-05, "loss": 4.739, "step": 841300 }, { "epoch": 5.073202204507905, "grad_norm": 4.196062088012695, "learning_rate": 5.293113546469864e-05, "loss": 4.4689, "step": 841350 }, { "epoch": 5.073503696289239, "grad_norm": 4.823468208312988, "learning_rate": 5.2915622647254134e-05, "loss": 4.1364, "step": 841400 }, { "epoch": 5.0738051880705735, "grad_norm": 3.7894434928894043, "learning_rate": 5.290011161650898e-05, "loss": 4.1151, "step": 841450 }, { "epoch": 5.074106679851908, "grad_norm": 3.9992494583129883, "learning_rate": 5.2884602372748724e-05, "loss": 4.5457, "step": 841500 }, { "epoch": 5.074408171633241, "grad_norm": 2.0282201766967773, "learning_rate": 5.28690949162588e-05, "loss": 4.3099, "step": 841550 }, { "epoch": 5.074709663414575, "grad_norm": 3.679757833480835, "learning_rate": 5.285358924732454e-05, "loss": 4.5791, "step": 841600 }, { "epoch": 5.075011155195909, "grad_norm": 2.5088729858398438, "learning_rate": 5.283808536623133e-05, "loss": 4.6362, "step": 841650 }, { "epoch": 5.075312646977244, "grad_norm": 2.563318967819214, "learning_rate": 5.282258327326454e-05, "loss": 4.9599, "step": 841700 }, { "epoch": 5.075614138758578, "grad_norm": 3.1265995502471924, "learning_rate": 5.280708296870942e-05, "loss": 4.7444, "step": 841750 }, { "epoch": 5.075915630539911, "grad_norm": 2.916193962097168, "learning_rate": 5.27915844528512e-05, "loss": 4.0648, "step": 841800 }, { "epoch": 5.076217122321245, "grad_norm": 2.636091470718384, "learning_rate": 5.2776087725975184e-05, "loss": 4.6103, "step": 841850 }, { "epoch": 5.0765186141025795, "grad_norm": 2.863583564758301, "learning_rate": 5.276059278836649e-05, "loss": 4.337, "step": 841900 }, { "epoch": 5.076820105883914, "grad_norm": 3.198166608810425, "learning_rate": 5.2745099640310316e-05, "loss": 4.9206, "step": 841950 }, { "epoch": 5.077121597665248, "grad_norm": 2.2037131786346436, "learning_rate": 5.272960828209184e-05, "loss": 4.6856, "step": 842000 }, { "epoch": 5.077423089446581, "grad_norm": 3.381286144256592, "learning_rate": 5.271411871399606e-05, "loss": 4.6033, "step": 842050 }, { "epoch": 5.0777245812279155, "grad_norm": 3.621337652206421, "learning_rate": 5.269863093630814e-05, "loss": 4.7657, "step": 842100 }, { "epoch": 5.07802607300925, "grad_norm": 2.7864723205566406, "learning_rate": 5.268314494931306e-05, "loss": 4.3822, "step": 842150 }, { "epoch": 5.078327564790584, "grad_norm": 3.3143041133880615, "learning_rate": 5.266766075329576e-05, "loss": 4.259, "step": 842200 }, { "epoch": 5.078629056571918, "grad_norm": 4.739815711975098, "learning_rate": 5.265217834854129e-05, "loss": 4.3371, "step": 842250 }, { "epoch": 5.078930548353252, "grad_norm": 2.0120248794555664, "learning_rate": 5.263669773533458e-05, "loss": 4.3088, "step": 842300 }, { "epoch": 5.079232040134586, "grad_norm": 3.726879119873047, "learning_rate": 5.262121891396046e-05, "loss": 4.5806, "step": 842350 }, { "epoch": 5.07953353191592, "grad_norm": 2.8753576278686523, "learning_rate": 5.260574188470385e-05, "loss": 4.4945, "step": 842400 }, { "epoch": 5.079835023697254, "grad_norm": 5.022063255310059, "learning_rate": 5.2590266647849596e-05, "loss": 4.7425, "step": 842450 }, { "epoch": 5.080136515478588, "grad_norm": 3.2523767948150635, "learning_rate": 5.25747932036825e-05, "loss": 4.4348, "step": 842500 }, { "epoch": 5.0804380072599225, "grad_norm": 2.6602299213409424, "learning_rate": 5.2559321552487236e-05, "loss": 4.5834, "step": 842550 }, { "epoch": 5.080739499041256, "grad_norm": 4.04872989654541, "learning_rate": 5.254385169454866e-05, "loss": 4.3959, "step": 842600 }, { "epoch": 5.08104099082259, "grad_norm": 2.918849468231201, "learning_rate": 5.252838363015137e-05, "loss": 4.4806, "step": 842650 }, { "epoch": 5.081342482603924, "grad_norm": 3.6528971195220947, "learning_rate": 5.251291735958007e-05, "loss": 4.3442, "step": 842700 }, { "epoch": 5.081643974385258, "grad_norm": 4.893467903137207, "learning_rate": 5.249745288311946e-05, "loss": 4.5082, "step": 842750 }, { "epoch": 5.081945466166593, "grad_norm": 3.3058552742004395, "learning_rate": 5.2481990201054033e-05, "loss": 4.5575, "step": 842800 }, { "epoch": 5.082246957947926, "grad_norm": 3.139845371246338, "learning_rate": 5.246652931366841e-05, "loss": 4.6316, "step": 842850 }, { "epoch": 5.08254844972926, "grad_norm": 2.667879819869995, "learning_rate": 5.245107022124719e-05, "loss": 4.811, "step": 842900 }, { "epoch": 5.082849941510594, "grad_norm": 3.0683977603912354, "learning_rate": 5.243561292407479e-05, "loss": 4.8869, "step": 842950 }, { "epoch": 5.0831514332919285, "grad_norm": 3.2600297927856445, "learning_rate": 5.242015742243567e-05, "loss": 4.4277, "step": 843000 }, { "epoch": 5.0831514332919285, "eval_loss": 5.023134231567383, "eval_runtime": 38.9747, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.568, "eval_tts_loss": 7.840570562501685, "step": 843000 }, { "epoch": 5.083452925073263, "grad_norm": 3.9016692638397217, "learning_rate": 5.240470371661434e-05, "loss": 4.6689, "step": 843050 }, { "epoch": 5.083754416854597, "grad_norm": 2.8101987838745117, "learning_rate": 5.2389251806895135e-05, "loss": 4.7103, "step": 843100 }, { "epoch": 5.08405590863593, "grad_norm": 2.974557638168335, "learning_rate": 5.237380169356243e-05, "loss": 4.368, "step": 843150 }, { "epoch": 5.0843574004172645, "grad_norm": 1.3388638496398926, "learning_rate": 5.2358353376900633e-05, "loss": 3.9613, "step": 843200 }, { "epoch": 5.084658892198599, "grad_norm": 3.7507638931274414, "learning_rate": 5.234290685719396e-05, "loss": 3.921, "step": 843250 }, { "epoch": 5.084960383979933, "grad_norm": 4.239660739898682, "learning_rate": 5.232746213472676e-05, "loss": 4.5531, "step": 843300 }, { "epoch": 5.085261875761267, "grad_norm": 3.8969385623931885, "learning_rate": 5.231201920978324e-05, "loss": 4.4626, "step": 843350 }, { "epoch": 5.0855633675426, "grad_norm": 2.876664161682129, "learning_rate": 5.2296578082647556e-05, "loss": 4.8025, "step": 843400 }, { "epoch": 5.085864859323935, "grad_norm": 4.327976226806641, "learning_rate": 5.228113875360391e-05, "loss": 4.4807, "step": 843450 }, { "epoch": 5.086166351105269, "grad_norm": 3.029655694961548, "learning_rate": 5.22657012229365e-05, "loss": 4.7609, "step": 843500 }, { "epoch": 5.086467842886603, "grad_norm": 4.409794807434082, "learning_rate": 5.225026549092935e-05, "loss": 4.9425, "step": 843550 }, { "epoch": 5.086769334667937, "grad_norm": 3.648730516433716, "learning_rate": 5.2234831557866566e-05, "loss": 4.6524, "step": 843600 }, { "epoch": 5.0870708264492706, "grad_norm": 3.5150768756866455, "learning_rate": 5.221939942403223e-05, "loss": 4.407, "step": 843650 }, { "epoch": 5.087372318230605, "grad_norm": 3.196237325668335, "learning_rate": 5.2203969089710254e-05, "loss": 4.1882, "step": 843700 }, { "epoch": 5.087673810011939, "grad_norm": 2.8114137649536133, "learning_rate": 5.218854055518473e-05, "loss": 4.5892, "step": 843750 }, { "epoch": 5.087975301793273, "grad_norm": 3.063173294067383, "learning_rate": 5.217311382073951e-05, "loss": 4.8532, "step": 843800 }, { "epoch": 5.088276793574607, "grad_norm": 1.591347336769104, "learning_rate": 5.2157688886658495e-05, "loss": 4.1474, "step": 843850 }, { "epoch": 5.088578285355942, "grad_norm": 3.056333541870117, "learning_rate": 5.214226575322557e-05, "loss": 4.6858, "step": 843900 }, { "epoch": 5.088879777137275, "grad_norm": 3.2715821266174316, "learning_rate": 5.2126844420724664e-05, "loss": 4.5773, "step": 843950 }, { "epoch": 5.089181268918609, "grad_norm": 3.3813796043395996, "learning_rate": 5.211142488943944e-05, "loss": 4.6208, "step": 844000 }, { "epoch": 5.089482760699943, "grad_norm": 3.6831462383270264, "learning_rate": 5.209600715965376e-05, "loss": 4.4115, "step": 844050 }, { "epoch": 5.0897842524812775, "grad_norm": 3.4312996864318848, "learning_rate": 5.208059123165141e-05, "loss": 4.6971, "step": 844100 }, { "epoch": 5.090085744262612, "grad_norm": 1.8793085813522339, "learning_rate": 5.2065177105716035e-05, "loss": 4.5634, "step": 844150 }, { "epoch": 5.090387236043945, "grad_norm": 3.376796245574951, "learning_rate": 5.2049764782131256e-05, "loss": 4.9056, "step": 844200 }, { "epoch": 5.090688727825279, "grad_norm": 3.7033891677856445, "learning_rate": 5.203435426118078e-05, "loss": 4.1424, "step": 844250 }, { "epoch": 5.0909902196066135, "grad_norm": 3.298844337463379, "learning_rate": 5.2018945543148264e-05, "loss": 4.4782, "step": 844300 }, { "epoch": 5.091291711387948, "grad_norm": 4.688598155975342, "learning_rate": 5.200353862831716e-05, "loss": 4.6453, "step": 844350 }, { "epoch": 5.091593203169282, "grad_norm": 4.996551990509033, "learning_rate": 5.19881335169711e-05, "loss": 4.7169, "step": 844400 }, { "epoch": 5.091894694950616, "grad_norm": 2.6778318881988525, "learning_rate": 5.19727302093936e-05, "loss": 4.5553, "step": 844450 }, { "epoch": 5.092196186731949, "grad_norm": 4.12401008605957, "learning_rate": 5.1957328705868125e-05, "loss": 4.6796, "step": 844500 }, { "epoch": 5.092497678513284, "grad_norm": 2.2196218967437744, "learning_rate": 5.1941929006678046e-05, "loss": 4.0949, "step": 844550 }, { "epoch": 5.092799170294618, "grad_norm": 2.9127984046936035, "learning_rate": 5.192653111210689e-05, "loss": 4.6756, "step": 844600 }, { "epoch": 5.093100662075952, "grad_norm": 1.7083476781845093, "learning_rate": 5.19111350224379e-05, "loss": 4.6462, "step": 844650 }, { "epoch": 5.093402153857286, "grad_norm": 4.0071940422058105, "learning_rate": 5.18957407379545e-05, "loss": 4.6673, "step": 844700 }, { "epoch": 5.0937036456386195, "grad_norm": 3.043034791946411, "learning_rate": 5.188034825894003e-05, "loss": 4.5742, "step": 844750 }, { "epoch": 5.094005137419954, "grad_norm": 3.1610162258148193, "learning_rate": 5.1864957585677694e-05, "loss": 4.1168, "step": 844800 }, { "epoch": 5.094306629201288, "grad_norm": 5.331333160400391, "learning_rate": 5.1849568718450755e-05, "loss": 4.3721, "step": 844850 }, { "epoch": 5.094608120982622, "grad_norm": 3.4003732204437256, "learning_rate": 5.183418165754248e-05, "loss": 4.413, "step": 844900 }, { "epoch": 5.094909612763956, "grad_norm": 3.80961012840271, "learning_rate": 5.181879640323601e-05, "loss": 4.3192, "step": 844950 }, { "epoch": 5.09521110454529, "grad_norm": 2.8535141944885254, "learning_rate": 5.180341295581443e-05, "loss": 4.5155, "step": 845000 }, { "epoch": 5.095512596326624, "grad_norm": 2.040172576904297, "learning_rate": 5.178803131556095e-05, "loss": 3.9349, "step": 845050 }, { "epoch": 5.095814088107958, "grad_norm": 4.114627361297607, "learning_rate": 5.177265148275856e-05, "loss": 4.6159, "step": 845100 }, { "epoch": 5.096115579889292, "grad_norm": 4.566989421844482, "learning_rate": 5.175727345769033e-05, "loss": 4.481, "step": 845150 }, { "epoch": 5.0964170716706265, "grad_norm": 3.661912441253662, "learning_rate": 5.174189724063932e-05, "loss": 4.6109, "step": 845200 }, { "epoch": 5.096718563451961, "grad_norm": 3.870042324066162, "learning_rate": 5.172652283188843e-05, "loss": 4.8312, "step": 845250 }, { "epoch": 5.097020055233294, "grad_norm": 2.8405263423919678, "learning_rate": 5.17111502317207e-05, "loss": 4.1124, "step": 845300 }, { "epoch": 5.097321547014628, "grad_norm": 2.832364559173584, "learning_rate": 5.1695779440418985e-05, "loss": 4.4782, "step": 845350 }, { "epoch": 5.0976230387959625, "grad_norm": 3.0236427783966064, "learning_rate": 5.168041045826611e-05, "loss": 4.7236, "step": 845400 }, { "epoch": 5.097924530577297, "grad_norm": 1.8876746892929077, "learning_rate": 5.166504328554495e-05, "loss": 4.1415, "step": 845450 }, { "epoch": 5.098226022358631, "grad_norm": 3.6714863777160645, "learning_rate": 5.1649677922538404e-05, "loss": 4.27, "step": 845500 }, { "epoch": 5.098527514139964, "grad_norm": 5.5856404304504395, "learning_rate": 5.163431436952914e-05, "loss": 4.445, "step": 845550 }, { "epoch": 5.098829005921298, "grad_norm": 4.041572093963623, "learning_rate": 5.161895262679994e-05, "loss": 4.9265, "step": 845600 }, { "epoch": 5.099130497702633, "grad_norm": 1.9692306518554688, "learning_rate": 5.1603592694633566e-05, "loss": 4.5699, "step": 845650 }, { "epoch": 5.099431989483967, "grad_norm": 4.485453128814697, "learning_rate": 5.158823457331264e-05, "loss": 4.4845, "step": 845700 }, { "epoch": 5.099733481265301, "grad_norm": 4.532214641571045, "learning_rate": 5.157287826311977e-05, "loss": 4.7401, "step": 845750 }, { "epoch": 5.100034973046634, "grad_norm": 3.6800451278686523, "learning_rate": 5.1557523764337675e-05, "loss": 4.4458, "step": 845800 }, { "epoch": 5.1003364648279685, "grad_norm": 2.74082350730896, "learning_rate": 5.154217107724881e-05, "loss": 4.5266, "step": 845850 }, { "epoch": 5.100637956609303, "grad_norm": 3.8386998176574707, "learning_rate": 5.1526820202135784e-05, "loss": 4.5102, "step": 845900 }, { "epoch": 5.100939448390637, "grad_norm": 2.6844282150268555, "learning_rate": 5.151147113928115e-05, "loss": 4.4989, "step": 845950 }, { "epoch": 5.101240940171971, "grad_norm": 2.454134941101074, "learning_rate": 5.149612388896729e-05, "loss": 4.4449, "step": 846000 }, { "epoch": 5.101240940171971, "eval_loss": 5.030444145202637, "eval_runtime": 38.852, "eval_samples_per_second": 13.178, "eval_steps_per_second": 6.589, "eval_tts_loss": 7.894688140374876, "step": 846000 }, { "epoch": 5.101542431953305, "grad_norm": 2.6058762073516846, "learning_rate": 5.14807784514767e-05, "loss": 4.4034, "step": 846050 }, { "epoch": 5.101843923734639, "grad_norm": 3.9381449222564697, "learning_rate": 5.146543482709184e-05, "loss": 4.7563, "step": 846100 }, { "epoch": 5.102145415515973, "grad_norm": 4.570047855377197, "learning_rate": 5.145009301609502e-05, "loss": 4.8778, "step": 846150 }, { "epoch": 5.102446907297307, "grad_norm": 4.401335716247559, "learning_rate": 5.143475301876856e-05, "loss": 4.4765, "step": 846200 }, { "epoch": 5.102748399078641, "grad_norm": 3.2058353424072266, "learning_rate": 5.141941483539487e-05, "loss": 4.318, "step": 846250 }, { "epoch": 5.1030498908599755, "grad_norm": 4.85622501373291, "learning_rate": 5.14040784662561e-05, "loss": 4.6752, "step": 846300 }, { "epoch": 5.103351382641309, "grad_norm": 3.4581332206726074, "learning_rate": 5.138874391163458e-05, "loss": 4.6399, "step": 846350 }, { "epoch": 5.103652874422643, "grad_norm": 2.5469460487365723, "learning_rate": 5.137341117181253e-05, "loss": 4.5332, "step": 846400 }, { "epoch": 5.103954366203977, "grad_norm": 3.2393689155578613, "learning_rate": 5.1358080247072075e-05, "loss": 4.7026, "step": 846450 }, { "epoch": 5.1042558579853115, "grad_norm": 3.164034843444824, "learning_rate": 5.134275113769542e-05, "loss": 4.4438, "step": 846500 }, { "epoch": 5.104557349766646, "grad_norm": 5.004789352416992, "learning_rate": 5.1327423843964644e-05, "loss": 4.6966, "step": 846550 }, { "epoch": 5.104858841547979, "grad_norm": 4.359361171722412, "learning_rate": 5.1312098366161766e-05, "loss": 4.6818, "step": 846600 }, { "epoch": 5.105160333329313, "grad_norm": 2.6655807495117188, "learning_rate": 5.129677470456888e-05, "loss": 4.1771, "step": 846650 }, { "epoch": 5.105461825110647, "grad_norm": 3.4255383014678955, "learning_rate": 5.1281452859468044e-05, "loss": 4.6275, "step": 846700 }, { "epoch": 5.105763316891982, "grad_norm": 2.8707237243652344, "learning_rate": 5.1266132831141153e-05, "loss": 4.701, "step": 846750 }, { "epoch": 5.106064808673316, "grad_norm": 6.041104793548584, "learning_rate": 5.1250814619870175e-05, "loss": 4.2477, "step": 846800 }, { "epoch": 5.10636630045465, "grad_norm": 3.9994800090789795, "learning_rate": 5.123549822593709e-05, "loss": 4.5215, "step": 846850 }, { "epoch": 5.106667792235983, "grad_norm": 3.3042922019958496, "learning_rate": 5.122018364962371e-05, "loss": 4.4241, "step": 846900 }, { "epoch": 5.1069692840173175, "grad_norm": 3.1424400806427, "learning_rate": 5.120487089121183e-05, "loss": 4.5789, "step": 846950 }, { "epoch": 5.107270775798652, "grad_norm": 4.103091239929199, "learning_rate": 5.1189559950983375e-05, "loss": 4.5808, "step": 847000 }, { "epoch": 5.107572267579986, "grad_norm": 3.4160027503967285, "learning_rate": 5.117425082921999e-05, "loss": 4.4959, "step": 847050 }, { "epoch": 5.10787375936132, "grad_norm": 3.222658395767212, "learning_rate": 5.115894352620351e-05, "loss": 4.5775, "step": 847100 }, { "epoch": 5.1081752511426535, "grad_norm": 4.033851146697998, "learning_rate": 5.1143638042215655e-05, "loss": 4.7009, "step": 847150 }, { "epoch": 5.108476742923988, "grad_norm": 3.862483501434326, "learning_rate": 5.1128334377538024e-05, "loss": 4.3475, "step": 847200 }, { "epoch": 5.108778234705322, "grad_norm": 4.4450459480285645, "learning_rate": 5.1113032532452294e-05, "loss": 4.5494, "step": 847250 }, { "epoch": 5.109079726486656, "grad_norm": 3.405351161956787, "learning_rate": 5.109773250724012e-05, "loss": 4.4875, "step": 847300 }, { "epoch": 5.10938121826799, "grad_norm": 3.618680953979492, "learning_rate": 5.108243430218306e-05, "loss": 4.5529, "step": 847350 }, { "epoch": 5.109682710049324, "grad_norm": 3.527554750442505, "learning_rate": 5.106713791756258e-05, "loss": 3.935, "step": 847400 }, { "epoch": 5.109984201830658, "grad_norm": 2.961820125579834, "learning_rate": 5.105184335366023e-05, "loss": 4.0331, "step": 847450 }, { "epoch": 5.110285693611992, "grad_norm": 3.080085515975952, "learning_rate": 5.103655061075757e-05, "loss": 4.798, "step": 847500 }, { "epoch": 5.110587185393326, "grad_norm": 1.0844818353652954, "learning_rate": 5.102125968913589e-05, "loss": 4.3129, "step": 847550 }, { "epoch": 5.11088867717466, "grad_norm": 3.8226263523101807, "learning_rate": 5.10059705890767e-05, "loss": 4.6756, "step": 847600 }, { "epoch": 5.111190168955995, "grad_norm": 2.1658644676208496, "learning_rate": 5.099068331086138e-05, "loss": 4.2299, "step": 847650 }, { "epoch": 5.111491660737328, "grad_norm": 6.036841869354248, "learning_rate": 5.0975397854771235e-05, "loss": 4.7959, "step": 847700 }, { "epoch": 5.111793152518662, "grad_norm": 3.508312225341797, "learning_rate": 5.096011422108755e-05, "loss": 4.5713, "step": 847750 }, { "epoch": 5.112094644299996, "grad_norm": 4.651446342468262, "learning_rate": 5.094483241009166e-05, "loss": 4.225, "step": 847800 }, { "epoch": 5.112396136081331, "grad_norm": 3.753606081008911, "learning_rate": 5.0929552422064717e-05, "loss": 4.4997, "step": 847850 }, { "epoch": 5.112697627862665, "grad_norm": 3.580209732055664, "learning_rate": 5.091427425728799e-05, "loss": 4.7241, "step": 847900 }, { "epoch": 5.112999119643998, "grad_norm": 2.9865269660949707, "learning_rate": 5.089899791604269e-05, "loss": 4.3535, "step": 847950 }, { "epoch": 5.113300611425332, "grad_norm": 2.5545380115509033, "learning_rate": 5.0883723398609875e-05, "loss": 4.5291, "step": 848000 }, { "epoch": 5.1136021032066665, "grad_norm": 2.668511390686035, "learning_rate": 5.0868450705270714e-05, "loss": 4.4979, "step": 848050 }, { "epoch": 5.113903594988001, "grad_norm": 3.451680898666382, "learning_rate": 5.085317983630619e-05, "loss": 4.7986, "step": 848100 }, { "epoch": 5.114205086769335, "grad_norm": 3.9379520416259766, "learning_rate": 5.083791079199746e-05, "loss": 4.3191, "step": 848150 }, { "epoch": 5.114506578550669, "grad_norm": 4.383261203765869, "learning_rate": 5.082264357262543e-05, "loss": 4.3311, "step": 848200 }, { "epoch": 5.1148080703320025, "grad_norm": 2.0865015983581543, "learning_rate": 5.080737817847117e-05, "loss": 4.3864, "step": 848250 }, { "epoch": 5.115109562113337, "grad_norm": 3.1010873317718506, "learning_rate": 5.0792114609815485e-05, "loss": 4.559, "step": 848300 }, { "epoch": 5.115411053894671, "grad_norm": 1.0776761770248413, "learning_rate": 5.077685286693936e-05, "loss": 4.652, "step": 848350 }, { "epoch": 5.115712545676005, "grad_norm": 3.3070740699768066, "learning_rate": 5.076159295012371e-05, "loss": 4.4968, "step": 848400 }, { "epoch": 5.116014037457339, "grad_norm": 4.1215901374816895, "learning_rate": 5.074633485964927e-05, "loss": 4.7391, "step": 848450 }, { "epoch": 5.116315529238673, "grad_norm": 2.9781999588012695, "learning_rate": 5.073107859579695e-05, "loss": 4.2119, "step": 848500 }, { "epoch": 5.116617021020007, "grad_norm": 3.0633814334869385, "learning_rate": 5.0715824158847455e-05, "loss": 4.8408, "step": 848550 }, { "epoch": 5.116918512801341, "grad_norm": 2.042135238647461, "learning_rate": 5.0700571549081486e-05, "loss": 4.5515, "step": 848600 }, { "epoch": 5.117220004582675, "grad_norm": 1.4808753728866577, "learning_rate": 5.0685320766779796e-05, "loss": 4.6362, "step": 848650 }, { "epoch": 5.117521496364009, "grad_norm": 3.117791175842285, "learning_rate": 5.0670071812223093e-05, "loss": 4.1946, "step": 848700 }, { "epoch": 5.117822988145343, "grad_norm": 2.5993528366088867, "learning_rate": 5.065482468569191e-05, "loss": 4.1252, "step": 848750 }, { "epoch": 5.118124479926677, "grad_norm": 3.414780378341675, "learning_rate": 5.063957938746691e-05, "loss": 4.3386, "step": 848800 }, { "epoch": 5.118425971708011, "grad_norm": 4.439657688140869, "learning_rate": 5.062433591782871e-05, "loss": 4.1349, "step": 848850 }, { "epoch": 5.118727463489345, "grad_norm": 2.1681556701660156, "learning_rate": 5.060909427705779e-05, "loss": 4.3925, "step": 848900 }, { "epoch": 5.11902895527068, "grad_norm": 1.9322630167007446, "learning_rate": 5.0593854465434594e-05, "loss": 4.3215, "step": 848950 }, { "epoch": 5.119330447052014, "grad_norm": 3.8179521560668945, "learning_rate": 5.057861648323972e-05, "loss": 4.5335, "step": 849000 }, { "epoch": 5.119330447052014, "eval_loss": 5.019659042358398, "eval_runtime": 39.0368, "eval_samples_per_second": 13.116, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.821965862172142, "step": 849000 }, { "epoch": 5.119631938833347, "grad_norm": 2.3029825687408447, "learning_rate": 5.0563380330753453e-05, "loss": 3.9529, "step": 849050 }, { "epoch": 5.119933430614681, "grad_norm": 3.1154654026031494, "learning_rate": 5.0548146008256265e-05, "loss": 4.759, "step": 849100 }, { "epoch": 5.1202349223960155, "grad_norm": 3.419945240020752, "learning_rate": 5.053291351602859e-05, "loss": 4.4928, "step": 849150 }, { "epoch": 5.12053641417735, "grad_norm": 3.939945697784424, "learning_rate": 5.051768285435064e-05, "loss": 4.5263, "step": 849200 }, { "epoch": 5.120837905958684, "grad_norm": 3.823791980743408, "learning_rate": 5.050245402350282e-05, "loss": 4.515, "step": 849250 }, { "epoch": 5.121139397740017, "grad_norm": 4.042923450469971, "learning_rate": 5.048722702376528e-05, "loss": 4.3878, "step": 849300 }, { "epoch": 5.1214408895213515, "grad_norm": 1.1724894046783447, "learning_rate": 5.047200185541838e-05, "loss": 4.5108, "step": 849350 }, { "epoch": 5.121742381302686, "grad_norm": 4.820743083953857, "learning_rate": 5.0456778518742195e-05, "loss": 4.5852, "step": 849400 }, { "epoch": 5.12204387308402, "grad_norm": 4.081212997436523, "learning_rate": 5.044155701401699e-05, "loss": 4.8418, "step": 849450 }, { "epoch": 5.122345364865354, "grad_norm": 4.626194477081299, "learning_rate": 5.042633734152281e-05, "loss": 4.152, "step": 849500 }, { "epoch": 5.122646856646687, "grad_norm": 2.553769588470459, "learning_rate": 5.041111950153979e-05, "loss": 4.5002, "step": 849550 }, { "epoch": 5.122948348428022, "grad_norm": 2.5617756843566895, "learning_rate": 5.039590349434804e-05, "loss": 4.239, "step": 849600 }, { "epoch": 5.123249840209356, "grad_norm": 3.610837459564209, "learning_rate": 5.0380689320227494e-05, "loss": 4.3696, "step": 849650 }, { "epoch": 5.12355133199069, "grad_norm": 3.7590131759643555, "learning_rate": 5.036547697945824e-05, "loss": 4.3683, "step": 849700 }, { "epoch": 5.123852823772024, "grad_norm": 3.8495733737945557, "learning_rate": 5.035026647232022e-05, "loss": 4.5006, "step": 849750 }, { "epoch": 5.124154315553358, "grad_norm": 2.706717014312744, "learning_rate": 5.033505779909326e-05, "loss": 4.3715, "step": 849800 }, { "epoch": 5.124455807334692, "grad_norm": 3.0297515392303467, "learning_rate": 5.031985096005735e-05, "loss": 4.3035, "step": 849850 }, { "epoch": 5.124757299116026, "grad_norm": 3.0203700065612793, "learning_rate": 5.0304645955492375e-05, "loss": 4.327, "step": 849900 }, { "epoch": 5.12505879089736, "grad_norm": 6.722344875335693, "learning_rate": 5.028944278567807e-05, "loss": 4.976, "step": 849950 }, { "epoch": 5.125360282678694, "grad_norm": 1.0601376295089722, "learning_rate": 5.027424145089428e-05, "loss": 4.0726, "step": 850000 }, { "epoch": 5.125661774460029, "grad_norm": 2.2024905681610107, "learning_rate": 5.0259041951420806e-05, "loss": 4.6669, "step": 850050 }, { "epoch": 5.125963266241362, "grad_norm": 3.091891288757324, "learning_rate": 5.024384428753733e-05, "loss": 4.338, "step": 850100 }, { "epoch": 5.126264758022696, "grad_norm": 1.6163164377212524, "learning_rate": 5.022864845952347e-05, "loss": 4.8328, "step": 850150 }, { "epoch": 5.12656624980403, "grad_norm": 1.687753677368164, "learning_rate": 5.021345446765904e-05, "loss": 4.2415, "step": 850200 }, { "epoch": 5.1268677415853645, "grad_norm": 4.003787994384766, "learning_rate": 5.019826231222349e-05, "loss": 4.5617, "step": 850250 }, { "epoch": 5.127169233366699, "grad_norm": 3.0574851036071777, "learning_rate": 5.018307199349653e-05, "loss": 4.524, "step": 850300 }, { "epoch": 5.127470725148032, "grad_norm": 4.155770301818848, "learning_rate": 5.0167883511757715e-05, "loss": 4.9288, "step": 850350 }, { "epoch": 5.127772216929366, "grad_norm": 2.0201175212860107, "learning_rate": 5.015269686728649e-05, "loss": 4.4795, "step": 850400 }, { "epoch": 5.1280737087107005, "grad_norm": 3.0574631690979004, "learning_rate": 5.013751206036239e-05, "loss": 4.4451, "step": 850450 }, { "epoch": 5.128375200492035, "grad_norm": 3.154890537261963, "learning_rate": 5.0122329091264926e-05, "loss": 4.6406, "step": 850500 }, { "epoch": 5.128676692273369, "grad_norm": 3.551640510559082, "learning_rate": 5.0107147960273455e-05, "loss": 4.488, "step": 850550 }, { "epoch": 5.128978184054703, "grad_norm": 0.690365731716156, "learning_rate": 5.009196866766731e-05, "loss": 4.2626, "step": 850600 }, { "epoch": 5.129279675836036, "grad_norm": 3.6104636192321777, "learning_rate": 5.007679121372597e-05, "loss": 4.711, "step": 850650 }, { "epoch": 5.129581167617371, "grad_norm": 1.3287426233291626, "learning_rate": 5.006161559872863e-05, "loss": 4.6145, "step": 850700 }, { "epoch": 5.129882659398705, "grad_norm": 3.9368059635162354, "learning_rate": 5.004644182295464e-05, "loss": 4.4761, "step": 850750 }, { "epoch": 5.130184151180039, "grad_norm": 1.4183045625686646, "learning_rate": 5.0031269886683256e-05, "loss": 4.3712, "step": 850800 }, { "epoch": 5.130485642961373, "grad_norm": 4.270002365112305, "learning_rate": 5.001609979019372e-05, "loss": 4.7503, "step": 850850 }, { "epoch": 5.1307871347427065, "grad_norm": 1.262454628944397, "learning_rate": 5.0000931533765194e-05, "loss": 4.3178, "step": 850900 }, { "epoch": 5.131088626524041, "grad_norm": 3.914260149002075, "learning_rate": 4.9985765117676765e-05, "loss": 4.4681, "step": 850950 }, { "epoch": 5.131390118305375, "grad_norm": 3.5204708576202393, "learning_rate": 4.997060054220765e-05, "loss": 4.6155, "step": 851000 }, { "epoch": 5.131691610086709, "grad_norm": 1.626630425453186, "learning_rate": 4.9955437807636836e-05, "loss": 4.0256, "step": 851050 }, { "epoch": 5.131993101868043, "grad_norm": 3.2038841247558594, "learning_rate": 4.994027691424341e-05, "loss": 4.3104, "step": 851100 }, { "epoch": 5.132294593649377, "grad_norm": 3.4520740509033203, "learning_rate": 4.992511786230644e-05, "loss": 4.1873, "step": 851150 }, { "epoch": 5.132596085430711, "grad_norm": 4.198259353637695, "learning_rate": 4.990996065210482e-05, "loss": 4.3965, "step": 851200 }, { "epoch": 5.132897577212045, "grad_norm": 4.244449138641357, "learning_rate": 4.9894805283917586e-05, "loss": 4.8078, "step": 851250 }, { "epoch": 5.133199068993379, "grad_norm": 5.542426586151123, "learning_rate": 4.987965175802361e-05, "loss": 4.5948, "step": 851300 }, { "epoch": 5.1335005607747135, "grad_norm": 4.940261363983154, "learning_rate": 4.986450007470171e-05, "loss": 4.7204, "step": 851350 }, { "epoch": 5.133802052556048, "grad_norm": 3.488121509552002, "learning_rate": 4.984935023423078e-05, "loss": 4.5271, "step": 851400 }, { "epoch": 5.134103544337381, "grad_norm": 2.359079599380493, "learning_rate": 4.9834202236889694e-05, "loss": 4.8632, "step": 851450 }, { "epoch": 5.134405036118715, "grad_norm": 2.8476240634918213, "learning_rate": 4.981905608295711e-05, "loss": 4.788, "step": 851500 }, { "epoch": 5.1347065279000494, "grad_norm": 3.4376442432403564, "learning_rate": 4.980391177271184e-05, "loss": 4.3441, "step": 851550 }, { "epoch": 5.135008019681384, "grad_norm": 2.8286380767822266, "learning_rate": 4.9788769306432636e-05, "loss": 4.843, "step": 851600 }, { "epoch": 5.135309511462718, "grad_norm": 3.7481327056884766, "learning_rate": 4.977362868439807e-05, "loss": 4.9747, "step": 851650 }, { "epoch": 5.135611003244051, "grad_norm": 2.1741912364959717, "learning_rate": 4.975848990688691e-05, "loss": 4.0776, "step": 851700 }, { "epoch": 5.135912495025385, "grad_norm": 3.9144842624664307, "learning_rate": 4.974335297417767e-05, "loss": 4.5365, "step": 851750 }, { "epoch": 5.13621398680672, "grad_norm": 3.290396213531494, "learning_rate": 4.972821788654889e-05, "loss": 4.4089, "step": 851800 }, { "epoch": 5.136515478588054, "grad_norm": 2.627375602722168, "learning_rate": 4.971308464427917e-05, "loss": 4.2756, "step": 851850 }, { "epoch": 5.136816970369388, "grad_norm": 3.017449378967285, "learning_rate": 4.9697953247647066e-05, "loss": 4.8791, "step": 851900 }, { "epoch": 5.137118462150722, "grad_norm": 0.6188549399375916, "learning_rate": 4.968282369693093e-05, "loss": 4.7358, "step": 851950 }, { "epoch": 5.1374199539320555, "grad_norm": 2.0754711627960205, "learning_rate": 4.966769599240927e-05, "loss": 4.6414, "step": 852000 }, { "epoch": 5.1374199539320555, "eval_loss": 5.009816646575928, "eval_runtime": 39.0482, "eval_samples_per_second": 13.112, "eval_steps_per_second": 6.556, "eval_tts_loss": 7.829591650874526, "step": 852000 }, { "epoch": 5.13772144571339, "grad_norm": 3.3062846660614014, "learning_rate": 4.9652570134360545e-05, "loss": 4.6934, "step": 852050 }, { "epoch": 5.138022937494724, "grad_norm": 3.7689669132232666, "learning_rate": 4.9637446123063065e-05, "loss": 4.3392, "step": 852100 }, { "epoch": 5.138324429276058, "grad_norm": 4.89033317565918, "learning_rate": 4.962232395879509e-05, "loss": 4.5229, "step": 852150 }, { "epoch": 5.138625921057392, "grad_norm": 3.275444984436035, "learning_rate": 4.960720364183507e-05, "loss": 4.7014, "step": 852200 }, { "epoch": 5.138927412838726, "grad_norm": 3.809314012527466, "learning_rate": 4.959208517246113e-05, "loss": 4.0909, "step": 852250 }, { "epoch": 5.13922890462006, "grad_norm": 3.471188545227051, "learning_rate": 4.957696855095158e-05, "loss": 4.494, "step": 852300 }, { "epoch": 5.139530396401394, "grad_norm": 3.442392110824585, "learning_rate": 4.956185377758467e-05, "loss": 4.7351, "step": 852350 }, { "epoch": 5.139831888182728, "grad_norm": 2.8473522663116455, "learning_rate": 4.954674085263844e-05, "loss": 4.46, "step": 852400 }, { "epoch": 5.1401333799640625, "grad_norm": 2.4320032596588135, "learning_rate": 4.953162977639116e-05, "loss": 4.8674, "step": 852450 }, { "epoch": 5.140434871745396, "grad_norm": 3.5173428058624268, "learning_rate": 4.9516520549120856e-05, "loss": 4.6629, "step": 852500 }, { "epoch": 5.14073636352673, "grad_norm": 3.7435221672058105, "learning_rate": 4.950141317110552e-05, "loss": 4.0001, "step": 852550 }, { "epoch": 5.141037855308064, "grad_norm": 4.172060012817383, "learning_rate": 4.948630764262329e-05, "loss": 4.5799, "step": 852600 }, { "epoch": 5.141339347089398, "grad_norm": 2.8866357803344727, "learning_rate": 4.947120396395216e-05, "loss": 4.8214, "step": 852650 }, { "epoch": 5.141640838870733, "grad_norm": 1.9755676984786987, "learning_rate": 4.9456102135369995e-05, "loss": 4.696, "step": 852700 }, { "epoch": 5.141942330652067, "grad_norm": 4.457348823547363, "learning_rate": 4.944100215715481e-05, "loss": 4.3007, "step": 852750 }, { "epoch": 5.1422438224334, "grad_norm": 2.8482909202575684, "learning_rate": 4.9425904029584504e-05, "loss": 4.5433, "step": 852800 }, { "epoch": 5.142545314214734, "grad_norm": 3.6011123657226562, "learning_rate": 4.941080775293687e-05, "loss": 4.8985, "step": 852850 }, { "epoch": 5.142846805996069, "grad_norm": 3.4616048336029053, "learning_rate": 4.9395713327489827e-05, "loss": 4.3125, "step": 852900 }, { "epoch": 5.143148297777403, "grad_norm": 2.5606541633605957, "learning_rate": 4.938062075352111e-05, "loss": 4.7314, "step": 852950 }, { "epoch": 5.143449789558737, "grad_norm": 4.8662261962890625, "learning_rate": 4.9365530031308414e-05, "loss": 4.1424, "step": 853000 }, { "epoch": 5.14375128134007, "grad_norm": 2.9729177951812744, "learning_rate": 4.935044116112952e-05, "loss": 4.3062, "step": 853050 }, { "epoch": 5.1440527731214045, "grad_norm": 0.7464814186096191, "learning_rate": 4.933535414326219e-05, "loss": 4.1256, "step": 853100 }, { "epoch": 5.144354264902739, "grad_norm": 4.707556247711182, "learning_rate": 4.9320268977983945e-05, "loss": 4.3514, "step": 853150 }, { "epoch": 5.144655756684073, "grad_norm": 3.936490058898926, "learning_rate": 4.930518566557248e-05, "loss": 4.7515, "step": 853200 }, { "epoch": 5.144957248465407, "grad_norm": 2.8967955112457275, "learning_rate": 4.9290104206305434e-05, "loss": 4.802, "step": 853250 }, { "epoch": 5.1452587402467405, "grad_norm": 3.072585344314575, "learning_rate": 4.9275024600460285e-05, "loss": 4.2855, "step": 853300 }, { "epoch": 5.145560232028075, "grad_norm": 3.25268292427063, "learning_rate": 4.925994684831451e-05, "loss": 4.4348, "step": 853350 }, { "epoch": 5.145861723809409, "grad_norm": 3.9694879055023193, "learning_rate": 4.9244870950145715e-05, "loss": 4.8264, "step": 853400 }, { "epoch": 5.146163215590743, "grad_norm": 4.272956848144531, "learning_rate": 4.9229796906231216e-05, "loss": 4.4357, "step": 853450 }, { "epoch": 5.146464707372077, "grad_norm": 3.521718740463257, "learning_rate": 4.921472471684851e-05, "loss": 4.6107, "step": 853500 }, { "epoch": 5.1467661991534115, "grad_norm": 2.6152029037475586, "learning_rate": 4.919965438227502e-05, "loss": 4.3739, "step": 853550 }, { "epoch": 5.147067690934745, "grad_norm": 2.1185200214385986, "learning_rate": 4.9184585902787985e-05, "loss": 4.6126, "step": 853600 }, { "epoch": 5.147369182716079, "grad_norm": 4.123730182647705, "learning_rate": 4.916951927866484e-05, "loss": 4.7224, "step": 853650 }, { "epoch": 5.147670674497413, "grad_norm": 2.9968605041503906, "learning_rate": 4.915445451018273e-05, "loss": 4.6822, "step": 853700 }, { "epoch": 5.147972166278747, "grad_norm": 3.212367296218872, "learning_rate": 4.9139391597619035e-05, "loss": 4.5132, "step": 853750 }, { "epoch": 5.148273658060082, "grad_norm": 2.5107433795928955, "learning_rate": 4.912433054125084e-05, "loss": 4.7162, "step": 853800 }, { "epoch": 5.148575149841415, "grad_norm": 4.40142297744751, "learning_rate": 4.9109271341355445e-05, "loss": 4.6146, "step": 853850 }, { "epoch": 5.148876641622749, "grad_norm": 3.8647027015686035, "learning_rate": 4.909421399820988e-05, "loss": 4.4645, "step": 853900 }, { "epoch": 5.149178133404083, "grad_norm": 3.933789014816284, "learning_rate": 4.90791585120913e-05, "loss": 4.5369, "step": 853950 }, { "epoch": 5.149479625185418, "grad_norm": 3.004333019256592, "learning_rate": 4.9064104883276844e-05, "loss": 4.6401, "step": 854000 }, { "epoch": 5.149781116966752, "grad_norm": 4.788760662078857, "learning_rate": 4.904905311204343e-05, "loss": 4.2743, "step": 854050 }, { "epoch": 5.150082608748085, "grad_norm": 3.9516937732696533, "learning_rate": 4.9034003198668196e-05, "loss": 4.7461, "step": 854100 }, { "epoch": 5.150384100529419, "grad_norm": 3.4538707733154297, "learning_rate": 4.901895514342803e-05, "loss": 4.847, "step": 854150 }, { "epoch": 5.1506855923107535, "grad_norm": 3.080568790435791, "learning_rate": 4.900390894659985e-05, "loss": 4.6772, "step": 854200 }, { "epoch": 5.150987084092088, "grad_norm": 3.3202004432678223, "learning_rate": 4.898886460846059e-05, "loss": 4.3207, "step": 854250 }, { "epoch": 5.151288575873422, "grad_norm": 2.828404664993286, "learning_rate": 4.8973822129287134e-05, "loss": 4.872, "step": 854300 }, { "epoch": 5.151590067654756, "grad_norm": 1.2157843112945557, "learning_rate": 4.895878150935635e-05, "loss": 4.7084, "step": 854350 }, { "epoch": 5.1518915594360895, "grad_norm": 1.5338855981826782, "learning_rate": 4.894374274894495e-05, "loss": 4.4122, "step": 854400 }, { "epoch": 5.152193051217424, "grad_norm": 3.6576344966888428, "learning_rate": 4.892870584832981e-05, "loss": 4.532, "step": 854450 }, { "epoch": 5.152494542998758, "grad_norm": 3.116206407546997, "learning_rate": 4.891367080778759e-05, "loss": 4.422, "step": 854500 }, { "epoch": 5.152796034780092, "grad_norm": 4.767117500305176, "learning_rate": 4.889863762759497e-05, "loss": 4.6546, "step": 854550 }, { "epoch": 5.153097526561426, "grad_norm": 3.812497854232788, "learning_rate": 4.888360630802863e-05, "loss": 4.5637, "step": 854600 }, { "epoch": 5.15339901834276, "grad_norm": 4.706997871398926, "learning_rate": 4.886857684936525e-05, "loss": 4.9845, "step": 854650 }, { "epoch": 5.153700510124094, "grad_norm": 4.493904113769531, "learning_rate": 4.885354925188138e-05, "loss": 4.6463, "step": 854700 }, { "epoch": 5.154002001905428, "grad_norm": 3.912102222442627, "learning_rate": 4.883852351585355e-05, "loss": 4.2449, "step": 854750 }, { "epoch": 5.154303493686762, "grad_norm": 3.336244583129883, "learning_rate": 4.882349964155841e-05, "loss": 4.635, "step": 854800 }, { "epoch": 5.154604985468096, "grad_norm": 3.1448538303375244, "learning_rate": 4.8808477629272305e-05, "loss": 4.9936, "step": 854850 }, { "epoch": 5.15490647724943, "grad_norm": 2.3979320526123047, "learning_rate": 4.8793457479271816e-05, "loss": 4.4879, "step": 854900 }, { "epoch": 5.155207969030764, "grad_norm": 3.0932812690734863, "learning_rate": 4.8778439191833325e-05, "loss": 4.1813, "step": 854950 }, { "epoch": 5.155509460812098, "grad_norm": 2.903174638748169, "learning_rate": 4.876342276723315e-05, "loss": 4.6821, "step": 855000 }, { "epoch": 5.155509460812098, "eval_loss": 5.015732765197754, "eval_runtime": 38.9554, "eval_samples_per_second": 13.143, "eval_steps_per_second": 6.572, "eval_tts_loss": 7.7938825720269005, "step": 855000 }, { "epoch": 5.155810952593432, "grad_norm": 6.07728385925293, "learning_rate": 4.8748408205747695e-05, "loss": 4.5938, "step": 855050 }, { "epoch": 5.156112444374767, "grad_norm": 2.9575977325439453, "learning_rate": 4.873339550765336e-05, "loss": 4.5293, "step": 855100 }, { "epoch": 5.156413936156101, "grad_norm": 3.742058038711548, "learning_rate": 4.87183846732263e-05, "loss": 4.8167, "step": 855150 }, { "epoch": 5.156715427937434, "grad_norm": 2.859302043914795, "learning_rate": 4.870337570274285e-05, "loss": 4.6297, "step": 855200 }, { "epoch": 5.157016919718768, "grad_norm": 3.694094181060791, "learning_rate": 4.868836859647925e-05, "loss": 4.7146, "step": 855250 }, { "epoch": 5.1573184115001025, "grad_norm": 2.2984375953674316, "learning_rate": 4.8673363354711655e-05, "loss": 4.4194, "step": 855300 }, { "epoch": 5.157619903281437, "grad_norm": 4.012394428253174, "learning_rate": 4.8658359977716154e-05, "loss": 4.6814, "step": 855350 }, { "epoch": 5.157921395062771, "grad_norm": 3.4219086170196533, "learning_rate": 4.864335846576897e-05, "loss": 4.7049, "step": 855400 }, { "epoch": 5.158222886844104, "grad_norm": 2.580470085144043, "learning_rate": 4.862835881914607e-05, "loss": 4.4671, "step": 855450 }, { "epoch": 5.1585243786254384, "grad_norm": 2.8834705352783203, "learning_rate": 4.861336103812357e-05, "loss": 4.2539, "step": 855500 }, { "epoch": 5.158825870406773, "grad_norm": 5.552759170532227, "learning_rate": 4.859836512297751e-05, "loss": 4.605, "step": 855550 }, { "epoch": 5.159127362188107, "grad_norm": 2.784782648086548, "learning_rate": 4.85833710739838e-05, "loss": 4.2746, "step": 855600 }, { "epoch": 5.159428853969441, "grad_norm": 3.9586663246154785, "learning_rate": 4.856837889141845e-05, "loss": 4.3538, "step": 855650 }, { "epoch": 5.159730345750775, "grad_norm": 4.394894599914551, "learning_rate": 4.8553388575557345e-05, "loss": 4.8053, "step": 855700 }, { "epoch": 5.160031837532109, "grad_norm": 2.156506299972534, "learning_rate": 4.8538400126676294e-05, "loss": 4.4837, "step": 855750 }, { "epoch": 5.160333329313443, "grad_norm": 3.2843406200408936, "learning_rate": 4.8523413545051206e-05, "loss": 4.4816, "step": 855800 }, { "epoch": 5.160634821094777, "grad_norm": 3.2600409984588623, "learning_rate": 4.850842883095791e-05, "loss": 4.3047, "step": 855850 }, { "epoch": 5.160936312876111, "grad_norm": 2.788213014602661, "learning_rate": 4.849344598467211e-05, "loss": 4.756, "step": 855900 }, { "epoch": 5.161237804657445, "grad_norm": 3.0931286811828613, "learning_rate": 4.847846500646957e-05, "loss": 4.4962, "step": 855950 }, { "epoch": 5.161539296438779, "grad_norm": 3.7343666553497314, "learning_rate": 4.8463485896626056e-05, "loss": 4.1931, "step": 856000 }, { "epoch": 5.161840788220113, "grad_norm": 3.459068536758423, "learning_rate": 4.8448508655417144e-05, "loss": 4.6491, "step": 856050 }, { "epoch": 5.162142280001447, "grad_norm": 2.8550305366516113, "learning_rate": 4.843353328311854e-05, "loss": 4.4484, "step": 856100 }, { "epoch": 5.162443771782781, "grad_norm": 3.343672513961792, "learning_rate": 4.8418559780005826e-05, "loss": 4.2123, "step": 856150 }, { "epoch": 5.162745263564116, "grad_norm": 3.0749032497406006, "learning_rate": 4.84035881463545e-05, "loss": 4.5374, "step": 856200 }, { "epoch": 5.163046755345449, "grad_norm": 4.154108047485352, "learning_rate": 4.8388618382440155e-05, "loss": 4.5661, "step": 856250 }, { "epoch": 5.163348247126783, "grad_norm": 3.840207576751709, "learning_rate": 4.837365048853832e-05, "loss": 4.7036, "step": 856300 }, { "epoch": 5.163649738908117, "grad_norm": 3.910512685775757, "learning_rate": 4.8358684464924375e-05, "loss": 4.7193, "step": 856350 }, { "epoch": 5.1639512306894515, "grad_norm": 4.637889385223389, "learning_rate": 4.83437203118738e-05, "loss": 4.427, "step": 856400 }, { "epoch": 5.164252722470786, "grad_norm": 3.267773151397705, "learning_rate": 4.832875802966203e-05, "loss": 4.5412, "step": 856450 }, { "epoch": 5.16455421425212, "grad_norm": 3.6442582607269287, "learning_rate": 4.831379761856435e-05, "loss": 4.3936, "step": 856500 }, { "epoch": 5.164855706033453, "grad_norm": 3.0978190898895264, "learning_rate": 4.829883907885609e-05, "loss": 4.7555, "step": 856550 }, { "epoch": 5.165157197814787, "grad_norm": 3.4412593841552734, "learning_rate": 4.82838824108126e-05, "loss": 4.8831, "step": 856600 }, { "epoch": 5.165458689596122, "grad_norm": 4.913561820983887, "learning_rate": 4.826892761470906e-05, "loss": 4.4048, "step": 856650 }, { "epoch": 5.165760181377456, "grad_norm": 3.652255058288574, "learning_rate": 4.825397469082072e-05, "loss": 4.8918, "step": 856700 }, { "epoch": 5.16606167315879, "grad_norm": 3.060126781463623, "learning_rate": 4.823902363942281e-05, "loss": 4.3234, "step": 856750 }, { "epoch": 5.166363164940123, "grad_norm": 3.87152099609375, "learning_rate": 4.822407446079042e-05, "loss": 4.4189, "step": 856800 }, { "epoch": 5.166664656721458, "grad_norm": 2.7583670616149902, "learning_rate": 4.820912715519874e-05, "loss": 4.6353, "step": 856850 }, { "epoch": 5.166966148502792, "grad_norm": 2.904437780380249, "learning_rate": 4.819418172292282e-05, "loss": 4.5444, "step": 856900 }, { "epoch": 5.167267640284126, "grad_norm": 3.96697998046875, "learning_rate": 4.817923816423764e-05, "loss": 4.6782, "step": 856950 }, { "epoch": 5.16756913206546, "grad_norm": 3.660736083984375, "learning_rate": 4.816429647941828e-05, "loss": 4.6384, "step": 857000 }, { "epoch": 5.1678706238467935, "grad_norm": 1.888892412185669, "learning_rate": 4.814935666873976e-05, "loss": 4.1057, "step": 857050 }, { "epoch": 5.168172115628128, "grad_norm": 3.4980950355529785, "learning_rate": 4.8134418732476946e-05, "loss": 4.8363, "step": 857100 }, { "epoch": 5.168473607409462, "grad_norm": 3.729865550994873, "learning_rate": 4.811948267090476e-05, "loss": 4.4375, "step": 857150 }, { "epoch": 5.168775099190796, "grad_norm": 1.76591956615448, "learning_rate": 4.810454848429818e-05, "loss": 4.4225, "step": 857200 }, { "epoch": 5.16907659097213, "grad_norm": 1.3105978965759277, "learning_rate": 4.808961617293192e-05, "loss": 4.5623, "step": 857250 }, { "epoch": 5.1693780827534646, "grad_norm": 2.025240421295166, "learning_rate": 4.8074685737080876e-05, "loss": 4.4324, "step": 857300 }, { "epoch": 5.169679574534798, "grad_norm": 2.8942947387695312, "learning_rate": 4.8059757177019795e-05, "loss": 4.564, "step": 857350 }, { "epoch": 5.169981066316132, "grad_norm": 1.681901216506958, "learning_rate": 4.804483049302336e-05, "loss": 4.6324, "step": 857400 }, { "epoch": 5.170282558097466, "grad_norm": 4.601197242736816, "learning_rate": 4.8029905685366314e-05, "loss": 4.486, "step": 857450 }, { "epoch": 5.1705840498788005, "grad_norm": 2.7047770023345947, "learning_rate": 4.8014982754323386e-05, "loss": 4.9128, "step": 857500 }, { "epoch": 5.170885541660135, "grad_norm": 1.862407922744751, "learning_rate": 4.800006170016911e-05, "loss": 4.767, "step": 857550 }, { "epoch": 5.171187033441468, "grad_norm": 3.1996006965637207, "learning_rate": 4.798514252317813e-05, "loss": 4.6679, "step": 857600 }, { "epoch": 5.171488525222802, "grad_norm": 3.359405755996704, "learning_rate": 4.797022522362509e-05, "loss": 4.5413, "step": 857650 }, { "epoch": 5.171790017004136, "grad_norm": 3.016165018081665, "learning_rate": 4.7955309801784436e-05, "loss": 4.6148, "step": 857700 }, { "epoch": 5.172091508785471, "grad_norm": 0.6138824820518494, "learning_rate": 4.794039625793063e-05, "loss": 4.6095, "step": 857750 }, { "epoch": 5.172393000566805, "grad_norm": 3.257011890411377, "learning_rate": 4.792548459233819e-05, "loss": 4.3242, "step": 857800 }, { "epoch": 5.172694492348138, "grad_norm": 3.6024515628814697, "learning_rate": 4.791057480528157e-05, "loss": 4.5066, "step": 857850 }, { "epoch": 5.172995984129472, "grad_norm": 2.518732786178589, "learning_rate": 4.7895666897035094e-05, "loss": 4.2821, "step": 857900 }, { "epoch": 5.173297475910807, "grad_norm": 3.396660566329956, "learning_rate": 4.788076086787316e-05, "loss": 4.588, "step": 857950 }, { "epoch": 5.173598967692141, "grad_norm": 3.2286431789398193, "learning_rate": 4.786585671807014e-05, "loss": 4.733, "step": 858000 }, { "epoch": 5.173598967692141, "eval_loss": 5.004810333251953, "eval_runtime": 39.0407, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.557, "eval_tts_loss": 7.852303166306544, "step": 858000 }, { "epoch": 5.173900459473475, "grad_norm": 3.2128663063049316, "learning_rate": 4.785095444790027e-05, "loss": 4.6602, "step": 858050 }, { "epoch": 5.174201951254809, "grad_norm": 3.068809747695923, "learning_rate": 4.783605405763776e-05, "loss": 4.9675, "step": 858100 }, { "epoch": 5.1745034430361425, "grad_norm": 2.744260549545288, "learning_rate": 4.782115554755693e-05, "loss": 4.7519, "step": 858150 }, { "epoch": 5.174804934817477, "grad_norm": 4.084619045257568, "learning_rate": 4.780625891793188e-05, "loss": 4.5542, "step": 858200 }, { "epoch": 5.175106426598811, "grad_norm": 2.3202004432678223, "learning_rate": 4.779136416903678e-05, "loss": 4.0292, "step": 858250 }, { "epoch": 5.175407918380145, "grad_norm": 0.8826591968536377, "learning_rate": 4.777647130114581e-05, "loss": 4.2539, "step": 858300 }, { "epoch": 5.175709410161479, "grad_norm": 2.398207187652588, "learning_rate": 4.7761580314532965e-05, "loss": 4.2767, "step": 858350 }, { "epoch": 5.176010901942813, "grad_norm": 2.674879550933838, "learning_rate": 4.7746691209472315e-05, "loss": 4.7419, "step": 858400 }, { "epoch": 5.176312393724147, "grad_norm": 2.826953411102295, "learning_rate": 4.773180398623796e-05, "loss": 4.3345, "step": 858450 }, { "epoch": 5.176613885505481, "grad_norm": 3.9703004360198975, "learning_rate": 4.7716918645103796e-05, "loss": 4.7477, "step": 858500 }, { "epoch": 5.176915377286815, "grad_norm": 3.818619966506958, "learning_rate": 4.770203518634374e-05, "loss": 4.3392, "step": 858550 }, { "epoch": 5.1772168690681495, "grad_norm": 4.482570171356201, "learning_rate": 4.768715361023177e-05, "loss": 4.492, "step": 858600 }, { "epoch": 5.177518360849483, "grad_norm": 3.7288246154785156, "learning_rate": 4.7672273917041705e-05, "loss": 4.5918, "step": 858650 }, { "epoch": 5.177819852630817, "grad_norm": 3.967822790145874, "learning_rate": 4.7657396107047385e-05, "loss": 4.5477, "step": 858700 }, { "epoch": 5.178121344412151, "grad_norm": 5.18989896774292, "learning_rate": 4.7642520180522695e-05, "loss": 4.4369, "step": 858750 }, { "epoch": 5.178422836193485, "grad_norm": 4.398643970489502, "learning_rate": 4.76276461377413e-05, "loss": 4.4841, "step": 858800 }, { "epoch": 5.17872432797482, "grad_norm": 3.1436996459960938, "learning_rate": 4.761277397897705e-05, "loss": 4.7509, "step": 858850 }, { "epoch": 5.179025819756154, "grad_norm": 1.571244716644287, "learning_rate": 4.759790370450356e-05, "loss": 4.237, "step": 858900 }, { "epoch": 5.179327311537487, "grad_norm": 4.039447784423828, "learning_rate": 4.7583035314594485e-05, "loss": 4.7647, "step": 858950 }, { "epoch": 5.179628803318821, "grad_norm": 2.1962573528289795, "learning_rate": 4.7568168809523485e-05, "loss": 4.7795, "step": 859000 }, { "epoch": 5.179930295100156, "grad_norm": 3.2500126361846924, "learning_rate": 4.755330418956419e-05, "loss": 4.4711, "step": 859050 }, { "epoch": 5.18023178688149, "grad_norm": 2.945075511932373, "learning_rate": 4.753844145499011e-05, "loss": 4.1681, "step": 859100 }, { "epoch": 5.180533278662824, "grad_norm": 2.722770929336548, "learning_rate": 4.752358060607478e-05, "loss": 4.3645, "step": 859150 }, { "epoch": 5.180834770444157, "grad_norm": 3.926853656768799, "learning_rate": 4.7508721643091746e-05, "loss": 4.5854, "step": 859200 }, { "epoch": 5.1811362622254915, "grad_norm": 3.854142189025879, "learning_rate": 4.7493864566314435e-05, "loss": 4.3848, "step": 859250 }, { "epoch": 5.181437754006826, "grad_norm": 2.6115970611572266, "learning_rate": 4.7479009376016216e-05, "loss": 4.627, "step": 859300 }, { "epoch": 5.18173924578816, "grad_norm": 4.8242998123168945, "learning_rate": 4.746415607247057e-05, "loss": 4.7075, "step": 859350 }, { "epoch": 5.182040737569494, "grad_norm": 4.249258995056152, "learning_rate": 4.7449304655950775e-05, "loss": 4.5622, "step": 859400 }, { "epoch": 5.182342229350828, "grad_norm": 4.012921333312988, "learning_rate": 4.743445512673017e-05, "loss": 4.508, "step": 859450 }, { "epoch": 5.182643721132162, "grad_norm": 3.2730700969696045, "learning_rate": 4.741960748508208e-05, "loss": 4.3776, "step": 859500 }, { "epoch": 5.182945212913496, "grad_norm": 2.9135491847991943, "learning_rate": 4.7404761731279685e-05, "loss": 4.4524, "step": 859550 }, { "epoch": 5.18324670469483, "grad_norm": 1.2648130655288696, "learning_rate": 4.738991786559624e-05, "loss": 4.1993, "step": 859600 }, { "epoch": 5.183548196476164, "grad_norm": 3.043184280395508, "learning_rate": 4.737507588830498e-05, "loss": 4.6037, "step": 859650 }, { "epoch": 5.1838496882574985, "grad_norm": 3.473205089569092, "learning_rate": 4.736023579967899e-05, "loss": 4.4316, "step": 859700 }, { "epoch": 5.184151180038832, "grad_norm": 5.821686744689941, "learning_rate": 4.7345397599991344e-05, "loss": 4.5025, "step": 859750 }, { "epoch": 5.184452671820166, "grad_norm": 3.5179076194763184, "learning_rate": 4.7330561289515194e-05, "loss": 4.3022, "step": 859800 }, { "epoch": 5.1847541636015, "grad_norm": 4.202495098114014, "learning_rate": 4.7315726868523504e-05, "loss": 4.3577, "step": 859850 }, { "epoch": 5.185055655382834, "grad_norm": 4.651796340942383, "learning_rate": 4.730089433728932e-05, "loss": 4.5617, "step": 859900 }, { "epoch": 5.185357147164169, "grad_norm": 3.5953497886657715, "learning_rate": 4.728606369608568e-05, "loss": 4.2013, "step": 859950 }, { "epoch": 5.185658638945502, "grad_norm": 2.4514856338500977, "learning_rate": 4.72712349451854e-05, "loss": 4.6701, "step": 860000 }, { "epoch": 5.185960130726836, "grad_norm": 2.742403507232666, "learning_rate": 4.725640808486148e-05, "loss": 4.4244, "step": 860050 }, { "epoch": 5.18626162250817, "grad_norm": 2.8565423488616943, "learning_rate": 4.7241583115386737e-05, "loss": 4.4917, "step": 860100 }, { "epoch": 5.186563114289505, "grad_norm": 3.237487554550171, "learning_rate": 4.722676003703399e-05, "loss": 4.2759, "step": 860150 }, { "epoch": 5.186864606070839, "grad_norm": 3.9960715770721436, "learning_rate": 4.721193885007605e-05, "loss": 4.5462, "step": 860200 }, { "epoch": 5.187166097852173, "grad_norm": 1.9523290395736694, "learning_rate": 4.719711955478573e-05, "loss": 4.3144, "step": 860250 }, { "epoch": 5.187467589633506, "grad_norm": 2.836134910583496, "learning_rate": 4.718230215143566e-05, "loss": 4.5288, "step": 860300 }, { "epoch": 5.1877690814148405, "grad_norm": 4.291683197021484, "learning_rate": 4.716748664029862e-05, "loss": 4.3244, "step": 860350 }, { "epoch": 5.188070573196175, "grad_norm": 6.528193473815918, "learning_rate": 4.715267302164726e-05, "loss": 4.3915, "step": 860400 }, { "epoch": 5.188372064977509, "grad_norm": 3.0332634449005127, "learning_rate": 4.7137861295754145e-05, "loss": 4.6535, "step": 860450 }, { "epoch": 5.188673556758843, "grad_norm": 3.6728551387786865, "learning_rate": 4.7123051462891954e-05, "loss": 4.6146, "step": 860500 }, { "epoch": 5.188975048540176, "grad_norm": 5.594315528869629, "learning_rate": 4.7108243523333176e-05, "loss": 4.3705, "step": 860550 }, { "epoch": 5.189276540321511, "grad_norm": 3.5837738513946533, "learning_rate": 4.70934374773503e-05, "loss": 4.78, "step": 860600 }, { "epoch": 5.189578032102845, "grad_norm": 1.9580111503601074, "learning_rate": 4.707863332521585e-05, "loss": 4.7484, "step": 860650 }, { "epoch": 5.189879523884179, "grad_norm": 3.100057363510132, "learning_rate": 4.706383106720232e-05, "loss": 4.9041, "step": 860700 }, { "epoch": 5.190181015665513, "grad_norm": 2.535518169403076, "learning_rate": 4.704903070358205e-05, "loss": 4.2135, "step": 860750 }, { "epoch": 5.190482507446847, "grad_norm": 0.9517030119895935, "learning_rate": 4.7034232234627434e-05, "loss": 4.4711, "step": 860800 }, { "epoch": 5.190783999228181, "grad_norm": 1.315659523010254, "learning_rate": 4.701943566061088e-05, "loss": 4.595, "step": 860850 }, { "epoch": 5.191085491009515, "grad_norm": 4.565047264099121, "learning_rate": 4.700464098180466e-05, "loss": 4.2693, "step": 860900 }, { "epoch": 5.191386982790849, "grad_norm": 3.1549243927001953, "learning_rate": 4.6989848198480995e-05, "loss": 4.5953, "step": 860950 }, { "epoch": 5.191688474572183, "grad_norm": 4.3853302001953125, "learning_rate": 4.697505731091217e-05, "loss": 4.5887, "step": 861000 }, { "epoch": 5.191688474572183, "eval_loss": 5.003710746765137, "eval_runtime": 39.0821, "eval_samples_per_second": 13.101, "eval_steps_per_second": 6.55, "eval_tts_loss": 7.889424778436123, "step": 861000 }, { "epoch": 5.191989966353518, "grad_norm": 4.129936218261719, "learning_rate": 4.696026831937042e-05, "loss": 4.5289, "step": 861050 }, { "epoch": 5.192291458134851, "grad_norm": 2.9999520778656006, "learning_rate": 4.6945481224127867e-05, "loss": 4.8567, "step": 861100 }, { "epoch": 5.192592949916185, "grad_norm": 2.9304065704345703, "learning_rate": 4.6930696025456644e-05, "loss": 4.4785, "step": 861150 }, { "epoch": 5.192894441697519, "grad_norm": 3.6412363052368164, "learning_rate": 4.6915912723628936e-05, "loss": 4.6462, "step": 861200 }, { "epoch": 5.1931959334788536, "grad_norm": 3.9417307376861572, "learning_rate": 4.690113131891674e-05, "loss": 4.7951, "step": 861250 }, { "epoch": 5.193497425260188, "grad_norm": 5.698966979980469, "learning_rate": 4.688635181159203e-05, "loss": 4.5263, "step": 861300 }, { "epoch": 5.193798917041521, "grad_norm": 6.198087215423584, "learning_rate": 4.6871574201926924e-05, "loss": 4.4898, "step": 861350 }, { "epoch": 5.194100408822855, "grad_norm": 3.424631357192993, "learning_rate": 4.6856798490193284e-05, "loss": 4.2293, "step": 861400 }, { "epoch": 5.1944019006041895, "grad_norm": 0.9148363471031189, "learning_rate": 4.684202467666306e-05, "loss": 4.2968, "step": 861450 }, { "epoch": 5.194703392385524, "grad_norm": 3.328731060028076, "learning_rate": 4.682725276160821e-05, "loss": 4.7551, "step": 861500 }, { "epoch": 5.195004884166858, "grad_norm": 2.902942180633545, "learning_rate": 4.6812482745300486e-05, "loss": 4.5391, "step": 861550 }, { "epoch": 5.195306375948191, "grad_norm": 4.2251057624816895, "learning_rate": 4.679771462801177e-05, "loss": 4.3926, "step": 861600 }, { "epoch": 5.195607867729525, "grad_norm": 3.5058350563049316, "learning_rate": 4.6782948410013874e-05, "loss": 4.5854, "step": 861650 }, { "epoch": 5.19590935951086, "grad_norm": 2.4929730892181396, "learning_rate": 4.6768184091578526e-05, "loss": 4.2428, "step": 861700 }, { "epoch": 5.196210851292194, "grad_norm": 5.481373310089111, "learning_rate": 4.675342167297738e-05, "loss": 4.5046, "step": 861750 }, { "epoch": 5.196512343073528, "grad_norm": 5.766800403594971, "learning_rate": 4.673866115448221e-05, "loss": 4.5099, "step": 861800 }, { "epoch": 5.196813834854862, "grad_norm": 1.3236984014511108, "learning_rate": 4.672390253636458e-05, "loss": 4.576, "step": 861850 }, { "epoch": 5.197115326636196, "grad_norm": 2.8855772018432617, "learning_rate": 4.670914581889613e-05, "loss": 3.8875, "step": 861900 }, { "epoch": 5.19741681841753, "grad_norm": 2.7039976119995117, "learning_rate": 4.669439100234851e-05, "loss": 4.4106, "step": 861950 }, { "epoch": 5.197718310198864, "grad_norm": 1.7850278615951538, "learning_rate": 4.667963808699314e-05, "loss": 4.4728, "step": 862000 }, { "epoch": 5.198019801980198, "grad_norm": 3.1049234867095947, "learning_rate": 4.666488707310162e-05, "loss": 4.8077, "step": 862050 }, { "epoch": 5.198321293761532, "grad_norm": 1.7922579050064087, "learning_rate": 4.6650137960945404e-05, "loss": 4.6465, "step": 862100 }, { "epoch": 5.198622785542866, "grad_norm": 2.2995176315307617, "learning_rate": 4.663539075079586e-05, "loss": 4.253, "step": 862150 }, { "epoch": 5.1989242773242, "grad_norm": 2.6322021484375, "learning_rate": 4.662064544292443e-05, "loss": 4.7076, "step": 862200 }, { "epoch": 5.199225769105534, "grad_norm": 3.170684814453125, "learning_rate": 4.6605902037602556e-05, "loss": 4.446, "step": 862250 }, { "epoch": 5.199527260886868, "grad_norm": 2.988419532775879, "learning_rate": 4.659116053510145e-05, "loss": 4.5277, "step": 862300 }, { "epoch": 5.1998287526682025, "grad_norm": 4.176224708557129, "learning_rate": 4.657642093569245e-05, "loss": 4.3597, "step": 862350 }, { "epoch": 5.200130244449536, "grad_norm": 5.281150817871094, "learning_rate": 4.65616832396469e-05, "loss": 4.3895, "step": 862400 }, { "epoch": 5.20043173623087, "grad_norm": 4.476017951965332, "learning_rate": 4.654694744723594e-05, "loss": 4.3663, "step": 862450 }, { "epoch": 5.200733228012204, "grad_norm": 3.617155075073242, "learning_rate": 4.653221355873073e-05, "loss": 4.2096, "step": 862500 }, { "epoch": 5.2010347197935385, "grad_norm": 3.138658285140991, "learning_rate": 4.651748157440252e-05, "loss": 4.6033, "step": 862550 }, { "epoch": 5.201336211574873, "grad_norm": 3.678283929824829, "learning_rate": 4.650275149452233e-05, "loss": 4.5144, "step": 862600 }, { "epoch": 5.201637703356207, "grad_norm": 2.967444896697998, "learning_rate": 4.648802331936132e-05, "loss": 4.4284, "step": 862650 }, { "epoch": 5.20193919513754, "grad_norm": 6.3852996826171875, "learning_rate": 4.647329704919056e-05, "loss": 4.4204, "step": 862700 }, { "epoch": 5.202240686918874, "grad_norm": 4.00788688659668, "learning_rate": 4.645857268428096e-05, "loss": 4.3328, "step": 862750 }, { "epoch": 5.202542178700209, "grad_norm": 2.8246309757232666, "learning_rate": 4.644385022490359e-05, "loss": 4.403, "step": 862800 }, { "epoch": 5.202843670481543, "grad_norm": 3.821382999420166, "learning_rate": 4.642912967132941e-05, "loss": 4.6204, "step": 862850 }, { "epoch": 5.203145162262877, "grad_norm": 2.516477108001709, "learning_rate": 4.641441102382931e-05, "loss": 4.5175, "step": 862900 }, { "epoch": 5.20344665404421, "grad_norm": 4.020590782165527, "learning_rate": 4.639969428267409e-05, "loss": 4.6702, "step": 862950 }, { "epoch": 5.203748145825545, "grad_norm": 4.151229381561279, "learning_rate": 4.638497944813471e-05, "loss": 4.6271, "step": 863000 }, { "epoch": 5.204049637606879, "grad_norm": 3.2040154933929443, "learning_rate": 4.637026652048185e-05, "loss": 4.7358, "step": 863050 }, { "epoch": 5.204351129388213, "grad_norm": 3.2146029472351074, "learning_rate": 4.635555549998636e-05, "loss": 4.58, "step": 863100 }, { "epoch": 5.204652621169547, "grad_norm": 1.8803097009658813, "learning_rate": 4.634084638691902e-05, "loss": 4.6274, "step": 863150 }, { "epoch": 5.204954112950881, "grad_norm": 4.280269145965576, "learning_rate": 4.6326139181550416e-05, "loss": 4.4496, "step": 863200 }, { "epoch": 5.205255604732215, "grad_norm": 4.432158470153809, "learning_rate": 4.631143388415131e-05, "loss": 4.6102, "step": 863250 }, { "epoch": 5.205557096513549, "grad_norm": 4.573638439178467, "learning_rate": 4.6296730494992294e-05, "loss": 4.4939, "step": 863300 }, { "epoch": 5.205858588294883, "grad_norm": 3.06923508644104, "learning_rate": 4.628202901434392e-05, "loss": 4.1594, "step": 863350 }, { "epoch": 5.206160080076217, "grad_norm": 3.017165422439575, "learning_rate": 4.6267329442476794e-05, "loss": 4.7956, "step": 863400 }, { "epoch": 5.2064615718575515, "grad_norm": 3.781877040863037, "learning_rate": 4.625263177966147e-05, "loss": 4.4304, "step": 863450 }, { "epoch": 5.206763063638885, "grad_norm": 2.8950862884521484, "learning_rate": 4.6237936026168356e-05, "loss": 4.474, "step": 863500 }, { "epoch": 5.207064555420219, "grad_norm": 4.474878311157227, "learning_rate": 4.622324218226796e-05, "loss": 4.9936, "step": 863550 }, { "epoch": 5.207366047201553, "grad_norm": 3.9146344661712646, "learning_rate": 4.620855024823074e-05, "loss": 4.4272, "step": 863600 }, { "epoch": 5.2076675389828875, "grad_norm": 3.013841390609741, "learning_rate": 4.619386022432703e-05, "loss": 4.2069, "step": 863650 }, { "epoch": 5.207969030764222, "grad_norm": 2.006434917449951, "learning_rate": 4.617917211082714e-05, "loss": 4.5326, "step": 863700 }, { "epoch": 5.208270522545555, "grad_norm": 2.332218647003174, "learning_rate": 4.616448590800147e-05, "loss": 4.411, "step": 863750 }, { "epoch": 5.208572014326889, "grad_norm": 3.352876901626587, "learning_rate": 4.6149801616120205e-05, "loss": 4.5248, "step": 863800 }, { "epoch": 5.208873506108223, "grad_norm": 3.673501491546631, "learning_rate": 4.613511923545364e-05, "loss": 4.4831, "step": 863850 }, { "epoch": 5.209174997889558, "grad_norm": 5.060515403747559, "learning_rate": 4.6120438766272046e-05, "loss": 4.6241, "step": 863900 }, { "epoch": 5.209476489670892, "grad_norm": 3.6918349266052246, "learning_rate": 4.610576020884547e-05, "loss": 4.3834, "step": 863950 }, { "epoch": 5.209777981452226, "grad_norm": 4.236481189727783, "learning_rate": 4.609108356344412e-05, "loss": 4.4, "step": 864000 }, { "epoch": 5.209777981452226, "eval_loss": 5.003841400146484, "eval_runtime": 39.1331, "eval_samples_per_second": 13.084, "eval_steps_per_second": 6.542, "eval_tts_loss": 7.840813354060155, "step": 864000 }, { "epoch": 5.210079473233559, "grad_norm": 2.9746527671813965, "learning_rate": 4.6076408830338124e-05, "loss": 4.7891, "step": 864050 }, { "epoch": 5.210380965014894, "grad_norm": 4.411005020141602, "learning_rate": 4.6061736009797533e-05, "loss": 4.7757, "step": 864100 }, { "epoch": 5.210682456796228, "grad_norm": 4.920834064483643, "learning_rate": 4.60470651020923e-05, "loss": 4.8455, "step": 864150 }, { "epoch": 5.210983948577562, "grad_norm": 3.975304126739502, "learning_rate": 4.603239610749254e-05, "loss": 4.3364, "step": 864200 }, { "epoch": 5.211285440358896, "grad_norm": 5.467555522918701, "learning_rate": 4.601772902626812e-05, "loss": 4.4943, "step": 864250 }, { "epoch": 5.2115869321402295, "grad_norm": 2.637460470199585, "learning_rate": 4.6003063858689e-05, "loss": 4.6071, "step": 864300 }, { "epoch": 5.211888423921564, "grad_norm": 2.3723576068878174, "learning_rate": 4.5988400605025085e-05, "loss": 4.1553, "step": 864350 }, { "epoch": 5.212189915702898, "grad_norm": 3.670635938644409, "learning_rate": 4.597373926554626e-05, "loss": 4.5949, "step": 864400 }, { "epoch": 5.212491407484232, "grad_norm": 2.5508577823638916, "learning_rate": 4.5959079840522314e-05, "loss": 4.4332, "step": 864450 }, { "epoch": 5.212792899265566, "grad_norm": 3.2266924381256104, "learning_rate": 4.5944422330222984e-05, "loss": 4.822, "step": 864500 }, { "epoch": 5.2130943910469, "grad_norm": 5.061856269836426, "learning_rate": 4.592976673491812e-05, "loss": 4.0298, "step": 864550 }, { "epoch": 5.213395882828234, "grad_norm": 2.7520453929901123, "learning_rate": 4.5915113054877326e-05, "loss": 4.2783, "step": 864600 }, { "epoch": 5.213697374609568, "grad_norm": 3.415038824081421, "learning_rate": 4.590046129037034e-05, "loss": 4.6867, "step": 864650 }, { "epoch": 5.213998866390902, "grad_norm": 4.421231746673584, "learning_rate": 4.588581144166686e-05, "loss": 4.0347, "step": 864700 }, { "epoch": 5.2143003581722365, "grad_norm": 3.3338658809661865, "learning_rate": 4.587116350903639e-05, "loss": 4.5146, "step": 864750 }, { "epoch": 5.214601849953571, "grad_norm": 3.5584299564361572, "learning_rate": 4.58565174927486e-05, "loss": 4.7157, "step": 864800 }, { "epoch": 5.214903341734904, "grad_norm": 2.486276149749756, "learning_rate": 4.584187339307298e-05, "loss": 4.7165, "step": 864850 }, { "epoch": 5.215204833516238, "grad_norm": 3.7980964183807373, "learning_rate": 4.5827231210279e-05, "loss": 4.434, "step": 864900 }, { "epoch": 5.215506325297572, "grad_norm": 3.1454238891601562, "learning_rate": 4.581259094463617e-05, "loss": 4.9174, "step": 864950 }, { "epoch": 5.215807817078907, "grad_norm": 3.5900022983551025, "learning_rate": 4.579795259641396e-05, "loss": 4.7356, "step": 865000 }, { "epoch": 5.216109308860241, "grad_norm": 3.0929832458496094, "learning_rate": 4.5783316165881675e-05, "loss": 4.4931, "step": 865050 }, { "epoch": 5.216410800641574, "grad_norm": 2.5840752124786377, "learning_rate": 4.5768681653308756e-05, "loss": 4.5404, "step": 865100 }, { "epoch": 5.216712292422908, "grad_norm": 3.4587514400482178, "learning_rate": 4.5754049058964535e-05, "loss": 4.2858, "step": 865150 }, { "epoch": 5.2170137842042426, "grad_norm": 1.5861490964889526, "learning_rate": 4.5739418383118213e-05, "loss": 4.7457, "step": 865200 }, { "epoch": 5.217315275985577, "grad_norm": 3.6629228591918945, "learning_rate": 4.572478962603919e-05, "loss": 4.4825, "step": 865250 }, { "epoch": 5.217616767766911, "grad_norm": 2.977792739868164, "learning_rate": 4.571016278799657e-05, "loss": 4.5147, "step": 865300 }, { "epoch": 5.217918259548244, "grad_norm": 4.90057897567749, "learning_rate": 4.5695537869259543e-05, "loss": 4.1574, "step": 865350 }, { "epoch": 5.2182197513295785, "grad_norm": 4.439706802368164, "learning_rate": 4.568091487009729e-05, "loss": 4.5462, "step": 865400 }, { "epoch": 5.218521243110913, "grad_norm": 4.1804890632629395, "learning_rate": 4.5666293790778976e-05, "loss": 4.3297, "step": 865450 }, { "epoch": 5.218822734892247, "grad_norm": 2.635939836502075, "learning_rate": 4.5651674631573594e-05, "loss": 4.5085, "step": 865500 }, { "epoch": 5.219124226673581, "grad_norm": 3.4919755458831787, "learning_rate": 4.563705739275022e-05, "loss": 4.5311, "step": 865550 }, { "epoch": 5.219425718454915, "grad_norm": 3.027963876724243, "learning_rate": 4.562244207457791e-05, "loss": 4.5283, "step": 865600 }, { "epoch": 5.219727210236249, "grad_norm": 2.375629425048828, "learning_rate": 4.560782867732561e-05, "loss": 4.7299, "step": 865650 }, { "epoch": 5.220028702017583, "grad_norm": 3.6446707248687744, "learning_rate": 4.5593217201262186e-05, "loss": 4.5713, "step": 865700 }, { "epoch": 5.220330193798917, "grad_norm": 4.56107234954834, "learning_rate": 4.557860764665665e-05, "loss": 4.8274, "step": 865750 }, { "epoch": 5.220631685580251, "grad_norm": 5.230074405670166, "learning_rate": 4.55640000137778e-05, "loss": 4.6077, "step": 865800 }, { "epoch": 5.2209331773615855, "grad_norm": 3.470212936401367, "learning_rate": 4.5549394302894474e-05, "loss": 4.1374, "step": 865850 }, { "epoch": 5.221234669142919, "grad_norm": 3.8926308155059814, "learning_rate": 4.553479051427553e-05, "loss": 4.843, "step": 865900 }, { "epoch": 5.221536160924253, "grad_norm": 3.3730721473693848, "learning_rate": 4.552018864818965e-05, "loss": 4.2779, "step": 865950 }, { "epoch": 5.221837652705587, "grad_norm": 3.738233804702759, "learning_rate": 4.550558870490564e-05, "loss": 4.5036, "step": 866000 }, { "epoch": 5.222139144486921, "grad_norm": 2.4103305339813232, "learning_rate": 4.54909906846921e-05, "loss": 4.6404, "step": 866050 }, { "epoch": 5.222440636268256, "grad_norm": 6.395610809326172, "learning_rate": 4.547639458781778e-05, "loss": 5.09, "step": 866100 }, { "epoch": 5.222742128049589, "grad_norm": 4.989628314971924, "learning_rate": 4.546180041455121e-05, "loss": 4.4107, "step": 866150 }, { "epoch": 5.223043619830923, "grad_norm": 3.1394882202148438, "learning_rate": 4.544720816516107e-05, "loss": 4.2267, "step": 866200 }, { "epoch": 5.223345111612257, "grad_norm": 3.137531280517578, "learning_rate": 4.543261783991581e-05, "loss": 4.5236, "step": 866250 }, { "epoch": 5.2236466033935915, "grad_norm": 3.201934814453125, "learning_rate": 4.541802943908398e-05, "loss": 4.7146, "step": 866300 }, { "epoch": 5.223948095174926, "grad_norm": 4.01151180267334, "learning_rate": 4.540344296293415e-05, "loss": 4.6174, "step": 866350 }, { "epoch": 5.22424958695626, "grad_norm": 9.499741554260254, "learning_rate": 4.5388858411734605e-05, "loss": 4.3698, "step": 866400 }, { "epoch": 5.224551078737593, "grad_norm": 1.0529216527938843, "learning_rate": 4.537427578575389e-05, "loss": 4.16, "step": 866450 }, { "epoch": 5.2248525705189275, "grad_norm": 3.552649974822998, "learning_rate": 4.535969508526032e-05, "loss": 4.7635, "step": 866500 }, { "epoch": 5.225154062300262, "grad_norm": 4.006450653076172, "learning_rate": 4.5345116310522195e-05, "loss": 4.4763, "step": 866550 }, { "epoch": 5.225455554081596, "grad_norm": 6.197049617767334, "learning_rate": 4.533053946180784e-05, "loss": 4.654, "step": 866600 }, { "epoch": 5.22575704586293, "grad_norm": 4.049698829650879, "learning_rate": 4.53159645393856e-05, "loss": 4.8166, "step": 866650 }, { "epoch": 5.226058537644263, "grad_norm": 4.837828159332275, "learning_rate": 4.530139154352358e-05, "loss": 4.5523, "step": 866700 }, { "epoch": 5.226360029425598, "grad_norm": 3.5183780193328857, "learning_rate": 4.5286820474490025e-05, "loss": 4.691, "step": 866750 }, { "epoch": 5.226661521206932, "grad_norm": 3.4933571815490723, "learning_rate": 4.5272251332553175e-05, "loss": 4.444, "step": 866800 }, { "epoch": 5.226963012988266, "grad_norm": 3.166654586791992, "learning_rate": 4.5257684117981066e-05, "loss": 4.4227, "step": 866850 }, { "epoch": 5.2272645047696, "grad_norm": 2.892529249191284, "learning_rate": 4.524311883104177e-05, "loss": 4.4304, "step": 866900 }, { "epoch": 5.2275659965509345, "grad_norm": 4.664366245269775, "learning_rate": 4.5228555472003424e-05, "loss": 4.7841, "step": 866950 }, { "epoch": 5.227867488332268, "grad_norm": 4.142714023590088, "learning_rate": 4.5213994041133954e-05, "loss": 4.6129, "step": 867000 }, { "epoch": 5.227867488332268, "eval_loss": 4.990819931030273, "eval_runtime": 39.034, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.854794672955998, "step": 867000 }, { "epoch": 5.228168980113602, "grad_norm": 1.8430235385894775, "learning_rate": 4.519943453870138e-05, "loss": 4.3442, "step": 867050 }, { "epoch": 5.228470471894936, "grad_norm": 3.547689437866211, "learning_rate": 4.5184876964973696e-05, "loss": 4.5312, "step": 867100 }, { "epoch": 5.22877196367627, "grad_norm": 3.4213287830352783, "learning_rate": 4.517032132021873e-05, "loss": 4.2507, "step": 867150 }, { "epoch": 5.229073455457605, "grad_norm": 2.9456112384796143, "learning_rate": 4.51557676047044e-05, "loss": 4.3411, "step": 867200 }, { "epoch": 5.229374947238938, "grad_norm": 4.304017543792725, "learning_rate": 4.514121581869859e-05, "loss": 4.3742, "step": 867250 }, { "epoch": 5.229676439020272, "grad_norm": 3.1138744354248047, "learning_rate": 4.5126665962469064e-05, "loss": 4.4856, "step": 867300 }, { "epoch": 5.229977930801606, "grad_norm": 3.1425137519836426, "learning_rate": 4.511211803628355e-05, "loss": 4.6453, "step": 867350 }, { "epoch": 5.2302794225829405, "grad_norm": 4.38671875, "learning_rate": 4.5097572040409865e-05, "loss": 4.3087, "step": 867400 }, { "epoch": 5.230580914364275, "grad_norm": 3.0252304077148438, "learning_rate": 4.508302797511561e-05, "loss": 4.5745, "step": 867450 }, { "epoch": 5.230882406145608, "grad_norm": 5.00927209854126, "learning_rate": 4.5068485840668494e-05, "loss": 4.4982, "step": 867500 }, { "epoch": 5.231183897926942, "grad_norm": 3.6082186698913574, "learning_rate": 4.50539456373362e-05, "loss": 4.3231, "step": 867550 }, { "epoch": 5.2314853897082765, "grad_norm": 2.7193892002105713, "learning_rate": 4.503940736538623e-05, "loss": 4.6812, "step": 867600 }, { "epoch": 5.231786881489611, "grad_norm": 3.7277536392211914, "learning_rate": 4.502487102508623e-05, "loss": 4.9781, "step": 867650 }, { "epoch": 5.232088373270945, "grad_norm": 3.391965627670288, "learning_rate": 4.501033661670363e-05, "loss": 4.6028, "step": 867700 }, { "epoch": 5.232389865052279, "grad_norm": 1.8028079271316528, "learning_rate": 4.4995804140505986e-05, "loss": 4.8082, "step": 867750 }, { "epoch": 5.232691356833612, "grad_norm": 3.461707830429077, "learning_rate": 4.498127359676067e-05, "loss": 4.7366, "step": 867800 }, { "epoch": 5.232992848614947, "grad_norm": 4.050750732421875, "learning_rate": 4.4966744985735166e-05, "loss": 4.6051, "step": 867850 }, { "epoch": 5.233294340396281, "grad_norm": 3.285299301147461, "learning_rate": 4.495221830769686e-05, "loss": 4.4975, "step": 867900 }, { "epoch": 5.233595832177615, "grad_norm": 4.469858169555664, "learning_rate": 4.493769356291303e-05, "loss": 4.3643, "step": 867950 }, { "epoch": 5.233897323958949, "grad_norm": 2.2851216793060303, "learning_rate": 4.492317075165105e-05, "loss": 4.5393, "step": 868000 }, { "epoch": 5.234198815740283, "grad_norm": 3.5151991844177246, "learning_rate": 4.490864987417817e-05, "loss": 4.3835, "step": 868050 }, { "epoch": 5.234500307521617, "grad_norm": 2.545868158340454, "learning_rate": 4.4894130930761554e-05, "loss": 4.4883, "step": 868100 }, { "epoch": 5.234801799302951, "grad_norm": 3.0707387924194336, "learning_rate": 4.4879613921668475e-05, "loss": 4.4576, "step": 868150 }, { "epoch": 5.235103291084285, "grad_norm": 3.5752196311950684, "learning_rate": 4.4865098847166134e-05, "loss": 4.2731, "step": 868200 }, { "epoch": 5.235404782865619, "grad_norm": 3.80863881111145, "learning_rate": 4.485058570752157e-05, "loss": 4.1661, "step": 868250 }, { "epoch": 5.235706274646953, "grad_norm": 2.418445110321045, "learning_rate": 4.4836074503001915e-05, "loss": 4.7326, "step": 868300 }, { "epoch": 5.236007766428287, "grad_norm": 3.3980937004089355, "learning_rate": 4.482156523387428e-05, "loss": 5.0573, "step": 868350 }, { "epoch": 5.236309258209621, "grad_norm": 4.121992111206055, "learning_rate": 4.4807057900405593e-05, "loss": 4.7169, "step": 868400 }, { "epoch": 5.236610749990955, "grad_norm": 4.788830757141113, "learning_rate": 4.479255250286293e-05, "loss": 4.1556, "step": 868450 }, { "epoch": 5.2369122417722895, "grad_norm": 3.1140732765197754, "learning_rate": 4.477804904151321e-05, "loss": 4.5902, "step": 868500 }, { "epoch": 5.237213733553624, "grad_norm": 3.756601572036743, "learning_rate": 4.476354751662327e-05, "loss": 4.6058, "step": 868550 }, { "epoch": 5.237515225334957, "grad_norm": 4.620804786682129, "learning_rate": 4.474904792846008e-05, "loss": 4.6659, "step": 868600 }, { "epoch": 5.237816717116291, "grad_norm": 2.2235355377197266, "learning_rate": 4.473455027729049e-05, "loss": 4.3993, "step": 868650 }, { "epoch": 5.2381182088976255, "grad_norm": 4.179457664489746, "learning_rate": 4.4720054563381244e-05, "loss": 4.6812, "step": 868700 }, { "epoch": 5.23841970067896, "grad_norm": 2.662271738052368, "learning_rate": 4.470556078699915e-05, "loss": 4.3937, "step": 868750 }, { "epoch": 5.238721192460294, "grad_norm": 5.439343452453613, "learning_rate": 4.469106894841099e-05, "loss": 4.7026, "step": 868800 }, { "epoch": 5.239022684241627, "grad_norm": 3.6714508533477783, "learning_rate": 4.467657904788344e-05, "loss": 4.9844, "step": 868850 }, { "epoch": 5.239324176022961, "grad_norm": 4.869085311889648, "learning_rate": 4.4662091085683084e-05, "loss": 4.0533, "step": 868900 }, { "epoch": 5.239625667804296, "grad_norm": 3.8602914810180664, "learning_rate": 4.464760506207666e-05, "loss": 4.5091, "step": 868950 }, { "epoch": 5.23992715958563, "grad_norm": 4.38851261138916, "learning_rate": 4.463312097733067e-05, "loss": 4.5385, "step": 869000 }, { "epoch": 5.240228651366964, "grad_norm": 3.182098388671875, "learning_rate": 4.461863883171173e-05, "loss": 4.538, "step": 869050 }, { "epoch": 5.240530143148297, "grad_norm": 4.277127265930176, "learning_rate": 4.4604158625486405e-05, "loss": 4.6732, "step": 869100 }, { "epoch": 5.2408316349296316, "grad_norm": 3.910876512527466, "learning_rate": 4.458968035892107e-05, "loss": 4.4366, "step": 869150 }, { "epoch": 5.241133126710966, "grad_norm": 4.38349723815918, "learning_rate": 4.45752040322823e-05, "loss": 4.4307, "step": 869200 }, { "epoch": 5.2414346184923, "grad_norm": 4.402797222137451, "learning_rate": 4.4560729645836453e-05, "loss": 4.3804, "step": 869250 }, { "epoch": 5.241736110273634, "grad_norm": 5.885738849639893, "learning_rate": 4.454625719984986e-05, "loss": 5.0991, "step": 869300 }, { "epoch": 5.242037602054968, "grad_norm": 1.9033896923065186, "learning_rate": 4.45317866945889e-05, "loss": 4.2676, "step": 869350 }, { "epoch": 5.242339093836302, "grad_norm": 4.62483024597168, "learning_rate": 4.451731813031995e-05, "loss": 4.7499, "step": 869400 }, { "epoch": 5.242640585617636, "grad_norm": 3.202893018722534, "learning_rate": 4.4502851507309184e-05, "loss": 4.1989, "step": 869450 }, { "epoch": 5.24294207739897, "grad_norm": 4.49053955078125, "learning_rate": 4.4488386825822884e-05, "loss": 4.9404, "step": 869500 }, { "epoch": 5.243243569180304, "grad_norm": 3.0244078636169434, "learning_rate": 4.44739240861273e-05, "loss": 4.4026, "step": 869550 }, { "epoch": 5.2435450609616385, "grad_norm": 4.038372993469238, "learning_rate": 4.4459463288488494e-05, "loss": 4.4556, "step": 869600 }, { "epoch": 5.243846552742972, "grad_norm": 3.8620858192443848, "learning_rate": 4.444500443317269e-05, "loss": 4.5999, "step": 869650 }, { "epoch": 5.244148044524306, "grad_norm": 1.2750909328460693, "learning_rate": 4.443054752044597e-05, "loss": 4.9223, "step": 869700 }, { "epoch": 5.24444953630564, "grad_norm": 4.962123394012451, "learning_rate": 4.44160925505743e-05, "loss": 4.3997, "step": 869750 }, { "epoch": 5.2447510280869745, "grad_norm": 3.6239163875579834, "learning_rate": 4.4401639523823774e-05, "loss": 4.6979, "step": 869800 }, { "epoch": 5.245052519868309, "grad_norm": 3.676541805267334, "learning_rate": 4.438718844046042e-05, "loss": 4.7536, "step": 869850 }, { "epoch": 5.245354011649642, "grad_norm": 4.3319196701049805, "learning_rate": 4.43727393007501e-05, "loss": 4.6087, "step": 869900 }, { "epoch": 5.245655503430976, "grad_norm": 4.556450843811035, "learning_rate": 4.435829210495878e-05, "loss": 4.3273, "step": 869950 }, { "epoch": 5.24595699521231, "grad_norm": 3.2920820713043213, "learning_rate": 4.434384685335239e-05, "loss": 4.7497, "step": 870000 }, { "epoch": 5.24595699521231, "eval_loss": 4.991202354431152, "eval_runtime": 39.2513, "eval_samples_per_second": 13.044, "eval_steps_per_second": 6.522, "eval_tts_loss": 7.8180199062598765, "step": 870000 }, { "epoch": 5.246258486993645, "grad_norm": 1.5558388233184814, "learning_rate": 4.43294035461967e-05, "loss": 4.1457, "step": 870050 }, { "epoch": 5.246559978774979, "grad_norm": 3.5029006004333496, "learning_rate": 4.431496218375751e-05, "loss": 4.6268, "step": 870100 }, { "epoch": 5.246861470556313, "grad_norm": 3.6913273334503174, "learning_rate": 4.4300522766300645e-05, "loss": 4.686, "step": 870150 }, { "epoch": 5.247162962337646, "grad_norm": 3.9521992206573486, "learning_rate": 4.428608529409179e-05, "loss": 5.0984, "step": 870200 }, { "epoch": 5.2474644541189805, "grad_norm": 3.3283753395080566, "learning_rate": 4.4271649767396663e-05, "loss": 4.4, "step": 870250 }, { "epoch": 5.247765945900315, "grad_norm": 4.224586486816406, "learning_rate": 4.4257216186480996e-05, "loss": 4.8022, "step": 870300 }, { "epoch": 5.248067437681649, "grad_norm": 2.7206313610076904, "learning_rate": 4.424278455161031e-05, "loss": 4.3643, "step": 870350 }, { "epoch": 5.248368929462983, "grad_norm": 4.4921088218688965, "learning_rate": 4.4228354863050285e-05, "loss": 4.2033, "step": 870400 }, { "epoch": 5.2486704212443165, "grad_norm": 3.5054023265838623, "learning_rate": 4.421392712106641e-05, "loss": 4.8136, "step": 870450 }, { "epoch": 5.248971913025651, "grad_norm": 4.946503162384033, "learning_rate": 4.419950132592429e-05, "loss": 4.472, "step": 870500 }, { "epoch": 5.249273404806985, "grad_norm": 4.2068305015563965, "learning_rate": 4.4185077477889316e-05, "loss": 4.6437, "step": 870550 }, { "epoch": 5.249574896588319, "grad_norm": 3.973602771759033, "learning_rate": 4.417065557722703e-05, "loss": 4.2449, "step": 870600 }, { "epoch": 5.249876388369653, "grad_norm": 3.9059534072875977, "learning_rate": 4.4156235624202755e-05, "loss": 4.5465, "step": 870650 }, { "epoch": 5.2501778801509875, "grad_norm": 5.737473964691162, "learning_rate": 4.414181761908191e-05, "loss": 4.4865, "step": 870700 }, { "epoch": 5.250479371932321, "grad_norm": 3.2540695667266846, "learning_rate": 4.4127401562129895e-05, "loss": 4.4723, "step": 870750 }, { "epoch": 5.250780863713655, "grad_norm": 3.60421085357666, "learning_rate": 4.411298745361193e-05, "loss": 4.5236, "step": 870800 }, { "epoch": 5.251082355494989, "grad_norm": 3.0997097492218018, "learning_rate": 4.409857529379334e-05, "loss": 4.1513, "step": 870850 }, { "epoch": 5.2513838472763235, "grad_norm": 4.181849002838135, "learning_rate": 4.4084165082939364e-05, "loss": 4.7417, "step": 870900 }, { "epoch": 5.251685339057658, "grad_norm": 4.107240200042725, "learning_rate": 4.406975682131513e-05, "loss": 4.1674, "step": 870950 }, { "epoch": 5.251986830838991, "grad_norm": 3.845790147781372, "learning_rate": 4.405535050918585e-05, "loss": 4.4616, "step": 871000 }, { "epoch": 5.252288322620325, "grad_norm": 4.860854148864746, "learning_rate": 4.4040946146816654e-05, "loss": 4.2327, "step": 871050 }, { "epoch": 5.252589814401659, "grad_norm": 5.181025505065918, "learning_rate": 4.402654373447268e-05, "loss": 4.569, "step": 871100 }, { "epoch": 5.252891306182994, "grad_norm": 3.5658113956451416, "learning_rate": 4.401214327241889e-05, "loss": 4.4735, "step": 871150 }, { "epoch": 5.253192797964328, "grad_norm": 0.6663273572921753, "learning_rate": 4.3997744760920394e-05, "loss": 3.9661, "step": 871200 }, { "epoch": 5.253494289745661, "grad_norm": 3.0847198963165283, "learning_rate": 4.398334820024213e-05, "loss": 4.5211, "step": 871250 }, { "epoch": 5.253795781526995, "grad_norm": 3.825141429901123, "learning_rate": 4.3968953590649013e-05, "loss": 4.5539, "step": 871300 }, { "epoch": 5.2540972733083295, "grad_norm": 3.4477927684783936, "learning_rate": 4.3954560932405986e-05, "loss": 4.6041, "step": 871350 }, { "epoch": 5.254398765089664, "grad_norm": 2.7212421894073486, "learning_rate": 4.394017022577796e-05, "loss": 4.4835, "step": 871400 }, { "epoch": 5.254700256870998, "grad_norm": 3.115861654281616, "learning_rate": 4.392578147102972e-05, "loss": 4.9189, "step": 871450 }, { "epoch": 5.255001748652331, "grad_norm": 4.676978588104248, "learning_rate": 4.391139466842609e-05, "loss": 4.2589, "step": 871500 }, { "epoch": 5.2553032404336655, "grad_norm": 3.503913164138794, "learning_rate": 4.3897009818231884e-05, "loss": 4.3894, "step": 871550 }, { "epoch": 5.255604732215, "grad_norm": 4.065406799316406, "learning_rate": 4.38826269207118e-05, "loss": 4.6655, "step": 871600 }, { "epoch": 5.255906223996334, "grad_norm": 3.290574550628662, "learning_rate": 4.38682459761305e-05, "loss": 4.5726, "step": 871650 }, { "epoch": 5.256207715777668, "grad_norm": 3.6259264945983887, "learning_rate": 4.385386698475269e-05, "loss": 4.2398, "step": 871700 }, { "epoch": 5.256509207559002, "grad_norm": 3.347900629043579, "learning_rate": 4.3839489946842955e-05, "loss": 4.0819, "step": 871750 }, { "epoch": 5.256810699340336, "grad_norm": 3.1782240867614746, "learning_rate": 4.3825114862665896e-05, "loss": 4.6497, "step": 871800 }, { "epoch": 5.25711219112167, "grad_norm": 3.9380435943603516, "learning_rate": 4.381074173248613e-05, "loss": 4.4271, "step": 871850 }, { "epoch": 5.257413682903004, "grad_norm": 3.2858502864837646, "learning_rate": 4.3796370556568074e-05, "loss": 4.3293, "step": 871900 }, { "epoch": 5.257715174684338, "grad_norm": 4.0193328857421875, "learning_rate": 4.378200133517627e-05, "loss": 5.0691, "step": 871950 }, { "epoch": 5.2580166664656725, "grad_norm": 3.5122807025909424, "learning_rate": 4.376763406857518e-05, "loss": 4.6289, "step": 872000 }, { "epoch": 5.258318158247006, "grad_norm": 3.845303773880005, "learning_rate": 4.375326875702919e-05, "loss": 4.1468, "step": 872050 }, { "epoch": 5.25861965002834, "grad_norm": 3.2982797622680664, "learning_rate": 4.3738905400802624e-05, "loss": 4.3075, "step": 872100 }, { "epoch": 5.258921141809674, "grad_norm": 3.18601131439209, "learning_rate": 4.3724544000159904e-05, "loss": 4.4819, "step": 872150 }, { "epoch": 5.259222633591008, "grad_norm": 3.1240477561950684, "learning_rate": 4.371018455536523e-05, "loss": 4.4761, "step": 872200 }, { "epoch": 5.259524125372343, "grad_norm": 3.0323638916015625, "learning_rate": 4.3695827066682936e-05, "loss": 4.7265, "step": 872250 }, { "epoch": 5.259825617153677, "grad_norm": 4.520359516143799, "learning_rate": 4.36814715343773e-05, "loss": 4.3726, "step": 872300 }, { "epoch": 5.26012710893501, "grad_norm": 5.503946781158447, "learning_rate": 4.366711795871239e-05, "loss": 4.3196, "step": 872350 }, { "epoch": 5.260428600716344, "grad_norm": 5.2125654220581055, "learning_rate": 4.3652766339952476e-05, "loss": 4.4261, "step": 872400 }, { "epoch": 5.2607300924976785, "grad_norm": 2.32989239692688, "learning_rate": 4.363841667836164e-05, "loss": 4.3937, "step": 872450 }, { "epoch": 5.261031584279013, "grad_norm": 3.1283936500549316, "learning_rate": 4.362406897420391e-05, "loss": 5.0011, "step": 872500 }, { "epoch": 5.261333076060347, "grad_norm": 4.111610412597656, "learning_rate": 4.3609723227743395e-05, "loss": 4.4729, "step": 872550 }, { "epoch": 5.26163456784168, "grad_norm": 4.949134349822998, "learning_rate": 4.359537943924413e-05, "loss": 4.4932, "step": 872600 }, { "epoch": 5.2619360596230145, "grad_norm": 3.357565402984619, "learning_rate": 4.3581037608970004e-05, "loss": 4.6474, "step": 872650 }, { "epoch": 5.262237551404349, "grad_norm": 2.9391403198242188, "learning_rate": 4.356669773718503e-05, "loss": 4.3571, "step": 872700 }, { "epoch": 5.262539043185683, "grad_norm": 3.1921825408935547, "learning_rate": 4.355235982415313e-05, "loss": 4.1316, "step": 872750 }, { "epoch": 5.262840534967017, "grad_norm": 3.007689952850342, "learning_rate": 4.35380238701381e-05, "loss": 4.8173, "step": 872800 }, { "epoch": 5.263142026748351, "grad_norm": 2.9515929222106934, "learning_rate": 4.3523689875403856e-05, "loss": 4.519, "step": 872850 }, { "epoch": 5.263443518529685, "grad_norm": 3.248199939727783, "learning_rate": 4.350935784021415e-05, "loss": 4.6865, "step": 872900 }, { "epoch": 5.263745010311019, "grad_norm": 4.691450595855713, "learning_rate": 4.34950277648327e-05, "loss": 4.5531, "step": 872950 }, { "epoch": 5.264046502092353, "grad_norm": 4.6253180503845215, "learning_rate": 4.348069964952327e-05, "loss": 4.253, "step": 873000 }, { "epoch": 5.264046502092353, "eval_loss": 4.993055820465088, "eval_runtime": 39.1091, "eval_samples_per_second": 13.092, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.820947945853371, "step": 873000 }, { "epoch": 5.264347993873687, "grad_norm": 3.141437292098999, "learning_rate": 4.34663734945496e-05, "loss": 4.877, "step": 873050 }, { "epoch": 5.264649485655021, "grad_norm": 3.482697010040283, "learning_rate": 4.3452049300175247e-05, "loss": 4.4044, "step": 873100 }, { "epoch": 5.264950977436355, "grad_norm": 3.787961006164551, "learning_rate": 4.3437727066663864e-05, "loss": 4.5483, "step": 873150 }, { "epoch": 5.265252469217689, "grad_norm": 3.7909038066864014, "learning_rate": 4.342340679427911e-05, "loss": 4.546, "step": 873200 }, { "epoch": 5.265553960999023, "grad_norm": 2.5081698894500732, "learning_rate": 4.340908848328444e-05, "loss": 4.7049, "step": 873250 }, { "epoch": 5.265855452780357, "grad_norm": 3.5731570720672607, "learning_rate": 4.339477213394334e-05, "loss": 4.2118, "step": 873300 }, { "epoch": 5.266156944561692, "grad_norm": 3.747868061065674, "learning_rate": 4.338045774651939e-05, "loss": 4.3769, "step": 873350 }, { "epoch": 5.266458436343025, "grad_norm": 3.0434021949768066, "learning_rate": 4.3366145321275885e-05, "loss": 4.3491, "step": 873400 }, { "epoch": 5.266759928124359, "grad_norm": 4.684616565704346, "learning_rate": 4.335183485847632e-05, "loss": 4.2432, "step": 873450 }, { "epoch": 5.267061419905693, "grad_norm": 4.156881332397461, "learning_rate": 4.333752635838409e-05, "loss": 4.6243, "step": 873500 }, { "epoch": 5.2673629116870275, "grad_norm": 3.1411890983581543, "learning_rate": 4.3323219821262415e-05, "loss": 4.7044, "step": 873550 }, { "epoch": 5.267664403468362, "grad_norm": 0.542344331741333, "learning_rate": 4.33089152473747e-05, "loss": 4.551, "step": 873600 }, { "epoch": 5.267965895249695, "grad_norm": 3.968012571334839, "learning_rate": 4.329461263698413e-05, "loss": 4.2525, "step": 873650 }, { "epoch": 5.268267387031029, "grad_norm": 2.830832004547119, "learning_rate": 4.328031199035389e-05, "loss": 4.4333, "step": 873700 }, { "epoch": 5.2685688788123635, "grad_norm": 3.096250057220459, "learning_rate": 4.326601330774722e-05, "loss": 4.5419, "step": 873750 }, { "epoch": 5.268870370593698, "grad_norm": 2.8076605796813965, "learning_rate": 4.32517165894273e-05, "loss": 4.2098, "step": 873800 }, { "epoch": 5.269171862375032, "grad_norm": 3.26393723487854, "learning_rate": 4.323742183565715e-05, "loss": 4.5811, "step": 873850 }, { "epoch": 5.269473354156366, "grad_norm": 1.5234986543655396, "learning_rate": 4.32231290466999e-05, "loss": 4.4164, "step": 873900 }, { "epoch": 5.269774845937699, "grad_norm": 4.345839023590088, "learning_rate": 4.3208838222818626e-05, "loss": 4.2368, "step": 873950 }, { "epoch": 5.270076337719034, "grad_norm": 5.911123275756836, "learning_rate": 4.319454936427623e-05, "loss": 4.8306, "step": 874000 }, { "epoch": 5.270377829500368, "grad_norm": 5.611337661743164, "learning_rate": 4.3180262471335786e-05, "loss": 4.2167, "step": 874050 }, { "epoch": 5.270679321281702, "grad_norm": 2.835338592529297, "learning_rate": 4.316597754426018e-05, "loss": 4.4457, "step": 874100 }, { "epoch": 5.270980813063036, "grad_norm": 4.986259460449219, "learning_rate": 4.315169458331225e-05, "loss": 4.6981, "step": 874150 }, { "epoch": 5.2712823048443695, "grad_norm": 4.219120025634766, "learning_rate": 4.313741358875488e-05, "loss": 4.673, "step": 874200 }, { "epoch": 5.271583796625704, "grad_norm": 3.1597182750701904, "learning_rate": 4.3123134560850974e-05, "loss": 4.6942, "step": 874250 }, { "epoch": 5.271885288407038, "grad_norm": 3.3223981857299805, "learning_rate": 4.31088574998632e-05, "loss": 4.5163, "step": 874300 }, { "epoch": 5.272186780188372, "grad_norm": 4.196834564208984, "learning_rate": 4.309458240605437e-05, "loss": 4.1305, "step": 874350 }, { "epoch": 5.272488271969706, "grad_norm": 4.837257385253906, "learning_rate": 4.308030927968724e-05, "loss": 4.3615, "step": 874400 }, { "epoch": 5.272789763751041, "grad_norm": 4.236565589904785, "learning_rate": 4.306603812102443e-05, "loss": 4.5835, "step": 874450 }, { "epoch": 5.273091255532374, "grad_norm": 2.610884428024292, "learning_rate": 4.305176893032853e-05, "loss": 4.7096, "step": 874500 }, { "epoch": 5.273392747313708, "grad_norm": 3.240065813064575, "learning_rate": 4.303750170786221e-05, "loss": 4.5893, "step": 874550 }, { "epoch": 5.273694239095042, "grad_norm": 3.5808422565460205, "learning_rate": 4.302323645388806e-05, "loss": 4.4242, "step": 874600 }, { "epoch": 5.2739957308763765, "grad_norm": 4.359694480895996, "learning_rate": 4.300897316866854e-05, "loss": 4.7048, "step": 874650 }, { "epoch": 5.274297222657711, "grad_norm": 0.8600373864173889, "learning_rate": 4.299471185246619e-05, "loss": 4.2759, "step": 874700 }, { "epoch": 5.274598714439044, "grad_norm": 5.4498610496521, "learning_rate": 4.298045250554351e-05, "loss": 4.6823, "step": 874750 }, { "epoch": 5.274900206220378, "grad_norm": 3.856264591217041, "learning_rate": 4.2966195128162876e-05, "loss": 4.4941, "step": 874800 }, { "epoch": 5.2752016980017125, "grad_norm": 3.0006513595581055, "learning_rate": 4.2951939720586633e-05, "loss": 4.3982, "step": 874850 }, { "epoch": 5.275503189783047, "grad_norm": 2.2218587398529053, "learning_rate": 4.2937686283077214e-05, "loss": 4.5682, "step": 874900 }, { "epoch": 5.275804681564381, "grad_norm": 3.01198148727417, "learning_rate": 4.292343481589684e-05, "loss": 4.6308, "step": 874950 }, { "epoch": 5.276106173345714, "grad_norm": 2.272279977798462, "learning_rate": 4.2909185319307845e-05, "loss": 4.2276, "step": 875000 }, { "epoch": 5.276407665127048, "grad_norm": 3.7043018341064453, "learning_rate": 4.2894937793572506e-05, "loss": 4.592, "step": 875050 }, { "epoch": 5.276709156908383, "grad_norm": 4.041541576385498, "learning_rate": 4.2880692238952964e-05, "loss": 4.7378, "step": 875100 }, { "epoch": 5.277010648689717, "grad_norm": 2.9784841537475586, "learning_rate": 4.286644865571139e-05, "loss": 4.8686, "step": 875150 }, { "epoch": 5.277312140471051, "grad_norm": 2.7087252140045166, "learning_rate": 4.285220704411001e-05, "loss": 4.5299, "step": 875200 }, { "epoch": 5.277613632252384, "grad_norm": 4.174684524536133, "learning_rate": 4.283796740441085e-05, "loss": 4.3356, "step": 875250 }, { "epoch": 5.2779151240337185, "grad_norm": 3.000105619430542, "learning_rate": 4.2823729736875907e-05, "loss": 4.772, "step": 875300 }, { "epoch": 5.278216615815053, "grad_norm": 1.1141407489776611, "learning_rate": 4.280949404176733e-05, "loss": 4.6882, "step": 875350 }, { "epoch": 5.278518107596387, "grad_norm": 3.6867735385894775, "learning_rate": 4.279526031934699e-05, "loss": 4.5979, "step": 875400 }, { "epoch": 5.278819599377721, "grad_norm": 4.081775188446045, "learning_rate": 4.27810285698769e-05, "loss": 4.4813, "step": 875450 }, { "epoch": 5.279121091159055, "grad_norm": 3.6375200748443604, "learning_rate": 4.276679879361902e-05, "loss": 4.7246, "step": 875500 }, { "epoch": 5.279422582940389, "grad_norm": 3.047086000442505, "learning_rate": 4.2752570990835126e-05, "loss": 4.7628, "step": 875550 }, { "epoch": 5.279724074721723, "grad_norm": 1.5755367279052734, "learning_rate": 4.2738345161787156e-05, "loss": 4.5055, "step": 875600 }, { "epoch": 5.280025566503057, "grad_norm": 2.954690456390381, "learning_rate": 4.2724121306736885e-05, "loss": 4.2542, "step": 875650 }, { "epoch": 5.280327058284391, "grad_norm": 2.651244640350342, "learning_rate": 4.270989942594602e-05, "loss": 4.4155, "step": 875700 }, { "epoch": 5.2806285500657255, "grad_norm": 4.453850746154785, "learning_rate": 4.269567951967634e-05, "loss": 4.7673, "step": 875750 }, { "epoch": 5.280930041847059, "grad_norm": 2.3143184185028076, "learning_rate": 4.268146158818959e-05, "loss": 4.2825, "step": 875800 }, { "epoch": 5.281231533628393, "grad_norm": 7.294296741485596, "learning_rate": 4.266724563174735e-05, "loss": 4.3038, "step": 875850 }, { "epoch": 5.281533025409727, "grad_norm": 3.81744122505188, "learning_rate": 4.265303165061128e-05, "loss": 4.6172, "step": 875900 }, { "epoch": 5.2818345171910615, "grad_norm": 2.651590585708618, "learning_rate": 4.263881964504301e-05, "loss": 4.5693, "step": 875950 }, { "epoch": 5.282136008972396, "grad_norm": 5.754111289978027, "learning_rate": 4.262460961530406e-05, "loss": 4.1207, "step": 876000 }, { "epoch": 5.282136008972396, "eval_loss": 4.985393524169922, "eval_runtime": 38.9824, "eval_samples_per_second": 13.134, "eval_steps_per_second": 6.567, "eval_tts_loss": 7.919702780594276, "step": 876000 }, { "epoch": 5.28243750075373, "grad_norm": 5.160268783569336, "learning_rate": 4.2610401561655876e-05, "loss": 4.7496, "step": 876050 }, { "epoch": 5.282738992535063, "grad_norm": 7.108484268188477, "learning_rate": 4.259619548436006e-05, "loss": 4.5306, "step": 876100 }, { "epoch": 5.283040484316397, "grad_norm": 4.630651473999023, "learning_rate": 4.258199138367794e-05, "loss": 4.3163, "step": 876150 }, { "epoch": 5.283341976097732, "grad_norm": 1.1447216272354126, "learning_rate": 4.2567789259870975e-05, "loss": 4.4173, "step": 876200 }, { "epoch": 5.283643467879066, "grad_norm": 4.693379878997803, "learning_rate": 4.255358911320058e-05, "loss": 4.6872, "step": 876250 }, { "epoch": 5.2839449596604, "grad_norm": 3.1272459030151367, "learning_rate": 4.2539390943928e-05, "loss": 4.5037, "step": 876300 }, { "epoch": 5.284246451441733, "grad_norm": 3.469867467880249, "learning_rate": 4.252519475231459e-05, "loss": 4.5118, "step": 876350 }, { "epoch": 5.2845479432230675, "grad_norm": 3.005127429962158, "learning_rate": 4.251100053862162e-05, "loss": 4.6823, "step": 876400 }, { "epoch": 5.284849435004402, "grad_norm": 1.74160635471344, "learning_rate": 4.24968083031103e-05, "loss": 4.6272, "step": 876450 }, { "epoch": 5.285150926785736, "grad_norm": 2.903501272201538, "learning_rate": 4.248261804604176e-05, "loss": 4.8315, "step": 876500 }, { "epoch": 5.28545241856707, "grad_norm": 3.121809482574463, "learning_rate": 4.246842976767723e-05, "loss": 4.6195, "step": 876550 }, { "epoch": 5.285753910348404, "grad_norm": 1.18062424659729, "learning_rate": 4.245424346827776e-05, "loss": 4.0765, "step": 876600 }, { "epoch": 5.286055402129738, "grad_norm": 3.2149388790130615, "learning_rate": 4.2440059148104465e-05, "loss": 4.3205, "step": 876650 }, { "epoch": 5.286356893911072, "grad_norm": 4.995577335357666, "learning_rate": 4.242587680741844e-05, "loss": 4.38, "step": 876700 }, { "epoch": 5.286658385692406, "grad_norm": 2.7512123584747314, "learning_rate": 4.241169644648058e-05, "loss": 3.9808, "step": 876750 }, { "epoch": 5.28695987747374, "grad_norm": 4.345463275909424, "learning_rate": 4.239751806555196e-05, "loss": 4.4829, "step": 876800 }, { "epoch": 5.2872613692550745, "grad_norm": 3.1472156047821045, "learning_rate": 4.238334166489346e-05, "loss": 4.6171, "step": 876850 }, { "epoch": 5.287562861036408, "grad_norm": 3.691502571105957, "learning_rate": 4.236916724476592e-05, "loss": 4.3025, "step": 876900 }, { "epoch": 5.287864352817742, "grad_norm": 3.9019174575805664, "learning_rate": 4.235499480543029e-05, "loss": 4.6769, "step": 876950 }, { "epoch": 5.288165844599076, "grad_norm": 2.0306825637817383, "learning_rate": 4.234082434714739e-05, "loss": 4.5737, "step": 877000 }, { "epoch": 5.28846733638041, "grad_norm": 2.7128424644470215, "learning_rate": 4.2326655870177925e-05, "loss": 4.2471, "step": 877050 }, { "epoch": 5.288768828161745, "grad_norm": 4.330371856689453, "learning_rate": 4.231248937478271e-05, "loss": 4.4502, "step": 877100 }, { "epoch": 5.289070319943078, "grad_norm": 3.4936342239379883, "learning_rate": 4.229832486122251e-05, "loss": 4.4855, "step": 877150 }, { "epoch": 5.289371811724412, "grad_norm": 7.770698547363281, "learning_rate": 4.228416232975788e-05, "loss": 4.6789, "step": 877200 }, { "epoch": 5.289673303505746, "grad_norm": 4.180165767669678, "learning_rate": 4.227000178064956e-05, "loss": 4.0325, "step": 877250 }, { "epoch": 5.289974795287081, "grad_norm": 3.371145248413086, "learning_rate": 4.225584321415814e-05, "loss": 4.8478, "step": 877300 }, { "epoch": 5.290276287068415, "grad_norm": 4.109340667724609, "learning_rate": 4.224168663054412e-05, "loss": 4.3368, "step": 877350 }, { "epoch": 5.290577778849748, "grad_norm": 3.4272208213806152, "learning_rate": 4.222753203006806e-05, "loss": 4.4022, "step": 877400 }, { "epoch": 5.290879270631082, "grad_norm": 4.2350358963012695, "learning_rate": 4.221337941299054e-05, "loss": 4.3551, "step": 877450 }, { "epoch": 5.2911807624124165, "grad_norm": 3.59057354927063, "learning_rate": 4.219922877957189e-05, "loss": 4.1022, "step": 877500 }, { "epoch": 5.291482254193751, "grad_norm": 3.0772271156311035, "learning_rate": 4.2185080130072604e-05, "loss": 4.6312, "step": 877550 }, { "epoch": 5.291783745975085, "grad_norm": 2.676426649093628, "learning_rate": 4.217093346475309e-05, "loss": 4.5209, "step": 877600 }, { "epoch": 5.292085237756419, "grad_norm": 4.989583492279053, "learning_rate": 4.2156788783873674e-05, "loss": 4.5536, "step": 877650 }, { "epoch": 5.2923867295377525, "grad_norm": 2.8783507347106934, "learning_rate": 4.214264608769463e-05, "loss": 4.5755, "step": 877700 }, { "epoch": 5.292688221319087, "grad_norm": 5.924379348754883, "learning_rate": 4.212850537647629e-05, "loss": 4.2861, "step": 877750 }, { "epoch": 5.292989713100421, "grad_norm": 0.9780111312866211, "learning_rate": 4.211436665047883e-05, "loss": 4.177, "step": 877800 }, { "epoch": 5.293291204881755, "grad_norm": 2.9682047367095947, "learning_rate": 4.2100229909962474e-05, "loss": 4.2615, "step": 877850 }, { "epoch": 5.293592696663089, "grad_norm": 4.606528282165527, "learning_rate": 4.208609515518743e-05, "loss": 4.4718, "step": 877900 }, { "epoch": 5.293894188444423, "grad_norm": 4.309343338012695, "learning_rate": 4.2071962386413824e-05, "loss": 4.273, "step": 877950 }, { "epoch": 5.294195680225757, "grad_norm": 0.8042331337928772, "learning_rate": 4.205783160390175e-05, "loss": 4.4765, "step": 878000 }, { "epoch": 5.294497172007091, "grad_norm": 5.462357997894287, "learning_rate": 4.20437028079112e-05, "loss": 4.1092, "step": 878050 }, { "epoch": 5.294798663788425, "grad_norm": 3.564887046813965, "learning_rate": 4.2029575998702274e-05, "loss": 4.3487, "step": 878100 }, { "epoch": 5.295100155569759, "grad_norm": 4.538468837738037, "learning_rate": 4.2015451176534863e-05, "loss": 4.7954, "step": 878150 }, { "epoch": 5.295401647351094, "grad_norm": 4.021631240844727, "learning_rate": 4.200132834166899e-05, "loss": 4.3531, "step": 878200 }, { "epoch": 5.295703139132427, "grad_norm": 3.3064355850219727, "learning_rate": 4.198720749436459e-05, "loss": 4.4986, "step": 878250 }, { "epoch": 5.296004630913761, "grad_norm": 4.305317401885986, "learning_rate": 4.1973088634881443e-05, "loss": 4.9819, "step": 878300 }, { "epoch": 5.296306122695095, "grad_norm": 5.561415195465088, "learning_rate": 4.1958971763479425e-05, "loss": 4.5323, "step": 878350 }, { "epoch": 5.29660761447643, "grad_norm": 2.636384963989258, "learning_rate": 4.1944856880418404e-05, "loss": 4.4024, "step": 878400 }, { "epoch": 5.296909106257764, "grad_norm": 3.3138203620910645, "learning_rate": 4.19307439859581e-05, "loss": 4.2524, "step": 878450 }, { "epoch": 5.297210598039097, "grad_norm": 3.6110880374908447, "learning_rate": 4.1916633080358166e-05, "loss": 4.4399, "step": 878500 }, { "epoch": 5.297512089820431, "grad_norm": 3.3335556983947754, "learning_rate": 4.190252416387841e-05, "loss": 4.448, "step": 878550 }, { "epoch": 5.2978135816017655, "grad_norm": 4.176019191741943, "learning_rate": 4.1888417236778394e-05, "loss": 4.3695, "step": 878600 }, { "epoch": 5.2981150733831, "grad_norm": 3.91545033454895, "learning_rate": 4.187431229931777e-05, "loss": 4.5787, "step": 878650 }, { "epoch": 5.298416565164434, "grad_norm": 3.559415817260742, "learning_rate": 4.186020935175618e-05, "loss": 4.4356, "step": 878700 }, { "epoch": 5.298718056945767, "grad_norm": 3.456881284713745, "learning_rate": 4.184610839435305e-05, "loss": 4.7164, "step": 878750 }, { "epoch": 5.2990195487271015, "grad_norm": 4.407569885253906, "learning_rate": 4.1832009427368e-05, "loss": 4.9159, "step": 878800 }, { "epoch": 5.299321040508436, "grad_norm": 2.833174228668213, "learning_rate": 4.181791245106046e-05, "loss": 4.4785, "step": 878850 }, { "epoch": 5.29962253228977, "grad_norm": 4.008768558502197, "learning_rate": 4.180381746568982e-05, "loss": 4.2478, "step": 878900 }, { "epoch": 5.299924024071104, "grad_norm": 3.2219347953796387, "learning_rate": 4.1789724471515496e-05, "loss": 4.984, "step": 878950 }, { "epoch": 5.300225515852437, "grad_norm": 3.6685287952423096, "learning_rate": 4.177563346879693e-05, "loss": 4.4326, "step": 879000 }, { "epoch": 5.300225515852437, "eval_loss": 4.983826637268066, "eval_runtime": 38.9941, "eval_samples_per_second": 13.13, "eval_steps_per_second": 6.565, "eval_tts_loss": 7.84213688324426, "step": 879000 }, { "epoch": 5.300527007633772, "grad_norm": 3.0377838611602783, "learning_rate": 4.176154445779332e-05, "loss": 4.7035, "step": 879050 }, { "epoch": 5.300828499415106, "grad_norm": 2.838636636734009, "learning_rate": 4.1747457438764045e-05, "loss": 4.8919, "step": 879100 }, { "epoch": 5.30112999119644, "grad_norm": 3.1543383598327637, "learning_rate": 4.1733372411968365e-05, "loss": 4.8817, "step": 879150 }, { "epoch": 5.301431482977774, "grad_norm": 4.957577705383301, "learning_rate": 4.171928937766546e-05, "loss": 4.5772, "step": 879200 }, { "epoch": 5.301732974759108, "grad_norm": 3.684844970703125, "learning_rate": 4.170520833611446e-05, "loss": 4.599, "step": 879250 }, { "epoch": 5.302034466540442, "grad_norm": 2.7679898738861084, "learning_rate": 4.16911292875746e-05, "loss": 4.4377, "step": 879300 }, { "epoch": 5.302335958321776, "grad_norm": 3.333422899246216, "learning_rate": 4.1677052232304894e-05, "loss": 4.5725, "step": 879350 }, { "epoch": 5.30263745010311, "grad_norm": 3.2938902378082275, "learning_rate": 4.1662977170564464e-05, "loss": 4.7018, "step": 879400 }, { "epoch": 5.302938941884444, "grad_norm": 3.5950870513916016, "learning_rate": 4.164890410261239e-05, "loss": 4.5964, "step": 879450 }, { "epoch": 5.303240433665779, "grad_norm": 5.45695686340332, "learning_rate": 4.1634833028707546e-05, "loss": 4.334, "step": 879500 }, { "epoch": 5.303541925447112, "grad_norm": 2.936584711074829, "learning_rate": 4.162076394910897e-05, "loss": 4.4381, "step": 879550 }, { "epoch": 5.303843417228446, "grad_norm": 4.451938629150391, "learning_rate": 4.1606696864075594e-05, "loss": 4.5711, "step": 879600 }, { "epoch": 5.30414490900978, "grad_norm": 4.280570983886719, "learning_rate": 4.159263177386628e-05, "loss": 4.7961, "step": 879650 }, { "epoch": 5.3044464007911145, "grad_norm": 3.6963179111480713, "learning_rate": 4.157856867873985e-05, "loss": 4.6521, "step": 879700 }, { "epoch": 5.304747892572449, "grad_norm": 4.187257289886475, "learning_rate": 4.156450757895516e-05, "loss": 4.8118, "step": 879750 }, { "epoch": 5.305049384353783, "grad_norm": 3.4120266437530518, "learning_rate": 4.155044847477092e-05, "loss": 5.1085, "step": 879800 }, { "epoch": 5.305350876135116, "grad_norm": 3.742276430130005, "learning_rate": 4.153639136644593e-05, "loss": 4.6538, "step": 879850 }, { "epoch": 5.3056523679164505, "grad_norm": 2.9055376052856445, "learning_rate": 4.152233625423888e-05, "loss": 4.2299, "step": 879900 }, { "epoch": 5.305953859697785, "grad_norm": 3.298097848892212, "learning_rate": 4.150828313840841e-05, "loss": 4.5071, "step": 879950 }, { "epoch": 5.306255351479119, "grad_norm": 2.3419482707977295, "learning_rate": 4.149423201921321e-05, "loss": 4.2828, "step": 880000 }, { "epoch": 5.306556843260453, "grad_norm": 3.310049057006836, "learning_rate": 4.1480182896911804e-05, "loss": 4.572, "step": 880050 }, { "epoch": 5.306858335041786, "grad_norm": 3.930898666381836, "learning_rate": 4.1466135771762735e-05, "loss": 4.5131, "step": 880100 }, { "epoch": 5.307159826823121, "grad_norm": 2.2138442993164062, "learning_rate": 4.145209064402454e-05, "loss": 4.6475, "step": 880150 }, { "epoch": 5.307461318604455, "grad_norm": 4.502129554748535, "learning_rate": 4.143804751395577e-05, "loss": 4.6179, "step": 880200 }, { "epoch": 5.307762810385789, "grad_norm": 2.977597713470459, "learning_rate": 4.142400638181475e-05, "loss": 4.5064, "step": 880250 }, { "epoch": 5.308064302167123, "grad_norm": 4.506468296051025, "learning_rate": 4.140996724785995e-05, "loss": 4.3742, "step": 880300 }, { "epoch": 5.308365793948457, "grad_norm": 5.04126501083374, "learning_rate": 4.1395930112349776e-05, "loss": 4.7617, "step": 880350 }, { "epoch": 5.308667285729791, "grad_norm": 5.146033763885498, "learning_rate": 4.138189497554254e-05, "loss": 4.0291, "step": 880400 }, { "epoch": 5.308968777511125, "grad_norm": 3.1045117378234863, "learning_rate": 4.1367861837696456e-05, "loss": 4.5853, "step": 880450 }, { "epoch": 5.309270269292459, "grad_norm": 4.35404109954834, "learning_rate": 4.1353830699069904e-05, "loss": 4.5764, "step": 880500 }, { "epoch": 5.309571761073793, "grad_norm": 3.6539876461029053, "learning_rate": 4.1339801559921e-05, "loss": 4.5718, "step": 880550 }, { "epoch": 5.309873252855128, "grad_norm": 3.7338688373565674, "learning_rate": 4.1325774420507994e-05, "loss": 4.7466, "step": 880600 }, { "epoch": 5.310174744636461, "grad_norm": 3.893888473510742, "learning_rate": 4.1311749281089056e-05, "loss": 4.3411, "step": 880650 }, { "epoch": 5.310476236417795, "grad_norm": 3.1982297897338867, "learning_rate": 4.1297726141922234e-05, "loss": 4.7313, "step": 880700 }, { "epoch": 5.310777728199129, "grad_norm": 2.3466086387634277, "learning_rate": 4.1283705003265616e-05, "loss": 4.2739, "step": 880750 }, { "epoch": 5.3110792199804635, "grad_norm": 2.403125286102295, "learning_rate": 4.126968586537732e-05, "loss": 4.5951, "step": 880800 }, { "epoch": 5.311380711761798, "grad_norm": 3.5367531776428223, "learning_rate": 4.1255668728515296e-05, "loss": 4.8782, "step": 880850 }, { "epoch": 5.311682203543131, "grad_norm": 5.505914688110352, "learning_rate": 4.124165359293745e-05, "loss": 4.452, "step": 880900 }, { "epoch": 5.311983695324465, "grad_norm": 1.298803448677063, "learning_rate": 4.12276404589018e-05, "loss": 4.6254, "step": 880950 }, { "epoch": 5.3122851871057994, "grad_norm": 2.886197090148926, "learning_rate": 4.121362932666617e-05, "loss": 4.1724, "step": 881000 }, { "epoch": 5.312586678887134, "grad_norm": 3.983116865158081, "learning_rate": 4.1199620196488434e-05, "loss": 4.3128, "step": 881050 }, { "epoch": 5.312888170668468, "grad_norm": 4.4247894287109375, "learning_rate": 4.118561306862647e-05, "loss": 4.5602, "step": 881100 }, { "epoch": 5.313189662449801, "grad_norm": 4.810932636260986, "learning_rate": 4.117160794333797e-05, "loss": 4.5544, "step": 881150 }, { "epoch": 5.313491154231135, "grad_norm": 3.395904302597046, "learning_rate": 4.115760482088075e-05, "loss": 4.5579, "step": 881200 }, { "epoch": 5.31379264601247, "grad_norm": 3.646139621734619, "learning_rate": 4.114360370151246e-05, "loss": 4.419, "step": 881250 }, { "epoch": 5.314094137793804, "grad_norm": 4.287700653076172, "learning_rate": 4.112960458549083e-05, "loss": 4.3007, "step": 881300 }, { "epoch": 5.314395629575138, "grad_norm": 4.398752689361572, "learning_rate": 4.1115607473073426e-05, "loss": 4.5827, "step": 881350 }, { "epoch": 5.314697121356472, "grad_norm": 4.907151699066162, "learning_rate": 4.1101612364517884e-05, "loss": 4.6746, "step": 881400 }, { "epoch": 5.3149986131378055, "grad_norm": 6.044816493988037, "learning_rate": 4.108761926008179e-05, "loss": 4.3894, "step": 881450 }, { "epoch": 5.31530010491914, "grad_norm": 3.081918954849243, "learning_rate": 4.1073628160022605e-05, "loss": 4.8174, "step": 881500 }, { "epoch": 5.315601596700474, "grad_norm": 4.6418352127075195, "learning_rate": 4.1059639064597886e-05, "loss": 4.1896, "step": 881550 }, { "epoch": 5.315903088481808, "grad_norm": 5.031993865966797, "learning_rate": 4.104565197406504e-05, "loss": 4.5161, "step": 881600 }, { "epoch": 5.316204580263142, "grad_norm": 1.8190028667449951, "learning_rate": 4.1031666888681445e-05, "loss": 4.5083, "step": 881650 }, { "epoch": 5.316506072044476, "grad_norm": 2.4654088020324707, "learning_rate": 4.1017683808704515e-05, "loss": 4.5491, "step": 881700 }, { "epoch": 5.31680756382581, "grad_norm": 2.8265748023986816, "learning_rate": 4.1003702734391627e-05, "loss": 4.5297, "step": 881750 }, { "epoch": 5.317109055607144, "grad_norm": 3.2572405338287354, "learning_rate": 4.0989723666e-05, "loss": 4.2132, "step": 881800 }, { "epoch": 5.317410547388478, "grad_norm": 4.76715087890625, "learning_rate": 4.097574660378694e-05, "loss": 4.7574, "step": 881850 }, { "epoch": 5.3177120391698125, "grad_norm": 5.075124740600586, "learning_rate": 4.0961771548009736e-05, "loss": 4.8336, "step": 881900 }, { "epoch": 5.318013530951147, "grad_norm": 3.444728374481201, "learning_rate": 4.094779849892546e-05, "loss": 4.3461, "step": 881950 }, { "epoch": 5.31831502273248, "grad_norm": 2.5941789150238037, "learning_rate": 4.093382745679139e-05, "loss": 4.3644, "step": 882000 }, { "epoch": 5.31831502273248, "eval_loss": 4.974609851837158, "eval_runtime": 39.0015, "eval_samples_per_second": 13.128, "eval_steps_per_second": 6.564, "eval_tts_loss": 7.824926679244531, "step": 882000 }, { "epoch": 5.318616514513814, "grad_norm": 3.8494863510131836, "learning_rate": 4.091985842186457e-05, "loss": 4.2105, "step": 882050 }, { "epoch": 5.318918006295148, "grad_norm": 4.621703147888184, "learning_rate": 4.0905891394402047e-05, "loss": 4.5369, "step": 882100 }, { "epoch": 5.319219498076483, "grad_norm": 2.4311676025390625, "learning_rate": 4.089192637466089e-05, "loss": 4.2446, "step": 882150 }, { "epoch": 5.319520989857817, "grad_norm": 0.8289532661437988, "learning_rate": 4.087796336289819e-05, "loss": 4.2743, "step": 882200 }, { "epoch": 5.31982248163915, "grad_norm": 4.193682670593262, "learning_rate": 4.08640023593708e-05, "loss": 4.173, "step": 882250 }, { "epoch": 5.320123973420484, "grad_norm": 3.3086936473846436, "learning_rate": 4.08500433643357e-05, "loss": 4.1821, "step": 882300 }, { "epoch": 5.320425465201819, "grad_norm": 3.129054307937622, "learning_rate": 4.0836086378049834e-05, "loss": 4.2469, "step": 882350 }, { "epoch": 5.320726956983153, "grad_norm": 5.895461082458496, "learning_rate": 4.082213140077001e-05, "loss": 4.2246, "step": 882400 }, { "epoch": 5.321028448764487, "grad_norm": 2.934516668319702, "learning_rate": 4.080817843275299e-05, "loss": 4.4207, "step": 882450 }, { "epoch": 5.32132994054582, "grad_norm": 4.113605976104736, "learning_rate": 4.079422747425569e-05, "loss": 3.9439, "step": 882500 }, { "epoch": 5.3216314323271545, "grad_norm": 4.517127990722656, "learning_rate": 4.0780278525534734e-05, "loss": 4.6388, "step": 882550 }, { "epoch": 5.321932924108489, "grad_norm": 3.6459670066833496, "learning_rate": 4.076633158684689e-05, "loss": 4.6598, "step": 882600 }, { "epoch": 5.322234415889823, "grad_norm": 4.280463695526123, "learning_rate": 4.0752386658448876e-05, "loss": 4.4726, "step": 882650 }, { "epoch": 5.322535907671157, "grad_norm": 1.5937223434448242, "learning_rate": 4.073844374059723e-05, "loss": 4.5122, "step": 882700 }, { "epoch": 5.3228373994524905, "grad_norm": 1.6167353391647339, "learning_rate": 4.0724502833548635e-05, "loss": 4.3738, "step": 882750 }, { "epoch": 5.323138891233825, "grad_norm": 1.2825764417648315, "learning_rate": 4.0710563937559595e-05, "loss": 4.4952, "step": 882800 }, { "epoch": 5.323440383015159, "grad_norm": 2.866180896759033, "learning_rate": 4.06966270528867e-05, "loss": 4.5229, "step": 882850 }, { "epoch": 5.323741874796493, "grad_norm": 6.0867018699646, "learning_rate": 4.068269217978636e-05, "loss": 4.5088, "step": 882900 }, { "epoch": 5.324043366577827, "grad_norm": 3.252387046813965, "learning_rate": 4.06687593185151e-05, "loss": 4.6397, "step": 882950 }, { "epoch": 5.3243448583591615, "grad_norm": 4.534121036529541, "learning_rate": 4.0654828469329266e-05, "loss": 4.2675, "step": 883000 }, { "epoch": 5.324646350140495, "grad_norm": 3.0823614597320557, "learning_rate": 4.064089963248526e-05, "loss": 4.3859, "step": 883050 }, { "epoch": 5.324947841921829, "grad_norm": 3.6461565494537354, "learning_rate": 4.062697280823949e-05, "loss": 4.6702, "step": 883100 }, { "epoch": 5.325249333703163, "grad_norm": 1.003621220588684, "learning_rate": 4.061304799684815e-05, "loss": 4.4732, "step": 883150 }, { "epoch": 5.325550825484497, "grad_norm": 2.607306718826294, "learning_rate": 4.059912519856759e-05, "loss": 4.41, "step": 883200 }, { "epoch": 5.325852317265832, "grad_norm": 2.581697940826416, "learning_rate": 4.0585204413654025e-05, "loss": 4.4915, "step": 883250 }, { "epoch": 5.326153809047165, "grad_norm": 3.9291443824768066, "learning_rate": 4.057128564236359e-05, "loss": 4.87, "step": 883300 }, { "epoch": 5.326455300828499, "grad_norm": 2.6847312450408936, "learning_rate": 4.0557368884952476e-05, "loss": 4.4058, "step": 883350 }, { "epoch": 5.326756792609833, "grad_norm": 3.101228713989258, "learning_rate": 4.054345414167683e-05, "loss": 4.2162, "step": 883400 }, { "epoch": 5.327058284391168, "grad_norm": 2.6177151203155518, "learning_rate": 4.052954141279269e-05, "loss": 4.4295, "step": 883450 }, { "epoch": 5.327359776172502, "grad_norm": 3.2127559185028076, "learning_rate": 4.051563069855609e-05, "loss": 4.5432, "step": 883500 }, { "epoch": 5.327661267953836, "grad_norm": 3.1904563903808594, "learning_rate": 4.050172199922313e-05, "loss": 4.0981, "step": 883550 }, { "epoch": 5.327962759735169, "grad_norm": 3.4744632244110107, "learning_rate": 4.048781531504971e-05, "loss": 4.5102, "step": 883600 }, { "epoch": 5.3282642515165035, "grad_norm": 3.999555826187134, "learning_rate": 4.0473910646291715e-05, "loss": 4.477, "step": 883650 }, { "epoch": 5.328565743297838, "grad_norm": 5.940770149230957, "learning_rate": 4.046000799320513e-05, "loss": 4.7057, "step": 883700 }, { "epoch": 5.328867235079172, "grad_norm": 2.9178786277770996, "learning_rate": 4.044610735604575e-05, "loss": 4.8751, "step": 883750 }, { "epoch": 5.329168726860506, "grad_norm": 2.4396841526031494, "learning_rate": 4.043220873506941e-05, "loss": 4.4672, "step": 883800 }, { "epoch": 5.3294702186418395, "grad_norm": 3.462306261062622, "learning_rate": 4.041831213053196e-05, "loss": 4.2211, "step": 883850 }, { "epoch": 5.329771710423174, "grad_norm": 0.6699211001396179, "learning_rate": 4.0404417542689044e-05, "loss": 4.3726, "step": 883900 }, { "epoch": 5.330073202204508, "grad_norm": 3.1322169303894043, "learning_rate": 4.0390524971796426e-05, "loss": 4.3684, "step": 883950 }, { "epoch": 5.330374693985842, "grad_norm": 3.692687511444092, "learning_rate": 4.037663441810982e-05, "loss": 4.5951, "step": 884000 }, { "epoch": 5.330676185767176, "grad_norm": 4.0718488693237305, "learning_rate": 4.036274588188482e-05, "loss": 4.5841, "step": 884050 }, { "epoch": 5.3309776775485105, "grad_norm": 4.031057834625244, "learning_rate": 4.034885936337696e-05, "loss": 4.6986, "step": 884100 }, { "epoch": 5.331279169329844, "grad_norm": 3.038874864578247, "learning_rate": 4.033497486284191e-05, "loss": 4.4949, "step": 884150 }, { "epoch": 5.331580661111178, "grad_norm": 1.522263765335083, "learning_rate": 4.0321092380535116e-05, "loss": 4.3261, "step": 884200 }, { "epoch": 5.331882152892512, "grad_norm": 4.945465564727783, "learning_rate": 4.0307211916712096e-05, "loss": 4.3699, "step": 884250 }, { "epoch": 5.332183644673846, "grad_norm": 4.376889705657959, "learning_rate": 4.029333347162834e-05, "loss": 4.5623, "step": 884300 }, { "epoch": 5.332485136455181, "grad_norm": 4.947981357574463, "learning_rate": 4.027945704553918e-05, "loss": 4.1749, "step": 884350 }, { "epoch": 5.332786628236514, "grad_norm": 3.048419952392578, "learning_rate": 4.0265582638700065e-05, "loss": 4.0834, "step": 884400 }, { "epoch": 5.333088120017848, "grad_norm": 4.549026966094971, "learning_rate": 4.025171025136632e-05, "loss": 4.1078, "step": 884450 }, { "epoch": 5.333389611799182, "grad_norm": 2.334355354309082, "learning_rate": 4.023783988379317e-05, "loss": 4.5181, "step": 884500 }, { "epoch": 5.333691103580517, "grad_norm": 4.898207187652588, "learning_rate": 4.0223971536235946e-05, "loss": 4.2469, "step": 884550 }, { "epoch": 5.333992595361851, "grad_norm": 2.381955146789551, "learning_rate": 4.0210105208949865e-05, "loss": 4.7233, "step": 884600 }, { "epoch": 5.334294087143184, "grad_norm": 8.7001953125, "learning_rate": 4.019624090219017e-05, "loss": 4.1758, "step": 884650 }, { "epoch": 5.334595578924518, "grad_norm": 2.9318926334381104, "learning_rate": 4.01823786162119e-05, "loss": 4.3075, "step": 884700 }, { "epoch": 5.3348970707058525, "grad_norm": 3.9460930824279785, "learning_rate": 4.01685183512703e-05, "loss": 4.1962, "step": 884750 }, { "epoch": 5.335198562487187, "grad_norm": 5.155025959014893, "learning_rate": 4.015466010762036e-05, "loss": 4.7001, "step": 884800 }, { "epoch": 5.335500054268521, "grad_norm": 2.28918194770813, "learning_rate": 4.014080388551711e-05, "loss": 4.6874, "step": 884850 }, { "epoch": 5.335801546049854, "grad_norm": 2.7403886318206787, "learning_rate": 4.01269496852156e-05, "loss": 4.6317, "step": 884900 }, { "epoch": 5.3361030378311884, "grad_norm": 4.600287914276123, "learning_rate": 4.011309750697081e-05, "loss": 4.4397, "step": 884950 }, { "epoch": 5.336404529612523, "grad_norm": 3.827352285385132, "learning_rate": 4.0099247351037605e-05, "loss": 4.3472, "step": 885000 }, { "epoch": 5.336404529612523, "eval_loss": 4.980876445770264, "eval_runtime": 39.3378, "eval_samples_per_second": 13.015, "eval_steps_per_second": 6.508, "eval_tts_loss": 7.840021756970947, "step": 885000 }, { "epoch": 5.336706021393857, "grad_norm": 3.1385300159454346, "learning_rate": 4.008539921767092e-05, "loss": 4.6877, "step": 885050 }, { "epoch": 5.337007513175191, "grad_norm": 2.99350643157959, "learning_rate": 4.0071553107125635e-05, "loss": 4.5082, "step": 885100 }, { "epoch": 5.337309004956525, "grad_norm": 4.72139835357666, "learning_rate": 4.0057709019656517e-05, "loss": 4.3046, "step": 885150 }, { "epoch": 5.337610496737859, "grad_norm": 3.914628505706787, "learning_rate": 4.004386695551839e-05, "loss": 4.5605, "step": 885200 }, { "epoch": 5.337911988519193, "grad_norm": 3.1803460121154785, "learning_rate": 4.003002691496599e-05, "loss": 4.4215, "step": 885250 }, { "epoch": 5.338213480300527, "grad_norm": 4.701462745666504, "learning_rate": 4.001618889825396e-05, "loss": 4.6551, "step": 885300 }, { "epoch": 5.338514972081861, "grad_norm": 3.084522008895874, "learning_rate": 4.000235290563701e-05, "loss": 4.7211, "step": 885350 }, { "epoch": 5.338816463863195, "grad_norm": 3.335024833679199, "learning_rate": 3.998851893736983e-05, "loss": 4.5186, "step": 885400 }, { "epoch": 5.339117955644529, "grad_norm": 3.3672378063201904, "learning_rate": 3.997468699370694e-05, "loss": 4.5174, "step": 885450 }, { "epoch": 5.339419447425863, "grad_norm": 2.2634224891662598, "learning_rate": 3.9960857074902894e-05, "loss": 4.3101, "step": 885500 }, { "epoch": 5.339720939207197, "grad_norm": 4.115593433380127, "learning_rate": 3.9947029181212286e-05, "loss": 4.3429, "step": 885550 }, { "epoch": 5.340022430988531, "grad_norm": 4.2649641036987305, "learning_rate": 3.9933203312889574e-05, "loss": 4.5076, "step": 885600 }, { "epoch": 5.340323922769866, "grad_norm": 3.546436071395874, "learning_rate": 3.991937947018911e-05, "loss": 4.5041, "step": 885650 }, { "epoch": 5.3406254145512, "grad_norm": 5.60295295715332, "learning_rate": 3.990555765336543e-05, "loss": 4.3367, "step": 885700 }, { "epoch": 5.340926906332533, "grad_norm": 3.9855518341064453, "learning_rate": 3.989173786267282e-05, "loss": 4.2616, "step": 885750 }, { "epoch": 5.341228398113867, "grad_norm": 3.720248222351074, "learning_rate": 3.987792009836561e-05, "loss": 4.3082, "step": 885800 }, { "epoch": 5.3415298898952015, "grad_norm": 4.191715240478516, "learning_rate": 3.9864104360698194e-05, "loss": 4.8498, "step": 885850 }, { "epoch": 5.341831381676536, "grad_norm": 5.4169182777404785, "learning_rate": 3.9850290649924714e-05, "loss": 4.4427, "step": 885900 }, { "epoch": 5.34213287345787, "grad_norm": 4.396515846252441, "learning_rate": 3.9836478966299486e-05, "loss": 4.5485, "step": 885950 }, { "epoch": 5.342434365239203, "grad_norm": 4.7302470207214355, "learning_rate": 3.982266931007665e-05, "loss": 4.5615, "step": 886000 }, { "epoch": 5.342735857020537, "grad_norm": 4.203479290008545, "learning_rate": 3.980886168151033e-05, "loss": 4.2823, "step": 886050 }, { "epoch": 5.343037348801872, "grad_norm": 4.231707572937012, "learning_rate": 3.979505608085462e-05, "loss": 4.6252, "step": 886100 }, { "epoch": 5.343338840583206, "grad_norm": 1.3641928434371948, "learning_rate": 3.97812525083637e-05, "loss": 4.2057, "step": 886150 }, { "epoch": 5.34364033236454, "grad_norm": 3.86933970451355, "learning_rate": 3.9767450964291504e-05, "loss": 4.7761, "step": 886200 }, { "epoch": 5.343941824145873, "grad_norm": 1.2610784769058228, "learning_rate": 3.975365144889205e-05, "loss": 4.6066, "step": 886250 }, { "epoch": 5.344243315927208, "grad_norm": 4.401784896850586, "learning_rate": 3.973985396241934e-05, "loss": 4.4679, "step": 886300 }, { "epoch": 5.344544807708542, "grad_norm": 5.311068534851074, "learning_rate": 3.972605850512725e-05, "loss": 4.6026, "step": 886350 }, { "epoch": 5.344846299489876, "grad_norm": 4.35365104675293, "learning_rate": 3.971226507726972e-05, "loss": 4.3293, "step": 886400 }, { "epoch": 5.34514779127121, "grad_norm": 4.561801433563232, "learning_rate": 3.969847367910055e-05, "loss": 4.0262, "step": 886450 }, { "epoch": 5.3454492830525435, "grad_norm": 4.4139084815979, "learning_rate": 3.968468431087353e-05, "loss": 4.4266, "step": 886500 }, { "epoch": 5.345750774833878, "grad_norm": 4.383113384246826, "learning_rate": 3.967089697284246e-05, "loss": 4.3817, "step": 886550 }, { "epoch": 5.346052266615212, "grad_norm": 4.421796798706055, "learning_rate": 3.965711166526113e-05, "loss": 4.5687, "step": 886600 }, { "epoch": 5.346353758396546, "grad_norm": 2.802805185317993, "learning_rate": 3.964332838838315e-05, "loss": 4.4158, "step": 886650 }, { "epoch": 5.34665525017788, "grad_norm": 1.5726109743118286, "learning_rate": 3.962954714246223e-05, "loss": 4.5583, "step": 886700 }, { "epoch": 5.3469567419592146, "grad_norm": 5.887285232543945, "learning_rate": 3.961576792775202e-05, "loss": 4.6638, "step": 886750 }, { "epoch": 5.347258233740548, "grad_norm": 3.226754665374756, "learning_rate": 3.960199074450607e-05, "loss": 4.1137, "step": 886800 }, { "epoch": 5.347559725521882, "grad_norm": 4.615200519561768, "learning_rate": 3.9588215592977906e-05, "loss": 4.3619, "step": 886850 }, { "epoch": 5.347861217303216, "grad_norm": 4.8035407066345215, "learning_rate": 3.9574442473421114e-05, "loss": 4.0882, "step": 886900 }, { "epoch": 5.3481627090845505, "grad_norm": 2.0811729431152344, "learning_rate": 3.956067138608906e-05, "loss": 4.3805, "step": 886950 }, { "epoch": 5.348464200865885, "grad_norm": 5.659428596496582, "learning_rate": 3.954690233123526e-05, "loss": 4.6444, "step": 887000 }, { "epoch": 5.348765692647218, "grad_norm": 4.330775737762451, "learning_rate": 3.953313530911314e-05, "loss": 4.549, "step": 887050 }, { "epoch": 5.349067184428552, "grad_norm": 1.356136679649353, "learning_rate": 3.951937031997597e-05, "loss": 4.5123, "step": 887100 }, { "epoch": 5.349368676209886, "grad_norm": 2.876807689666748, "learning_rate": 3.950560736407716e-05, "loss": 4.1301, "step": 887150 }, { "epoch": 5.349670167991221, "grad_norm": 4.581151962280273, "learning_rate": 3.9491846441669975e-05, "loss": 4.0132, "step": 887200 }, { "epoch": 5.349971659772555, "grad_norm": 5.32914400100708, "learning_rate": 3.9478087553007606e-05, "loss": 4.4096, "step": 887250 }, { "epoch": 5.350273151553889, "grad_norm": 4.948695182800293, "learning_rate": 3.9464330698343316e-05, "loss": 4.5903, "step": 887300 }, { "epoch": 5.350574643335222, "grad_norm": 3.4121015071868896, "learning_rate": 3.945057587793033e-05, "loss": 4.6514, "step": 887350 }, { "epoch": 5.350876135116557, "grad_norm": 3.2299931049346924, "learning_rate": 3.9436823092021664e-05, "loss": 4.6982, "step": 887400 }, { "epoch": 5.351177626897891, "grad_norm": 3.4370007514953613, "learning_rate": 3.9423072340870495e-05, "loss": 4.8247, "step": 887450 }, { "epoch": 5.351479118679225, "grad_norm": 3.8296995162963867, "learning_rate": 3.940932362472993e-05, "loss": 4.5169, "step": 887500 }, { "epoch": 5.351780610460559, "grad_norm": 3.5386428833007812, "learning_rate": 3.93955769438529e-05, "loss": 4.5641, "step": 887550 }, { "epoch": 5.3520821022418925, "grad_norm": 2.656191110610962, "learning_rate": 3.938183229849247e-05, "loss": 4.5595, "step": 887600 }, { "epoch": 5.352383594023227, "grad_norm": 4.567238807678223, "learning_rate": 3.936808968890155e-05, "loss": 4.692, "step": 887650 }, { "epoch": 5.352685085804561, "grad_norm": 3.1028196811676025, "learning_rate": 3.935434911533303e-05, "loss": 4.7311, "step": 887700 }, { "epoch": 5.352986577585895, "grad_norm": 2.945193290710449, "learning_rate": 3.934061057803981e-05, "loss": 4.4835, "step": 887750 }, { "epoch": 5.353288069367229, "grad_norm": 3.0814852714538574, "learning_rate": 3.9326874077274775e-05, "loss": 4.7356, "step": 887800 }, { "epoch": 5.3535895611485635, "grad_norm": 3.114187002182007, "learning_rate": 3.931313961329064e-05, "loss": 4.8611, "step": 887850 }, { "epoch": 5.353891052929897, "grad_norm": 2.716679334640503, "learning_rate": 3.9299407186340205e-05, "loss": 4.1998, "step": 887900 }, { "epoch": 5.354192544711231, "grad_norm": 5.4044647216796875, "learning_rate": 3.928567679667625e-05, "loss": 4.5749, "step": 887950 }, { "epoch": 5.354494036492565, "grad_norm": 2.5205860137939453, "learning_rate": 3.927194844455142e-05, "loss": 4.6387, "step": 888000 }, { "epoch": 5.354494036492565, "eval_loss": 4.967442512512207, "eval_runtime": 39.1112, "eval_samples_per_second": 13.091, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.919983669350635, "step": 888000 }, { "epoch": 5.3547955282738995, "grad_norm": 4.477883815765381, "learning_rate": 3.925822213021832e-05, "loss": 4.4708, "step": 888050 }, { "epoch": 5.355097020055234, "grad_norm": 2.396846055984497, "learning_rate": 3.924449785392961e-05, "loss": 5.0626, "step": 888100 }, { "epoch": 5.355398511836567, "grad_norm": 4.2632012367248535, "learning_rate": 3.92307756159379e-05, "loss": 4.4282, "step": 888150 }, { "epoch": 5.355700003617901, "grad_norm": 3.494781255722046, "learning_rate": 3.921705541649563e-05, "loss": 4.4316, "step": 888200 }, { "epoch": 5.356001495399235, "grad_norm": 2.6851940155029297, "learning_rate": 3.9203337255855374e-05, "loss": 4.1746, "step": 888250 }, { "epoch": 5.35630298718057, "grad_norm": 4.145962238311768, "learning_rate": 3.918962113426962e-05, "loss": 4.2933, "step": 888300 }, { "epoch": 5.356604478961904, "grad_norm": 3.9092843532562256, "learning_rate": 3.9175907051990745e-05, "loss": 4.7929, "step": 888350 }, { "epoch": 5.356905970743237, "grad_norm": 3.658604383468628, "learning_rate": 3.9162195009271123e-05, "loss": 4.7902, "step": 888400 }, { "epoch": 5.357207462524571, "grad_norm": 3.7803115844726562, "learning_rate": 3.9148485006363154e-05, "loss": 4.7813, "step": 888450 }, { "epoch": 5.357508954305906, "grad_norm": 4.806382179260254, "learning_rate": 3.913477704351909e-05, "loss": 4.8095, "step": 888500 }, { "epoch": 5.35781044608724, "grad_norm": 3.285597324371338, "learning_rate": 3.912107112099124e-05, "loss": 4.3524, "step": 888550 }, { "epoch": 5.358111937868574, "grad_norm": 6.767914772033691, "learning_rate": 3.9107367239031885e-05, "loss": 4.0949, "step": 888600 }, { "epoch": 5.358413429649907, "grad_norm": 4.579882621765137, "learning_rate": 3.909366539789314e-05, "loss": 4.6993, "step": 888650 }, { "epoch": 5.3587149214312415, "grad_norm": 2.1983706951141357, "learning_rate": 3.907996559782721e-05, "loss": 4.7824, "step": 888700 }, { "epoch": 5.359016413212576, "grad_norm": 2.9973602294921875, "learning_rate": 3.906626783908627e-05, "loss": 4.5088, "step": 888750 }, { "epoch": 5.35931790499391, "grad_norm": 3.4052555561065674, "learning_rate": 3.905257212192236e-05, "loss": 4.5337, "step": 888800 }, { "epoch": 5.359619396775244, "grad_norm": 3.066479206085205, "learning_rate": 3.9038878446587486e-05, "loss": 4.4042, "step": 888850 }, { "epoch": 5.359920888556578, "grad_norm": 3.728403091430664, "learning_rate": 3.902518681333375e-05, "loss": 4.804, "step": 888900 }, { "epoch": 5.360222380337912, "grad_norm": 3.7563610076904297, "learning_rate": 3.9011497222413026e-05, "loss": 4.765, "step": 888950 }, { "epoch": 5.360523872119246, "grad_norm": 5.296621799468994, "learning_rate": 3.8997809674077316e-05, "loss": 4.4148, "step": 889000 }, { "epoch": 5.36082536390058, "grad_norm": 4.057509899139404, "learning_rate": 3.898412416857855e-05, "loss": 4.408, "step": 889050 }, { "epoch": 5.361126855681914, "grad_norm": 3.0296950340270996, "learning_rate": 3.897044070616851e-05, "loss": 4.351, "step": 889100 }, { "epoch": 5.3614283474632485, "grad_norm": 3.9223313331604004, "learning_rate": 3.895675928709909e-05, "loss": 4.61, "step": 889150 }, { "epoch": 5.361729839244582, "grad_norm": 2.3848156929016113, "learning_rate": 3.894307991162206e-05, "loss": 4.439, "step": 889200 }, { "epoch": 5.362031331025916, "grad_norm": 1.4691938161849976, "learning_rate": 3.8929402579989127e-05, "loss": 4.0311, "step": 889250 }, { "epoch": 5.36233282280725, "grad_norm": 2.761249303817749, "learning_rate": 3.891572729245202e-05, "loss": 4.4782, "step": 889300 }, { "epoch": 5.362634314588584, "grad_norm": 3.5590474605560303, "learning_rate": 3.890205404926247e-05, "loss": 4.1488, "step": 889350 }, { "epoch": 5.362935806369919, "grad_norm": 1.5772309303283691, "learning_rate": 3.888838285067203e-05, "loss": 4.1047, "step": 889400 }, { "epoch": 5.363237298151253, "grad_norm": 2.3434605598449707, "learning_rate": 3.8874713696932324e-05, "loss": 4.6277, "step": 889450 }, { "epoch": 5.363538789932586, "grad_norm": 2.96457576751709, "learning_rate": 3.886104658829499e-05, "loss": 4.3741, "step": 889500 }, { "epoch": 5.36384028171392, "grad_norm": 1.1554923057556152, "learning_rate": 3.884738152501144e-05, "loss": 4.4536, "step": 889550 }, { "epoch": 5.364141773495255, "grad_norm": 2.2965519428253174, "learning_rate": 3.883371850733325e-05, "loss": 4.1377, "step": 889600 }, { "epoch": 5.364443265276589, "grad_norm": 3.310232639312744, "learning_rate": 3.882005753551184e-05, "loss": 4.3726, "step": 889650 }, { "epoch": 5.364744757057923, "grad_norm": 4.888609886169434, "learning_rate": 3.8806398609798544e-05, "loss": 4.1919, "step": 889700 }, { "epoch": 5.365046248839256, "grad_norm": 3.666712999343872, "learning_rate": 3.8792741730444816e-05, "loss": 4.498, "step": 889750 }, { "epoch": 5.3653477406205905, "grad_norm": 5.10930871963501, "learning_rate": 3.8779086897702015e-05, "loss": 4.1364, "step": 889800 }, { "epoch": 5.365649232401925, "grad_norm": 3.5550787448883057, "learning_rate": 3.8765434111821366e-05, "loss": 4.5515, "step": 889850 }, { "epoch": 5.365950724183259, "grad_norm": 4.57624626159668, "learning_rate": 3.875178337305414e-05, "loss": 4.5848, "step": 889900 }, { "epoch": 5.366252215964593, "grad_norm": 3.0069632530212402, "learning_rate": 3.873813468165164e-05, "loss": 4.2706, "step": 889950 }, { "epoch": 5.366553707745926, "grad_norm": 3.5298831462860107, "learning_rate": 3.8724488037864994e-05, "loss": 4.4536, "step": 890000 }, { "epoch": 5.366855199527261, "grad_norm": 2.0529637336730957, "learning_rate": 3.871084344194528e-05, "loss": 4.7231, "step": 890050 }, { "epoch": 5.367156691308595, "grad_norm": 7.956217288970947, "learning_rate": 3.869720089414375e-05, "loss": 4.4472, "step": 890100 }, { "epoch": 5.367458183089929, "grad_norm": 4.031665325164795, "learning_rate": 3.868356039471134e-05, "loss": 4.5762, "step": 890150 }, { "epoch": 5.367759674871263, "grad_norm": 3.102978229522705, "learning_rate": 3.866992194389913e-05, "loss": 4.2593, "step": 890200 }, { "epoch": 5.368061166652597, "grad_norm": 4.366133689880371, "learning_rate": 3.8656285541958206e-05, "loss": 4.1428, "step": 890250 }, { "epoch": 5.368362658433931, "grad_norm": 1.0888209342956543, "learning_rate": 3.864265118913938e-05, "loss": 4.0683, "step": 890300 }, { "epoch": 5.368664150215265, "grad_norm": 5.3773908615112305, "learning_rate": 3.862901888569371e-05, "loss": 4.3299, "step": 890350 }, { "epoch": 5.368965641996599, "grad_norm": 5.163787364959717, "learning_rate": 3.861538863187199e-05, "loss": 4.557, "step": 890400 }, { "epoch": 5.369267133777933, "grad_norm": 3.1791305541992188, "learning_rate": 3.860176042792506e-05, "loss": 4.2862, "step": 890450 }, { "epoch": 5.369568625559268, "grad_norm": 3.174495220184326, "learning_rate": 3.858813427410375e-05, "loss": 4.0775, "step": 890500 }, { "epoch": 5.369870117340601, "grad_norm": 7.232796669006348, "learning_rate": 3.857451017065887e-05, "loss": 3.9656, "step": 890550 }, { "epoch": 5.370171609121935, "grad_norm": 4.625428676605225, "learning_rate": 3.856088811784108e-05, "loss": 4.5872, "step": 890600 }, { "epoch": 5.370473100903269, "grad_norm": 5.110595703125, "learning_rate": 3.8547268115901126e-05, "loss": 4.4535, "step": 890650 }, { "epoch": 5.3707745926846036, "grad_norm": 4.916935443878174, "learning_rate": 3.8533650165089686e-05, "loss": 4.58, "step": 890700 }, { "epoch": 5.371076084465938, "grad_norm": 3.4406676292419434, "learning_rate": 3.85200342656573e-05, "loss": 4.1468, "step": 890750 }, { "epoch": 5.371377576247271, "grad_norm": 4.203329563140869, "learning_rate": 3.850642041785462e-05, "loss": 4.3855, "step": 890800 }, { "epoch": 5.371679068028605, "grad_norm": 3.5454835891723633, "learning_rate": 3.849280862193218e-05, "loss": 4.3681, "step": 890850 }, { "epoch": 5.3719805598099395, "grad_norm": 4.880011558532715, "learning_rate": 3.847919887814042e-05, "loss": 4.5983, "step": 890900 }, { "epoch": 5.372282051591274, "grad_norm": 2.0412821769714355, "learning_rate": 3.846559118672986e-05, "loss": 4.5904, "step": 890950 }, { "epoch": 5.372583543372608, "grad_norm": 3.621842622756958, "learning_rate": 3.845198554795096e-05, "loss": 4.5476, "step": 891000 }, { "epoch": 5.372583543372608, "eval_loss": 4.965810775756836, "eval_runtime": 39.0906, "eval_samples_per_second": 13.098, "eval_steps_per_second": 6.549, "eval_tts_loss": 7.864200174964437, "step": 891000 }, { "epoch": 5.372885035153942, "grad_norm": 3.326981782913208, "learning_rate": 3.843838196205405e-05, "loss": 4.8859, "step": 891050 }, { "epoch": 5.373186526935275, "grad_norm": 3.5861644744873047, "learning_rate": 3.8424780429289495e-05, "loss": 4.2699, "step": 891100 }, { "epoch": 5.37348801871661, "grad_norm": 2.6559441089630127, "learning_rate": 3.8411180949907675e-05, "loss": 4.7016, "step": 891150 }, { "epoch": 5.373789510497944, "grad_norm": 2.726419687271118, "learning_rate": 3.8397583524158834e-05, "loss": 4.4296, "step": 891200 }, { "epoch": 5.374091002279278, "grad_norm": 3.894496440887451, "learning_rate": 3.838398815229316e-05, "loss": 4.3924, "step": 891250 }, { "epoch": 5.374392494060612, "grad_norm": 3.5465097427368164, "learning_rate": 3.837039483456095e-05, "loss": 4.2248, "step": 891300 }, { "epoch": 5.374693985841946, "grad_norm": 2.932980537414551, "learning_rate": 3.835680357121226e-05, "loss": 4.3328, "step": 891350 }, { "epoch": 5.37499547762328, "grad_norm": 5.1074371337890625, "learning_rate": 3.834321436249728e-05, "loss": 4.4187, "step": 891400 }, { "epoch": 5.375296969404614, "grad_norm": 3.2231931686401367, "learning_rate": 3.8329627208666106e-05, "loss": 4.4823, "step": 891450 }, { "epoch": 5.375598461185948, "grad_norm": 1.5123032331466675, "learning_rate": 3.831604210996881e-05, "loss": 4.068, "step": 891500 }, { "epoch": 5.375899952967282, "grad_norm": 4.270966529846191, "learning_rate": 3.8302459066655376e-05, "loss": 4.2183, "step": 891550 }, { "epoch": 5.376201444748617, "grad_norm": 3.0356810092926025, "learning_rate": 3.828887807897576e-05, "loss": 4.5726, "step": 891600 }, { "epoch": 5.37650293652995, "grad_norm": 3.2867956161499023, "learning_rate": 3.8275299147179937e-05, "loss": 4.5539, "step": 891650 }, { "epoch": 5.376804428311284, "grad_norm": 3.6448781490325928, "learning_rate": 3.826172227151778e-05, "loss": 4.5172, "step": 891700 }, { "epoch": 5.377105920092618, "grad_norm": 4.766648292541504, "learning_rate": 3.824814745223915e-05, "loss": 5.0325, "step": 891750 }, { "epoch": 5.3774074118739525, "grad_norm": 4.3381028175354, "learning_rate": 3.823457468959392e-05, "loss": 4.5302, "step": 891800 }, { "epoch": 5.377708903655287, "grad_norm": 3.98781681060791, "learning_rate": 3.8221003983831825e-05, "loss": 4.6058, "step": 891850 }, { "epoch": 5.37801039543662, "grad_norm": 1.7837496995925903, "learning_rate": 3.820743533520262e-05, "loss": 4.5051, "step": 891900 }, { "epoch": 5.378311887217954, "grad_norm": 4.011618137359619, "learning_rate": 3.819386874395608e-05, "loss": 4.2399, "step": 891950 }, { "epoch": 5.3786133789992885, "grad_norm": 4.945970058441162, "learning_rate": 3.8180304210341824e-05, "loss": 4.1014, "step": 892000 }, { "epoch": 5.378914870780623, "grad_norm": 5.737361907958984, "learning_rate": 3.816674173460946e-05, "loss": 4.741, "step": 892050 }, { "epoch": 5.379216362561957, "grad_norm": 4.009129047393799, "learning_rate": 3.8153181317008654e-05, "loss": 4.5843, "step": 892100 }, { "epoch": 5.37951785434329, "grad_norm": 3.579990863800049, "learning_rate": 3.813962295778889e-05, "loss": 4.7884, "step": 892150 }, { "epoch": 5.379819346124624, "grad_norm": 5.118958950042725, "learning_rate": 3.8126066657199746e-05, "loss": 4.4899, "step": 892200 }, { "epoch": 5.380120837905959, "grad_norm": 5.178778648376465, "learning_rate": 3.811251241549071e-05, "loss": 4.4308, "step": 892250 }, { "epoch": 5.380422329687293, "grad_norm": 2.9985411167144775, "learning_rate": 3.809896023291118e-05, "loss": 4.6457, "step": 892300 }, { "epoch": 5.380723821468627, "grad_norm": 4.290802478790283, "learning_rate": 3.8085410109710655e-05, "loss": 4.4877, "step": 892350 }, { "epoch": 5.38102531324996, "grad_norm": 1.235442876815796, "learning_rate": 3.807186204613843e-05, "loss": 4.7106, "step": 892400 }, { "epoch": 5.381326805031295, "grad_norm": 3.5363171100616455, "learning_rate": 3.80583160424438e-05, "loss": 4.4408, "step": 892450 }, { "epoch": 5.381628296812629, "grad_norm": 4.825289726257324, "learning_rate": 3.804477209887613e-05, "loss": 4.4574, "step": 892500 }, { "epoch": 5.381929788593963, "grad_norm": 3.4774374961853027, "learning_rate": 3.8031230215684714e-05, "loss": 4.1237, "step": 892550 }, { "epoch": 5.382231280375297, "grad_norm": 4.281065464019775, "learning_rate": 3.801769039311866e-05, "loss": 4.4038, "step": 892600 }, { "epoch": 5.382532772156631, "grad_norm": 4.958920955657959, "learning_rate": 3.8004152631427195e-05, "loss": 4.4743, "step": 892650 }, { "epoch": 5.382834263937965, "grad_norm": 4.276713848114014, "learning_rate": 3.799061693085954e-05, "loss": 4.8442, "step": 892700 }, { "epoch": 5.383135755719299, "grad_norm": 7.8328704833984375, "learning_rate": 3.797708329166472e-05, "loss": 4.7158, "step": 892750 }, { "epoch": 5.383437247500633, "grad_norm": 6.339632987976074, "learning_rate": 3.796355171409178e-05, "loss": 4.4285, "step": 892800 }, { "epoch": 5.383738739281967, "grad_norm": 6.487315654754639, "learning_rate": 3.795002219838982e-05, "loss": 4.6399, "step": 892850 }, { "epoch": 5.3840402310633015, "grad_norm": 5.260771751403809, "learning_rate": 3.7936494744807775e-05, "loss": 4.3393, "step": 892900 }, { "epoch": 5.384341722844635, "grad_norm": 4.041215419769287, "learning_rate": 3.7922969353594606e-05, "loss": 4.5523, "step": 892950 }, { "epoch": 5.384643214625969, "grad_norm": 3.6324076652526855, "learning_rate": 3.7909446024999296e-05, "loss": 4.3052, "step": 893000 }, { "epoch": 5.384944706407303, "grad_norm": 3.041635036468506, "learning_rate": 3.789592475927061e-05, "loss": 4.3949, "step": 893050 }, { "epoch": 5.3852461981886375, "grad_norm": 3.9627294540405273, "learning_rate": 3.788240555665746e-05, "loss": 4.5187, "step": 893100 }, { "epoch": 5.385547689969972, "grad_norm": 3.471073865890503, "learning_rate": 3.786888841740868e-05, "loss": 4.6004, "step": 893150 }, { "epoch": 5.385849181751306, "grad_norm": 3.392207384109497, "learning_rate": 3.7855373341772986e-05, "loss": 4.4527, "step": 893200 }, { "epoch": 5.386150673532639, "grad_norm": 3.4499359130859375, "learning_rate": 3.784186032999906e-05, "loss": 4.5067, "step": 893250 }, { "epoch": 5.386452165313973, "grad_norm": 4.3100972175598145, "learning_rate": 3.7828349382335676e-05, "loss": 4.5126, "step": 893300 }, { "epoch": 5.386753657095308, "grad_norm": 5.264728546142578, "learning_rate": 3.781484049903141e-05, "loss": 4.0273, "step": 893350 }, { "epoch": 5.387055148876642, "grad_norm": 3.9798247814178467, "learning_rate": 3.78013336803349e-05, "loss": 4.6455, "step": 893400 }, { "epoch": 5.387356640657976, "grad_norm": 2.9797353744506836, "learning_rate": 3.778782892649479e-05, "loss": 4.3808, "step": 893450 }, { "epoch": 5.387658132439309, "grad_norm": 4.075174331665039, "learning_rate": 3.7774326237759485e-05, "loss": 4.3594, "step": 893500 }, { "epoch": 5.387959624220644, "grad_norm": 4.140774250030518, "learning_rate": 3.7760825614377596e-05, "loss": 4.6529, "step": 893550 }, { "epoch": 5.388261116001978, "grad_norm": 2.002833843231201, "learning_rate": 3.774732705659755e-05, "loss": 4.7711, "step": 893600 }, { "epoch": 5.388562607783312, "grad_norm": 5.050238132476807, "learning_rate": 3.7733830564667714e-05, "loss": 4.5307, "step": 893650 }, { "epoch": 5.388864099564646, "grad_norm": 4.7676520347595215, "learning_rate": 3.7720336138836503e-05, "loss": 4.7635, "step": 893700 }, { "epoch": 5.3891655913459795, "grad_norm": 4.175806045532227, "learning_rate": 3.770684377935234e-05, "loss": 4.4008, "step": 893750 }, { "epoch": 5.389467083127314, "grad_norm": 3.2691192626953125, "learning_rate": 3.7693353486463394e-05, "loss": 4.8947, "step": 893800 }, { "epoch": 5.389768574908648, "grad_norm": 5.903613090515137, "learning_rate": 3.767986526041802e-05, "loss": 4.1643, "step": 893850 }, { "epoch": 5.390070066689982, "grad_norm": 1.9170691967010498, "learning_rate": 3.766637910146449e-05, "loss": 4.6904, "step": 893900 }, { "epoch": 5.390371558471316, "grad_norm": 3.057417869567871, "learning_rate": 3.7652895009850925e-05, "loss": 4.4481, "step": 893950 }, { "epoch": 5.39067305025265, "grad_norm": 3.7252485752105713, "learning_rate": 3.763941298582546e-05, "loss": 4.5724, "step": 894000 }, { "epoch": 5.39067305025265, "eval_loss": 4.961812496185303, "eval_runtime": 39.1116, "eval_samples_per_second": 13.091, "eval_steps_per_second": 6.545, "eval_tts_loss": 7.886291087336796, "step": 894000 }, { "epoch": 5.390974542033984, "grad_norm": 1.9840335845947266, "learning_rate": 3.76259330296363e-05, "loss": 4.4273, "step": 894050 }, { "epoch": 5.391276033815318, "grad_norm": 3.1860787868499756, "learning_rate": 3.761245514153144e-05, "loss": 4.5128, "step": 894100 }, { "epoch": 5.391577525596652, "grad_norm": 3.9461781978607178, "learning_rate": 3.759897932175894e-05, "loss": 4.652, "step": 894150 }, { "epoch": 5.3918790173779865, "grad_norm": 3.835958957672119, "learning_rate": 3.7585505570566884e-05, "loss": 4.061, "step": 894200 }, { "epoch": 5.392180509159321, "grad_norm": 3.1817574501037598, "learning_rate": 3.7572033888203115e-05, "loss": 4.3183, "step": 894250 }, { "epoch": 5.392482000940654, "grad_norm": 2.2394676208496094, "learning_rate": 3.755856427491562e-05, "loss": 4.5019, "step": 894300 }, { "epoch": 5.392783492721988, "grad_norm": 3.579998016357422, "learning_rate": 3.7545096730952337e-05, "loss": 4.3885, "step": 894350 }, { "epoch": 5.393084984503322, "grad_norm": 4.487922191619873, "learning_rate": 3.753163125656107e-05, "loss": 4.2956, "step": 894400 }, { "epoch": 5.393386476284657, "grad_norm": 5.497716903686523, "learning_rate": 3.751816785198957e-05, "loss": 4.5117, "step": 894450 }, { "epoch": 5.393687968065991, "grad_norm": 2.091202974319458, "learning_rate": 3.7504706517485736e-05, "loss": 4.1667, "step": 894500 }, { "epoch": 5.393989459847324, "grad_norm": 1.8644788265228271, "learning_rate": 3.749124725329717e-05, "loss": 4.005, "step": 894550 }, { "epoch": 5.394290951628658, "grad_norm": 2.9789042472839355, "learning_rate": 3.7477790059671663e-05, "loss": 4.5348, "step": 894600 }, { "epoch": 5.3945924434099926, "grad_norm": 4.210984230041504, "learning_rate": 3.7464334936856894e-05, "loss": 4.6805, "step": 894650 }, { "epoch": 5.394893935191327, "grad_norm": 4.229691505432129, "learning_rate": 3.745088188510038e-05, "loss": 4.6902, "step": 894700 }, { "epoch": 5.395195426972661, "grad_norm": 3.871438980102539, "learning_rate": 3.7437430904649826e-05, "loss": 4.6293, "step": 894750 }, { "epoch": 5.395496918753995, "grad_norm": 3.071190118789673, "learning_rate": 3.742398199575267e-05, "loss": 4.518, "step": 894800 }, { "epoch": 5.3957984105353285, "grad_norm": 3.952174663543701, "learning_rate": 3.741053515865653e-05, "loss": 4.4241, "step": 894850 }, { "epoch": 5.396099902316663, "grad_norm": 3.650242328643799, "learning_rate": 3.739709039360875e-05, "loss": 4.5432, "step": 894900 }, { "epoch": 5.396401394097997, "grad_norm": 1.9776679277420044, "learning_rate": 3.738364770085685e-05, "loss": 4.4121, "step": 894950 }, { "epoch": 5.396702885879331, "grad_norm": 3.8623766899108887, "learning_rate": 3.737020708064824e-05, "loss": 4.3183, "step": 895000 }, { "epoch": 5.397004377660665, "grad_norm": 5.415629863739014, "learning_rate": 3.735676853323019e-05, "loss": 4.9639, "step": 895050 }, { "epoch": 5.397305869441999, "grad_norm": 3.7909562587738037, "learning_rate": 3.73433320588501e-05, "loss": 4.387, "step": 895100 }, { "epoch": 5.397607361223333, "grad_norm": 3.639801502227783, "learning_rate": 3.732989765775518e-05, "loss": 4.0982, "step": 895150 }, { "epoch": 5.397908853004667, "grad_norm": 1.381971001625061, "learning_rate": 3.7316465330192754e-05, "loss": 4.4202, "step": 895200 }, { "epoch": 5.398210344786001, "grad_norm": 4.886963844299316, "learning_rate": 3.7303035076409915e-05, "loss": 4.6535, "step": 895250 }, { "epoch": 5.3985118365673355, "grad_norm": 5.730031967163086, "learning_rate": 3.7289606896653946e-05, "loss": 4.8465, "step": 895300 }, { "epoch": 5.39881332834867, "grad_norm": 3.7177844047546387, "learning_rate": 3.7276180791171876e-05, "loss": 4.5138, "step": 895350 }, { "epoch": 5.399114820130003, "grad_norm": 3.4279966354370117, "learning_rate": 3.7262756760210815e-05, "loss": 4.8863, "step": 895400 }, { "epoch": 5.399416311911337, "grad_norm": 4.416679859161377, "learning_rate": 3.724933480401789e-05, "loss": 4.9696, "step": 895450 }, { "epoch": 5.399717803692671, "grad_norm": 4.412967681884766, "learning_rate": 3.7235914922839996e-05, "loss": 4.4697, "step": 895500 }, { "epoch": 5.400019295474006, "grad_norm": 2.1961026191711426, "learning_rate": 3.7222497116924235e-05, "loss": 4.3526, "step": 895550 }, { "epoch": 5.40032078725534, "grad_norm": 3.6386611461639404, "learning_rate": 3.720908138651745e-05, "loss": 4.3535, "step": 895600 }, { "epoch": 5.400622279036673, "grad_norm": 6.730104446411133, "learning_rate": 3.7195667731866516e-05, "loss": 4.2691, "step": 895650 }, { "epoch": 5.400923770818007, "grad_norm": 2.8093061447143555, "learning_rate": 3.718225615321835e-05, "loss": 4.086, "step": 895700 }, { "epoch": 5.4012252625993415, "grad_norm": 3.82904314994812, "learning_rate": 3.7168846650819786e-05, "loss": 4.4973, "step": 895750 }, { "epoch": 5.401526754380676, "grad_norm": 7.636054515838623, "learning_rate": 3.715543922491756e-05, "loss": 4.4084, "step": 895800 }, { "epoch": 5.40182824616201, "grad_norm": 3.254366159439087, "learning_rate": 3.714203387575842e-05, "loss": 4.9357, "step": 895850 }, { "epoch": 5.402129737943343, "grad_norm": 3.256549835205078, "learning_rate": 3.712863060358914e-05, "loss": 4.6482, "step": 895900 }, { "epoch": 5.4024312297246775, "grad_norm": 0.4439033269882202, "learning_rate": 3.711522940865633e-05, "loss": 4.2612, "step": 895950 }, { "epoch": 5.402732721506012, "grad_norm": 3.1245598793029785, "learning_rate": 3.7101830291206586e-05, "loss": 4.4372, "step": 896000 }, { "epoch": 5.403034213287346, "grad_norm": 2.6757028102874756, "learning_rate": 3.70884332514866e-05, "loss": 4.3924, "step": 896050 }, { "epoch": 5.40333570506868, "grad_norm": 4.166225910186768, "learning_rate": 3.70750382897428e-05, "loss": 4.5096, "step": 896100 }, { "epoch": 5.403637196850013, "grad_norm": 4.108001708984375, "learning_rate": 3.706164540622178e-05, "loss": 4.611, "step": 896150 }, { "epoch": 5.403938688631348, "grad_norm": 3.0871381759643555, "learning_rate": 3.704825460117004e-05, "loss": 4.2974, "step": 896200 }, { "epoch": 5.404240180412682, "grad_norm": 0.9811477661132812, "learning_rate": 3.7034865874833955e-05, "loss": 4.9448, "step": 896250 }, { "epoch": 5.404541672194016, "grad_norm": 4.2815752029418945, "learning_rate": 3.702147922745996e-05, "loss": 4.4104, "step": 896300 }, { "epoch": 5.40484316397535, "grad_norm": 3.768587827682495, "learning_rate": 3.700809465929442e-05, "loss": 4.3343, "step": 896350 }, { "epoch": 5.4051446557566845, "grad_norm": 4.027388095855713, "learning_rate": 3.699471217058369e-05, "loss": 4.2788, "step": 896400 }, { "epoch": 5.405446147538018, "grad_norm": 3.9644505977630615, "learning_rate": 3.698133176157394e-05, "loss": 4.6714, "step": 896450 }, { "epoch": 5.405747639319352, "grad_norm": 0.9846653342247009, "learning_rate": 3.6967953432511544e-05, "loss": 4.4946, "step": 896500 }, { "epoch": 5.406049131100686, "grad_norm": 3.913849115371704, "learning_rate": 3.695457718364263e-05, "loss": 4.4345, "step": 896550 }, { "epoch": 5.40635062288202, "grad_norm": 2.796626329421997, "learning_rate": 3.694120301521339e-05, "loss": 4.5805, "step": 896600 }, { "epoch": 5.406652114663355, "grad_norm": 3.2142961025238037, "learning_rate": 3.6927830927470006e-05, "loss": 4.7229, "step": 896650 }, { "epoch": 5.406953606444688, "grad_norm": 2.4453678131103516, "learning_rate": 3.691446092065848e-05, "loss": 4.5159, "step": 896700 }, { "epoch": 5.407255098226022, "grad_norm": 3.487729549407959, "learning_rate": 3.690109299502497e-05, "loss": 4.2402, "step": 896750 }, { "epoch": 5.407556590007356, "grad_norm": 3.3559813499450684, "learning_rate": 3.688772715081544e-05, "loss": 3.9366, "step": 896800 }, { "epoch": 5.4078580817886905, "grad_norm": 4.476135730743408, "learning_rate": 3.687436338827583e-05, "loss": 4.1707, "step": 896850 }, { "epoch": 5.408159573570025, "grad_norm": 4.846728324890137, "learning_rate": 3.6861001707652125e-05, "loss": 4.4511, "step": 896900 }, { "epoch": 5.408461065351359, "grad_norm": 4.535758018493652, "learning_rate": 3.684764210919027e-05, "loss": 4.7685, "step": 896950 }, { "epoch": 5.408762557132692, "grad_norm": 4.430819988250732, "learning_rate": 3.683428459313604e-05, "loss": 4.3685, "step": 897000 }, { "epoch": 5.408762557132692, "eval_loss": 4.966648101806641, "eval_runtime": 38.9771, "eval_samples_per_second": 13.136, "eval_steps_per_second": 6.568, "eval_tts_loss": 7.8394777298179354, "step": 897000 }, { "epoch": 5.4090640489140265, "grad_norm": 3.5523080825805664, "learning_rate": 3.682092915973531e-05, "loss": 4.6579, "step": 897050 }, { "epoch": 5.409365540695361, "grad_norm": 4.494332313537598, "learning_rate": 3.68075758092339e-05, "loss": 4.5907, "step": 897100 }, { "epoch": 5.409667032476695, "grad_norm": 6.6666765213012695, "learning_rate": 3.6794224541877524e-05, "loss": 4.6788, "step": 897150 }, { "epoch": 5.409968524258029, "grad_norm": 5.070326805114746, "learning_rate": 3.6780875357911865e-05, "loss": 4.3916, "step": 897200 }, { "epoch": 5.410270016039362, "grad_norm": 2.509700059890747, "learning_rate": 3.676752825758265e-05, "loss": 3.9629, "step": 897250 }, { "epoch": 5.410571507820697, "grad_norm": 4.243557929992676, "learning_rate": 3.6754183241135435e-05, "loss": 4.9225, "step": 897300 }, { "epoch": 5.410872999602031, "grad_norm": 3.866138219833374, "learning_rate": 3.6740840308815875e-05, "loss": 4.4826, "step": 897350 }, { "epoch": 5.411174491383365, "grad_norm": 4.278492450714111, "learning_rate": 3.672749946086958e-05, "loss": 4.784, "step": 897400 }, { "epoch": 5.411475983164699, "grad_norm": 3.0594065189361572, "learning_rate": 3.671416069754193e-05, "loss": 4.5026, "step": 897450 }, { "epoch": 5.411777474946033, "grad_norm": 3.4306116104125977, "learning_rate": 3.670082401907851e-05, "loss": 4.247, "step": 897500 }, { "epoch": 5.412078966727367, "grad_norm": 3.1974127292633057, "learning_rate": 3.6687489425724784e-05, "loss": 4.5703, "step": 897550 }, { "epoch": 5.412380458508701, "grad_norm": 2.9879744052886963, "learning_rate": 3.6674156917726095e-05, "loss": 4.309, "step": 897600 }, { "epoch": 5.412681950290035, "grad_norm": 0.9584088325500488, "learning_rate": 3.666082649532778e-05, "loss": 4.4808, "step": 897650 }, { "epoch": 5.412983442071369, "grad_norm": 5.0691423416137695, "learning_rate": 3.664749815877527e-05, "loss": 4.4465, "step": 897700 }, { "epoch": 5.413284933852703, "grad_norm": 1.0249704122543335, "learning_rate": 3.663417190831374e-05, "loss": 4.4825, "step": 897750 }, { "epoch": 5.413586425634037, "grad_norm": 3.648031711578369, "learning_rate": 3.662084774418849e-05, "loss": 4.6892, "step": 897800 }, { "epoch": 5.413887917415371, "grad_norm": 3.0361390113830566, "learning_rate": 3.6607525666644786e-05, "loss": 4.7146, "step": 897850 }, { "epoch": 5.414189409196705, "grad_norm": 2.7458040714263916, "learning_rate": 3.6594205675927694e-05, "loss": 4.492, "step": 897900 }, { "epoch": 5.4144909009780395, "grad_norm": 3.6762685775756836, "learning_rate": 3.658088777228247e-05, "loss": 4.711, "step": 897950 }, { "epoch": 5.414792392759374, "grad_norm": 3.794081449508667, "learning_rate": 3.6567571955954147e-05, "loss": 4.4969, "step": 898000 }, { "epoch": 5.415093884540707, "grad_norm": 4.275827407836914, "learning_rate": 3.655425822718773e-05, "loss": 4.1853, "step": 898050 }, { "epoch": 5.415395376322041, "grad_norm": 3.0558903217315674, "learning_rate": 3.6540946586228305e-05, "loss": 4.6384, "step": 898100 }, { "epoch": 5.4156968681033755, "grad_norm": 6.3887128829956055, "learning_rate": 3.652763703332085e-05, "loss": 4.7466, "step": 898150 }, { "epoch": 5.41599835988471, "grad_norm": 4.205146789550781, "learning_rate": 3.651432956871032e-05, "loss": 4.5696, "step": 898200 }, { "epoch": 5.416299851666044, "grad_norm": 4.1699347496032715, "learning_rate": 3.6501024192641585e-05, "loss": 4.0538, "step": 898250 }, { "epoch": 5.416601343447377, "grad_norm": 8.746088981628418, "learning_rate": 3.648772090535956e-05, "loss": 4.6442, "step": 898300 }, { "epoch": 5.416902835228711, "grad_norm": 3.495927333831787, "learning_rate": 3.6474419707109037e-05, "loss": 4.6047, "step": 898350 }, { "epoch": 5.417204327010046, "grad_norm": 3.527554512023926, "learning_rate": 3.6461120598134766e-05, "loss": 4.622, "step": 898400 }, { "epoch": 5.41750581879138, "grad_norm": 3.0464444160461426, "learning_rate": 3.644782357868154e-05, "loss": 4.2269, "step": 898450 }, { "epoch": 5.417807310572714, "grad_norm": 4.281116962432861, "learning_rate": 3.643452864899414e-05, "loss": 4.3293, "step": 898500 }, { "epoch": 5.418108802354048, "grad_norm": 1.2371809482574463, "learning_rate": 3.642123580931711e-05, "loss": 4.4291, "step": 898550 }, { "epoch": 5.4184102941353816, "grad_norm": 3.3697636127471924, "learning_rate": 3.640794505989514e-05, "loss": 4.3242, "step": 898600 }, { "epoch": 5.418711785916716, "grad_norm": 4.033542156219482, "learning_rate": 3.639465640097291e-05, "loss": 4.8567, "step": 898650 }, { "epoch": 5.41901327769805, "grad_norm": 1.2668979167938232, "learning_rate": 3.638136983279483e-05, "loss": 4.254, "step": 898700 }, { "epoch": 5.419314769479384, "grad_norm": 3.740993022918701, "learning_rate": 3.636808535560556e-05, "loss": 4.5795, "step": 898750 }, { "epoch": 5.419616261260718, "grad_norm": 5.163695812225342, "learning_rate": 3.63548029696495e-05, "loss": 4.3561, "step": 898800 }, { "epoch": 5.419917753042052, "grad_norm": 3.2793314456939697, "learning_rate": 3.634152267517109e-05, "loss": 4.5727, "step": 898850 }, { "epoch": 5.420219244823386, "grad_norm": 4.2296576499938965, "learning_rate": 3.632824447241474e-05, "loss": 4.5387, "step": 898900 }, { "epoch": 5.42052073660472, "grad_norm": 3.6833860874176025, "learning_rate": 3.6314968361624873e-05, "loss": 4.4474, "step": 898950 }, { "epoch": 5.420822228386054, "grad_norm": 3.8123929500579834, "learning_rate": 3.6301694343045744e-05, "loss": 4.3306, "step": 899000 }, { "epoch": 5.4211237201673885, "grad_norm": 5.990545749664307, "learning_rate": 3.6288422416921675e-05, "loss": 4.9412, "step": 899050 }, { "epoch": 5.421425211948723, "grad_norm": 3.26676082611084, "learning_rate": 3.627515258349695e-05, "loss": 4.8922, "step": 899100 }, { "epoch": 5.421726703730056, "grad_norm": 7.35116720199585, "learning_rate": 3.626188484301575e-05, "loss": 4.2325, "step": 899150 }, { "epoch": 5.42202819551139, "grad_norm": 4.889737129211426, "learning_rate": 3.62486191957222e-05, "loss": 4.5526, "step": 899200 }, { "epoch": 5.4223296872927245, "grad_norm": 1.4834468364715576, "learning_rate": 3.623535564186053e-05, "loss": 4.4517, "step": 899250 }, { "epoch": 5.422631179074059, "grad_norm": 3.938066005706787, "learning_rate": 3.622209418167476e-05, "loss": 4.5918, "step": 899300 }, { "epoch": 5.422932670855393, "grad_norm": 3.274956226348877, "learning_rate": 3.620883481540896e-05, "loss": 4.134, "step": 899350 }, { "epoch": 5.423234162636726, "grad_norm": 3.3438994884490967, "learning_rate": 3.61955775433072e-05, "loss": 4.4398, "step": 899400 }, { "epoch": 5.42353565441806, "grad_norm": 2.972778797149658, "learning_rate": 3.618232236561341e-05, "loss": 4.6528, "step": 899450 }, { "epoch": 5.423837146199395, "grad_norm": 4.175857067108154, "learning_rate": 3.616906928257157e-05, "loss": 4.3142, "step": 899500 }, { "epoch": 5.424138637980729, "grad_norm": 3.2063205242156982, "learning_rate": 3.615581829442553e-05, "loss": 4.4861, "step": 899550 }, { "epoch": 5.424440129762063, "grad_norm": 4.179772853851318, "learning_rate": 3.6142569401419245e-05, "loss": 4.7808, "step": 899600 }, { "epoch": 5.424741621543396, "grad_norm": 4.937626361846924, "learning_rate": 3.612932260379643e-05, "loss": 4.4161, "step": 899650 }, { "epoch": 5.4250431133247305, "grad_norm": 3.415754795074463, "learning_rate": 3.611607790180097e-05, "loss": 4.2497, "step": 899700 }, { "epoch": 5.425344605106065, "grad_norm": 4.243832111358643, "learning_rate": 3.610283529567654e-05, "loss": 4.2597, "step": 899750 }, { "epoch": 5.425646096887399, "grad_norm": 3.7786080837249756, "learning_rate": 3.608959478566688e-05, "loss": 4.66, "step": 899800 }, { "epoch": 5.425947588668733, "grad_norm": 4.087085723876953, "learning_rate": 3.6076356372015724e-05, "loss": 4.6091, "step": 899850 }, { "epoch": 5.4262490804500665, "grad_norm": 3.213265895843506, "learning_rate": 3.6063120054966596e-05, "loss": 4.3147, "step": 899900 }, { "epoch": 5.426550572231401, "grad_norm": 3.6542916297912598, "learning_rate": 3.6049885834763195e-05, "loss": 4.3193, "step": 899950 }, { "epoch": 5.426852064012735, "grad_norm": 4.111680030822754, "learning_rate": 3.603665371164903e-05, "loss": 4.5334, "step": 900000 }, { "epoch": 5.426852064012735, "eval_loss": 4.96249532699585, "eval_runtime": 39.1614, "eval_samples_per_second": 13.074, "eval_steps_per_second": 6.537, "eval_tts_loss": 7.858001226507847, "step": 900000 }, { "epoch": 5.427153555794069, "grad_norm": 3.2655913829803467, "learning_rate": 3.602342368586759e-05, "loss": 4.3897, "step": 900050 }, { "epoch": 5.427455047575403, "grad_norm": 4.105524063110352, "learning_rate": 3.6010195757662355e-05, "loss": 4.767, "step": 900100 }, { "epoch": 5.4277565393567375, "grad_norm": 1.1822668313980103, "learning_rate": 3.599696992727687e-05, "loss": 4.3568, "step": 900150 }, { "epoch": 5.428058031138071, "grad_norm": 4.869379043579102, "learning_rate": 3.5983746194954396e-05, "loss": 4.538, "step": 900200 }, { "epoch": 5.428359522919405, "grad_norm": 5.631143093109131, "learning_rate": 3.597052456093837e-05, "loss": 4.4054, "step": 900250 }, { "epoch": 5.428661014700739, "grad_norm": 3.4980030059814453, "learning_rate": 3.595730502547215e-05, "loss": 4.5699, "step": 900300 }, { "epoch": 5.4289625064820735, "grad_norm": 3.7929811477661133, "learning_rate": 3.594408758879897e-05, "loss": 4.411, "step": 900350 }, { "epoch": 5.429263998263408, "grad_norm": 3.685476064682007, "learning_rate": 3.593087225116207e-05, "loss": 4.7738, "step": 900400 }, { "epoch": 5.429565490044741, "grad_norm": 3.4926414489746094, "learning_rate": 3.591765901280473e-05, "loss": 4.9775, "step": 900450 }, { "epoch": 5.429866981826075, "grad_norm": 3.1115596294403076, "learning_rate": 3.590444787397001e-05, "loss": 4.2847, "step": 900500 }, { "epoch": 5.430168473607409, "grad_norm": 3.5464203357696533, "learning_rate": 3.58912388349011e-05, "loss": 4.4721, "step": 900550 }, { "epoch": 5.430469965388744, "grad_norm": 3.291388511657715, "learning_rate": 3.5878031895841163e-05, "loss": 4.4489, "step": 900600 }, { "epoch": 5.430771457170078, "grad_norm": 2.306464910507202, "learning_rate": 3.586482705703313e-05, "loss": 4.3481, "step": 900650 }, { "epoch": 5.431072948951412, "grad_norm": 2.0776240825653076, "learning_rate": 3.585162431872007e-05, "loss": 4.1103, "step": 900700 }, { "epoch": 5.431374440732745, "grad_norm": 4.7266764640808105, "learning_rate": 3.5838423681145036e-05, "loss": 4.4475, "step": 900750 }, { "epoch": 5.4316759325140795, "grad_norm": 3.4431803226470947, "learning_rate": 3.582522514455087e-05, "loss": 4.4485, "step": 900800 }, { "epoch": 5.431977424295414, "grad_norm": 3.95350980758667, "learning_rate": 3.581202870918046e-05, "loss": 4.3964, "step": 900850 }, { "epoch": 5.432278916076748, "grad_norm": 3.327208995819092, "learning_rate": 3.579883437527676e-05, "loss": 4.4463, "step": 900900 }, { "epoch": 5.432580407858082, "grad_norm": 1.8950066566467285, "learning_rate": 3.578564214308249e-05, "loss": 4.5512, "step": 900950 }, { "epoch": 5.4328818996394155, "grad_norm": 2.906675338745117, "learning_rate": 3.577245201284049e-05, "loss": 4.5145, "step": 901000 }, { "epoch": 5.43318339142075, "grad_norm": 3.7719881534576416, "learning_rate": 3.5759263984793526e-05, "loss": 4.2199, "step": 901050 }, { "epoch": 5.433484883202084, "grad_norm": 3.298153877258301, "learning_rate": 3.5746078059184245e-05, "loss": 4.3545, "step": 901100 }, { "epoch": 5.433786374983418, "grad_norm": 5.126485347747803, "learning_rate": 3.57328942362554e-05, "loss": 4.5031, "step": 901150 }, { "epoch": 5.434087866764752, "grad_norm": 4.571654796600342, "learning_rate": 3.5719712516249555e-05, "loss": 4.2403, "step": 901200 }, { "epoch": 5.434389358546086, "grad_norm": 2.3414459228515625, "learning_rate": 3.5706532899409286e-05, "loss": 4.3932, "step": 901250 }, { "epoch": 5.43469085032742, "grad_norm": 4.602268695831299, "learning_rate": 3.569335538597716e-05, "loss": 4.4703, "step": 901300 }, { "epoch": 5.434992342108754, "grad_norm": 4.461195945739746, "learning_rate": 3.568017997619574e-05, "loss": 4.2971, "step": 901350 }, { "epoch": 5.435293833890088, "grad_norm": 3.231722831726074, "learning_rate": 3.5667006670307436e-05, "loss": 4.6966, "step": 901400 }, { "epoch": 5.4355953256714225, "grad_norm": 3.7718276977539062, "learning_rate": 3.565383546855471e-05, "loss": 4.8241, "step": 901450 }, { "epoch": 5.435896817452756, "grad_norm": 4.2951483726501465, "learning_rate": 3.564066637118e-05, "loss": 4.5564, "step": 901500 }, { "epoch": 5.43619830923409, "grad_norm": 4.856678485870361, "learning_rate": 3.5627499378425645e-05, "loss": 4.6225, "step": 901550 }, { "epoch": 5.436499801015424, "grad_norm": 4.3116631507873535, "learning_rate": 3.5614334490533885e-05, "loss": 4.6563, "step": 901600 }, { "epoch": 5.436801292796758, "grad_norm": 4.135039806365967, "learning_rate": 3.560117170774705e-05, "loss": 4.4481, "step": 901650 }, { "epoch": 5.437102784578093, "grad_norm": 5.877639293670654, "learning_rate": 3.5588011030307465e-05, "loss": 4.4609, "step": 901700 }, { "epoch": 5.437404276359427, "grad_norm": 3.099806070327759, "learning_rate": 3.557485245845719e-05, "loss": 4.3042, "step": 901750 }, { "epoch": 5.43770576814076, "grad_norm": 3.240412950515747, "learning_rate": 3.556169599243848e-05, "loss": 4.78, "step": 901800 }, { "epoch": 5.438007259922094, "grad_norm": 3.341217041015625, "learning_rate": 3.554854163249349e-05, "loss": 4.2634, "step": 901850 }, { "epoch": 5.4383087517034285, "grad_norm": 4.459360122680664, "learning_rate": 3.55353893788642e-05, "loss": 4.662, "step": 901900 }, { "epoch": 5.438610243484763, "grad_norm": 4.767136096954346, "learning_rate": 3.552223923179277e-05, "loss": 4.542, "step": 901950 }, { "epoch": 5.438911735266097, "grad_norm": 5.026338577270508, "learning_rate": 3.550909119152116e-05, "loss": 4.6646, "step": 902000 }, { "epoch": 5.43921322704743, "grad_norm": 5.929046630859375, "learning_rate": 3.549594525829131e-05, "loss": 4.3543, "step": 902050 }, { "epoch": 5.4395147188287645, "grad_norm": 3.3129379749298096, "learning_rate": 3.5482801432345154e-05, "loss": 4.5041, "step": 902100 }, { "epoch": 5.439816210610099, "grad_norm": 5.045109748840332, "learning_rate": 3.546965971392467e-05, "loss": 4.747, "step": 902150 }, { "epoch": 5.440117702391433, "grad_norm": 4.870711803436279, "learning_rate": 3.5456520103271615e-05, "loss": 4.2501, "step": 902200 }, { "epoch": 5.440419194172767, "grad_norm": 4.503072738647461, "learning_rate": 3.544338260062784e-05, "loss": 4.4376, "step": 902250 }, { "epoch": 5.440720685954101, "grad_norm": 3.5465738773345947, "learning_rate": 3.5430247206235176e-05, "loss": 4.5685, "step": 902300 }, { "epoch": 5.441022177735435, "grad_norm": 4.844499111175537, "learning_rate": 3.5417113920335306e-05, "loss": 4.2485, "step": 902350 }, { "epoch": 5.441323669516769, "grad_norm": 3.9390268325805664, "learning_rate": 3.540398274316991e-05, "loss": 4.5676, "step": 902400 }, { "epoch": 5.441625161298103, "grad_norm": 3.4187724590301514, "learning_rate": 3.53908536749807e-05, "loss": 3.9799, "step": 902450 }, { "epoch": 5.441926653079437, "grad_norm": 4.154310703277588, "learning_rate": 3.537772671600923e-05, "loss": 4.3189, "step": 902500 }, { "epoch": 5.442228144860771, "grad_norm": 0.6559949517250061, "learning_rate": 3.5364601866497113e-05, "loss": 3.9117, "step": 902550 }, { "epoch": 5.442529636642105, "grad_norm": 6.925079345703125, "learning_rate": 3.5351479126685966e-05, "loss": 4.5697, "step": 902600 }, { "epoch": 5.442831128423439, "grad_norm": 3.877121686935425, "learning_rate": 3.5338358496817186e-05, "loss": 4.7179, "step": 902650 }, { "epoch": 5.443132620204773, "grad_norm": 1.2863895893096924, "learning_rate": 3.5325239977132334e-05, "loss": 4.3831, "step": 902700 }, { "epoch": 5.443434111986107, "grad_norm": 3.2634615898132324, "learning_rate": 3.531212356787279e-05, "loss": 4.67, "step": 902750 }, { "epoch": 5.443735603767442, "grad_norm": 4.459770679473877, "learning_rate": 3.529900926927989e-05, "loss": 4.3754, "step": 902800 }, { "epoch": 5.444037095548776, "grad_norm": 2.7112388610839844, "learning_rate": 3.528589708159505e-05, "loss": 4.3948, "step": 902850 }, { "epoch": 5.444338587330109, "grad_norm": 3.872079849243164, "learning_rate": 3.527278700505961e-05, "loss": 4.6193, "step": 902900 }, { "epoch": 5.444640079111443, "grad_norm": 3.2412707805633545, "learning_rate": 3.5259679039914754e-05, "loss": 4.8732, "step": 902950 }, { "epoch": 5.4449415708927775, "grad_norm": 2.227874755859375, "learning_rate": 3.524657318640176e-05, "loss": 4.4101, "step": 903000 }, { "epoch": 5.4449415708927775, "eval_loss": 4.961116790771484, "eval_runtime": 39.2193, "eval_samples_per_second": 13.055, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.876276139777627, "step": 903000 }, { "epoch": 5.445243062674112, "grad_norm": 3.0062496662139893, "learning_rate": 3.5233469444761865e-05, "loss": 4.5441, "step": 903050 }, { "epoch": 5.445544554455446, "grad_norm": 5.547059535980225, "learning_rate": 3.522036781523615e-05, "loss": 4.7049, "step": 903100 }, { "epoch": 5.445846046236779, "grad_norm": 2.3544578552246094, "learning_rate": 3.520726829806581e-05, "loss": 4.2073, "step": 903150 }, { "epoch": 5.4461475380181135, "grad_norm": 2.4694595336914062, "learning_rate": 3.519417089349187e-05, "loss": 4.2214, "step": 903200 }, { "epoch": 5.446449029799448, "grad_norm": 3.0309059619903564, "learning_rate": 3.518107560175533e-05, "loss": 4.7423, "step": 903250 }, { "epoch": 5.446750521580782, "grad_norm": 4.5818376541137695, "learning_rate": 3.516798242309726e-05, "loss": 4.7081, "step": 903300 }, { "epoch": 5.447052013362116, "grad_norm": 1.8890374898910522, "learning_rate": 3.5154891357758616e-05, "loss": 4.7117, "step": 903350 }, { "epoch": 5.447353505143449, "grad_norm": 1.668522596359253, "learning_rate": 3.514180240598027e-05, "loss": 4.5048, "step": 903400 }, { "epoch": 5.447654996924784, "grad_norm": 5.577354431152344, "learning_rate": 3.5128715568003146e-05, "loss": 4.5561, "step": 903450 }, { "epoch": 5.447956488706118, "grad_norm": 4.226103782653809, "learning_rate": 3.5115630844068124e-05, "loss": 4.4744, "step": 903500 }, { "epoch": 5.448257980487452, "grad_norm": 1.4766138792037964, "learning_rate": 3.510254823441597e-05, "loss": 4.4966, "step": 903550 }, { "epoch": 5.448559472268786, "grad_norm": 5.26249361038208, "learning_rate": 3.508946773928739e-05, "loss": 4.4484, "step": 903600 }, { "epoch": 5.4488609640501195, "grad_norm": 4.109043121337891, "learning_rate": 3.507638935892324e-05, "loss": 4.7623, "step": 903650 }, { "epoch": 5.449162455831454, "grad_norm": 4.843825817108154, "learning_rate": 3.5063313093564075e-05, "loss": 4.6269, "step": 903700 }, { "epoch": 5.449463947612788, "grad_norm": 3.4332337379455566, "learning_rate": 3.505023894345061e-05, "loss": 4.928, "step": 903750 }, { "epoch": 5.449765439394122, "grad_norm": 3.5912728309631348, "learning_rate": 3.503716690882351e-05, "loss": 4.3254, "step": 903800 }, { "epoch": 5.450066931175456, "grad_norm": 1.5284873247146606, "learning_rate": 3.5024096989923236e-05, "loss": 4.8376, "step": 903850 }, { "epoch": 5.450368422956791, "grad_norm": 6.35093355178833, "learning_rate": 3.501102918699042e-05, "loss": 4.3572, "step": 903900 }, { "epoch": 5.450669914738124, "grad_norm": 3.8806118965148926, "learning_rate": 3.4997963500265524e-05, "loss": 4.499, "step": 903950 }, { "epoch": 5.450971406519458, "grad_norm": 1.437355399131775, "learning_rate": 3.4984899929988943e-05, "loss": 4.6982, "step": 904000 }, { "epoch": 5.451272898300792, "grad_norm": 3.959723711013794, "learning_rate": 3.4971838476401145e-05, "loss": 4.5244, "step": 904050 }, { "epoch": 5.4515743900821265, "grad_norm": 4.617344379425049, "learning_rate": 3.495877913974255e-05, "loss": 4.235, "step": 904100 }, { "epoch": 5.451875881863461, "grad_norm": 3.454679250717163, "learning_rate": 3.494572192025342e-05, "loss": 4.2669, "step": 904150 }, { "epoch": 5.452177373644794, "grad_norm": 2.7187490463256836, "learning_rate": 3.493266681817407e-05, "loss": 4.4082, "step": 904200 }, { "epoch": 5.452478865426128, "grad_norm": 1.4452017545700073, "learning_rate": 3.491961383374482e-05, "loss": 4.5006, "step": 904250 }, { "epoch": 5.4527803572074625, "grad_norm": 2.0603749752044678, "learning_rate": 3.4906562967205826e-05, "loss": 4.5936, "step": 904300 }, { "epoch": 5.453081848988797, "grad_norm": 4.254958152770996, "learning_rate": 3.489351421879732e-05, "loss": 4.4184, "step": 904350 }, { "epoch": 5.453383340770131, "grad_norm": 3.8961100578308105, "learning_rate": 3.488046758875942e-05, "loss": 4.5438, "step": 904400 }, { "epoch": 5.453684832551465, "grad_norm": 5.940357685089111, "learning_rate": 3.4867423077332184e-05, "loss": 4.5282, "step": 904450 }, { "epoch": 5.453986324332798, "grad_norm": 3.582909345626831, "learning_rate": 3.4854380684755734e-05, "loss": 4.2231, "step": 904500 }, { "epoch": 5.454287816114133, "grad_norm": 3.111865758895874, "learning_rate": 3.4841340411270116e-05, "loss": 4.6435, "step": 904550 }, { "epoch": 5.454589307895467, "grad_norm": 2.9436073303222656, "learning_rate": 3.482830225711525e-05, "loss": 4.5675, "step": 904600 }, { "epoch": 5.454890799676801, "grad_norm": 2.954437732696533, "learning_rate": 3.481526622253113e-05, "loss": 4.5246, "step": 904650 }, { "epoch": 5.455192291458135, "grad_norm": 3.2362167835235596, "learning_rate": 3.480223230775769e-05, "loss": 4.5088, "step": 904700 }, { "epoch": 5.4554937832394685, "grad_norm": 4.760593414306641, "learning_rate": 3.478920051303477e-05, "loss": 4.3112, "step": 904750 }, { "epoch": 5.455795275020803, "grad_norm": 3.138737916946411, "learning_rate": 3.477617083860215e-05, "loss": 4.3378, "step": 904800 }, { "epoch": 5.456096766802137, "grad_norm": 3.796456813812256, "learning_rate": 3.476314328469969e-05, "loss": 4.8693, "step": 904850 }, { "epoch": 5.456398258583471, "grad_norm": 5.568893909454346, "learning_rate": 3.475011785156715e-05, "loss": 4.4292, "step": 904900 }, { "epoch": 5.456699750364805, "grad_norm": 3.5476982593536377, "learning_rate": 3.47370945394442e-05, "loss": 4.5029, "step": 904950 }, { "epoch": 5.457001242146139, "grad_norm": 1.0314373970031738, "learning_rate": 3.472407334857053e-05, "loss": 4.4158, "step": 905000 }, { "epoch": 5.457302733927473, "grad_norm": 4.493794918060303, "learning_rate": 3.471105427918581e-05, "loss": 4.2654, "step": 905050 }, { "epoch": 5.457604225708807, "grad_norm": 3.8877792358398438, "learning_rate": 3.4698037331529624e-05, "loss": 3.9415, "step": 905100 }, { "epoch": 5.457905717490141, "grad_norm": 3.784299612045288, "learning_rate": 3.468502250584147e-05, "loss": 4.4777, "step": 905150 }, { "epoch": 5.4582072092714755, "grad_norm": 3.6176671981811523, "learning_rate": 3.4672009802360945e-05, "loss": 4.7543, "step": 905200 }, { "epoch": 5.458508701052809, "grad_norm": 1.895725131034851, "learning_rate": 3.465899922132748e-05, "loss": 4.2721, "step": 905250 }, { "epoch": 5.458810192834143, "grad_norm": 4.0160651206970215, "learning_rate": 3.464599076298052e-05, "loss": 4.6455, "step": 905300 }, { "epoch": 5.459111684615477, "grad_norm": 5.885619163513184, "learning_rate": 3.463298442755951e-05, "loss": 4.7004, "step": 905350 }, { "epoch": 5.4594131763968115, "grad_norm": 6.0631303787231445, "learning_rate": 3.461998021530375e-05, "loss": 4.5909, "step": 905400 }, { "epoch": 5.459714668178146, "grad_norm": 5.7933831214904785, "learning_rate": 3.460697812645261e-05, "loss": 4.4485, "step": 905450 }, { "epoch": 5.46001615995948, "grad_norm": 3.2567028999328613, "learning_rate": 3.459397816124539e-05, "loss": 4.5307, "step": 905500 }, { "epoch": 5.460317651740813, "grad_norm": 2.9781150817871094, "learning_rate": 3.458098031992133e-05, "loss": 4.3967, "step": 905550 }, { "epoch": 5.460619143522147, "grad_norm": 3.7115092277526855, "learning_rate": 3.4567984602719556e-05, "loss": 4.1717, "step": 905600 }, { "epoch": 5.460920635303482, "grad_norm": 3.8602216243743896, "learning_rate": 3.455499100987934e-05, "loss": 4.1059, "step": 905650 }, { "epoch": 5.461222127084816, "grad_norm": 3.0719165802001953, "learning_rate": 3.454199954163972e-05, "loss": 4.614, "step": 905700 }, { "epoch": 5.46152361886615, "grad_norm": 4.087300777435303, "learning_rate": 3.4529010198239837e-05, "loss": 4.509, "step": 905750 }, { "epoch": 5.461825110647483, "grad_norm": 0.7911138534545898, "learning_rate": 3.451602297991877e-05, "loss": 4.5623, "step": 905800 }, { "epoch": 5.4621266024288175, "grad_norm": 3.893430709838867, "learning_rate": 3.4503037886915466e-05, "loss": 4.2587, "step": 905850 }, { "epoch": 5.462428094210152, "grad_norm": 6.221452236175537, "learning_rate": 3.449005491946895e-05, "loss": 4.7712, "step": 905900 }, { "epoch": 5.462729585991486, "grad_norm": 3.304131031036377, "learning_rate": 3.4477074077818155e-05, "loss": 4.3899, "step": 905950 }, { "epoch": 5.46303107777282, "grad_norm": 4.3532395362854, "learning_rate": 3.4464095362201894e-05, "loss": 4.7423, "step": 906000 }, { "epoch": 5.46303107777282, "eval_loss": 4.9569807052612305, "eval_runtime": 38.8, "eval_samples_per_second": 13.196, "eval_steps_per_second": 6.598, "eval_tts_loss": 7.889740264043441, "step": 906000 }, { "epoch": 5.463332569554154, "grad_norm": 5.005142688751221, "learning_rate": 3.445111877285909e-05, "loss": 4.8685, "step": 906050 }, { "epoch": 5.463634061335488, "grad_norm": 4.47836446762085, "learning_rate": 3.443814431002858e-05, "loss": 4.3967, "step": 906100 }, { "epoch": 5.463935553116822, "grad_norm": 2.719007968902588, "learning_rate": 3.442517197394907e-05, "loss": 4.2982, "step": 906150 }, { "epoch": 5.464237044898156, "grad_norm": 3.640695810317993, "learning_rate": 3.441220176485932e-05, "loss": 4.7339, "step": 906200 }, { "epoch": 5.46453853667949, "grad_norm": 4.002026081085205, "learning_rate": 3.43992336829981e-05, "loss": 4.4741, "step": 906250 }, { "epoch": 5.4648400284608245, "grad_norm": 2.0717551708221436, "learning_rate": 3.438626772860398e-05, "loss": 4.3913, "step": 906300 }, { "epoch": 5.465141520242158, "grad_norm": 2.6871328353881836, "learning_rate": 3.437330390191565e-05, "loss": 4.3834, "step": 906350 }, { "epoch": 5.465443012023492, "grad_norm": 6.5354509353637695, "learning_rate": 3.4360342203171646e-05, "loss": 4.6788, "step": 906400 }, { "epoch": 5.465744503804826, "grad_norm": 4.567416191101074, "learning_rate": 3.434738263261049e-05, "loss": 4.7067, "step": 906450 }, { "epoch": 5.4660459955861604, "grad_norm": 4.5625200271606445, "learning_rate": 3.433442519047069e-05, "loss": 4.5399, "step": 906500 }, { "epoch": 5.466347487367495, "grad_norm": 3.7190239429473877, "learning_rate": 3.432146987699079e-05, "loss": 4.3513, "step": 906550 }, { "epoch": 5.466648979148829, "grad_norm": 3.086164712905884, "learning_rate": 3.430851669240909e-05, "loss": 4.2092, "step": 906600 }, { "epoch": 5.466950470930162, "grad_norm": 4.996081829071045, "learning_rate": 3.429556563696407e-05, "loss": 4.7235, "step": 906650 }, { "epoch": 5.467251962711496, "grad_norm": 5.751585960388184, "learning_rate": 3.4282616710894065e-05, "loss": 4.4628, "step": 906700 }, { "epoch": 5.467553454492831, "grad_norm": 3.184148073196411, "learning_rate": 3.426966991443737e-05, "loss": 4.6998, "step": 906750 }, { "epoch": 5.467854946274165, "grad_norm": 2.9097769260406494, "learning_rate": 3.42567252478322e-05, "loss": 4.3001, "step": 906800 }, { "epoch": 5.468156438055499, "grad_norm": 1.859636664390564, "learning_rate": 3.424378271131687e-05, "loss": 4.9209, "step": 906850 }, { "epoch": 5.468457929836832, "grad_norm": 3.705509662628174, "learning_rate": 3.423084230512948e-05, "loss": 4.7998, "step": 906900 }, { "epoch": 5.4687594216181665, "grad_norm": 2.5260190963745117, "learning_rate": 3.4217904029508243e-05, "loss": 3.9772, "step": 906950 }, { "epoch": 5.469060913399501, "grad_norm": 2.053924560546875, "learning_rate": 3.420496788469128e-05, "loss": 4.4053, "step": 907000 }, { "epoch": 5.469362405180835, "grad_norm": 2.6778016090393066, "learning_rate": 3.41920338709166e-05, "loss": 4.5205, "step": 907050 }, { "epoch": 5.469663896962169, "grad_norm": 3.5174899101257324, "learning_rate": 3.417910198842231e-05, "loss": 4.0153, "step": 907100 }, { "epoch": 5.4699653887435025, "grad_norm": 3.3494718074798584, "learning_rate": 3.416617223744636e-05, "loss": 4.3398, "step": 907150 }, { "epoch": 5.470266880524837, "grad_norm": 3.4966063499450684, "learning_rate": 3.415324461822666e-05, "loss": 4.7245, "step": 907200 }, { "epoch": 5.470568372306171, "grad_norm": 3.412297487258911, "learning_rate": 3.4140319131001175e-05, "loss": 4.4746, "step": 907250 }, { "epoch": 5.470869864087505, "grad_norm": 3.9727447032928467, "learning_rate": 3.412739577600781e-05, "loss": 4.7081, "step": 907300 }, { "epoch": 5.471171355868839, "grad_norm": 6.28324031829834, "learning_rate": 3.411447455348434e-05, "loss": 4.6351, "step": 907350 }, { "epoch": 5.471472847650173, "grad_norm": 3.1412148475646973, "learning_rate": 3.4101555463668575e-05, "loss": 4.6873, "step": 907400 }, { "epoch": 5.471774339431507, "grad_norm": 3.5250697135925293, "learning_rate": 3.408863850679835e-05, "loss": 4.1861, "step": 907450 }, { "epoch": 5.472075831212841, "grad_norm": 4.996368408203125, "learning_rate": 3.407572368311125e-05, "loss": 4.4532, "step": 907500 }, { "epoch": 5.472377322994175, "grad_norm": 3.8036160469055176, "learning_rate": 3.406281099284507e-05, "loss": 4.7175, "step": 907550 }, { "epoch": 5.472678814775509, "grad_norm": 4.404510021209717, "learning_rate": 3.404990043623741e-05, "loss": 4.695, "step": 907600 }, { "epoch": 5.472980306556844, "grad_norm": 4.318177223205566, "learning_rate": 3.4036992013525826e-05, "loss": 4.5109, "step": 907650 }, { "epoch": 5.473281798338177, "grad_norm": 3.478793144226074, "learning_rate": 3.40240857249479e-05, "loss": 4.3795, "step": 907700 }, { "epoch": 5.473583290119511, "grad_norm": 1.51658296585083, "learning_rate": 3.4011181570741234e-05, "loss": 4.3662, "step": 907750 }, { "epoch": 5.473884781900845, "grad_norm": 2.9362058639526367, "learning_rate": 3.399827955114317e-05, "loss": 4.5816, "step": 907800 }, { "epoch": 5.47418627368218, "grad_norm": 3.9686038494110107, "learning_rate": 3.398537966639126e-05, "loss": 4.6664, "step": 907850 }, { "epoch": 5.474487765463514, "grad_norm": 2.7911860942840576, "learning_rate": 3.3972481916722896e-05, "loss": 4.2651, "step": 907900 }, { "epoch": 5.474789257244847, "grad_norm": 1.3844759464263916, "learning_rate": 3.395958630237543e-05, "loss": 4.5154, "step": 907950 }, { "epoch": 5.475090749026181, "grad_norm": 1.0357356071472168, "learning_rate": 3.394669282358615e-05, "loss": 4.2916, "step": 908000 }, { "epoch": 5.4753922408075155, "grad_norm": 5.078805446624756, "learning_rate": 3.3933801480592406e-05, "loss": 4.679, "step": 908050 }, { "epoch": 5.47569373258885, "grad_norm": 4.251710891723633, "learning_rate": 3.392091227363136e-05, "loss": 4.871, "step": 908100 }, { "epoch": 5.475995224370184, "grad_norm": 3.955176591873169, "learning_rate": 3.390802520294027e-05, "loss": 4.5439, "step": 908150 }, { "epoch": 5.476296716151518, "grad_norm": 4.37294864654541, "learning_rate": 3.3895140268756354e-05, "loss": 4.4368, "step": 908200 }, { "epoch": 5.4765982079328515, "grad_norm": 4.712566375732422, "learning_rate": 3.3882257471316645e-05, "loss": 4.3739, "step": 908250 }, { "epoch": 5.476899699714186, "grad_norm": 4.408198833465576, "learning_rate": 3.3869376810858324e-05, "loss": 4.4149, "step": 908300 }, { "epoch": 5.47720119149552, "grad_norm": 3.515787363052368, "learning_rate": 3.3856498287618336e-05, "loss": 4.6959, "step": 908350 }, { "epoch": 5.477502683276854, "grad_norm": 2.7229034900665283, "learning_rate": 3.384362190183382e-05, "loss": 4.3449, "step": 908400 }, { "epoch": 5.477804175058188, "grad_norm": 3.7759785652160645, "learning_rate": 3.383074765374162e-05, "loss": 4.4825, "step": 908450 }, { "epoch": 5.478105666839522, "grad_norm": 1.20309579372406, "learning_rate": 3.381787554357874e-05, "loss": 4.4579, "step": 908500 }, { "epoch": 5.478407158620856, "grad_norm": 0.8203766345977783, "learning_rate": 3.380500557158208e-05, "loss": 4.7635, "step": 908550 }, { "epoch": 5.47870865040219, "grad_norm": 3.832807779312134, "learning_rate": 3.379213773798845e-05, "loss": 4.4127, "step": 908600 }, { "epoch": 5.479010142183524, "grad_norm": 2.1663756370544434, "learning_rate": 3.377927204303468e-05, "loss": 4.4365, "step": 908650 }, { "epoch": 5.479311633964858, "grad_norm": 2.337214469909668, "learning_rate": 3.376640848695759e-05, "loss": 4.2569, "step": 908700 }, { "epoch": 5.479613125746192, "grad_norm": 3.8569390773773193, "learning_rate": 3.3753547069993886e-05, "loss": 4.7587, "step": 908750 }, { "epoch": 5.479914617527526, "grad_norm": 5.993983268737793, "learning_rate": 3.374068779238021e-05, "loss": 4.219, "step": 908800 }, { "epoch": 5.48021610930886, "grad_norm": 4.505828857421875, "learning_rate": 3.372783065435331e-05, "loss": 4.4314, "step": 908850 }, { "epoch": 5.480517601090194, "grad_norm": 4.1350274085998535, "learning_rate": 3.371497565614971e-05, "loss": 4.3467, "step": 908900 }, { "epoch": 5.480819092871529, "grad_norm": 3.68715500831604, "learning_rate": 3.370212279800604e-05, "loss": 4.5738, "step": 908950 }, { "epoch": 5.481120584652862, "grad_norm": 5.2594685554504395, "learning_rate": 3.368927208015888e-05, "loss": 4.444, "step": 909000 }, { "epoch": 5.481120584652862, "eval_loss": 4.950690269470215, "eval_runtime": 39.2968, "eval_samples_per_second": 13.029, "eval_steps_per_second": 6.515, "eval_tts_loss": 7.867395907202329, "step": 909000 }, { "epoch": 5.481422076434196, "grad_norm": 4.368719100952148, "learning_rate": 3.367642350284464e-05, "loss": 4.4746, "step": 909050 }, { "epoch": 5.48172356821553, "grad_norm": 1.5437792539596558, "learning_rate": 3.366357706629985e-05, "loss": 4.154, "step": 909100 }, { "epoch": 5.4820250599968645, "grad_norm": 3.357858180999756, "learning_rate": 3.3650732770760924e-05, "loss": 4.3997, "step": 909150 }, { "epoch": 5.482326551778199, "grad_norm": 3.622497797012329, "learning_rate": 3.3637890616464167e-05, "loss": 4.6499, "step": 909200 }, { "epoch": 5.482628043559533, "grad_norm": 4.181850910186768, "learning_rate": 3.362505060364598e-05, "loss": 4.9425, "step": 909250 }, { "epoch": 5.482929535340866, "grad_norm": 3.8818368911743164, "learning_rate": 3.3612212732542705e-05, "loss": 4.4953, "step": 909300 }, { "epoch": 5.4832310271222005, "grad_norm": 3.903632402420044, "learning_rate": 3.359937700339051e-05, "loss": 4.5628, "step": 909350 }, { "epoch": 5.483532518903535, "grad_norm": 3.1592037677764893, "learning_rate": 3.3586543416425675e-05, "loss": 4.6823, "step": 909400 }, { "epoch": 5.483834010684869, "grad_norm": 4.3107476234436035, "learning_rate": 3.3573711971884415e-05, "loss": 4.5848, "step": 909450 }, { "epoch": 5.484135502466203, "grad_norm": 3.667133331298828, "learning_rate": 3.3560882670002834e-05, "loss": 4.1454, "step": 909500 }, { "epoch": 5.484436994247536, "grad_norm": 4.323657512664795, "learning_rate": 3.3548055511017005e-05, "loss": 4.1859, "step": 909550 }, { "epoch": 5.484738486028871, "grad_norm": 3.324514865875244, "learning_rate": 3.353523049516307e-05, "loss": 4.5494, "step": 909600 }, { "epoch": 5.485039977810205, "grad_norm": 3.04355788230896, "learning_rate": 3.352240762267696e-05, "loss": 4.5725, "step": 909650 }, { "epoch": 5.485341469591539, "grad_norm": 6.28941011428833, "learning_rate": 3.3509586893794726e-05, "loss": 4.3176, "step": 909700 }, { "epoch": 5.485642961372873, "grad_norm": 6.879385948181152, "learning_rate": 3.3496768308752346e-05, "loss": 3.8202, "step": 909750 }, { "epoch": 5.485944453154207, "grad_norm": 5.031949996948242, "learning_rate": 3.3483951867785653e-05, "loss": 4.5312, "step": 909800 }, { "epoch": 5.486245944935541, "grad_norm": 3.456005334854126, "learning_rate": 3.347113757113054e-05, "loss": 4.4864, "step": 909850 }, { "epoch": 5.486547436716875, "grad_norm": 3.050764322280884, "learning_rate": 3.345832541902289e-05, "loss": 4.1694, "step": 909900 }, { "epoch": 5.486848928498209, "grad_norm": 2.9553749561309814, "learning_rate": 3.344551541169843e-05, "loss": 4.1868, "step": 909950 }, { "epoch": 5.487150420279543, "grad_norm": 3.995054006576538, "learning_rate": 3.34327075493929e-05, "loss": 4.5503, "step": 910000 }, { "epoch": 5.487451912060878, "grad_norm": 6.50631856918335, "learning_rate": 3.341990183234208e-05, "loss": 4.598, "step": 910050 }, { "epoch": 5.487753403842211, "grad_norm": 2.9798243045806885, "learning_rate": 3.340709826078156e-05, "loss": 4.3209, "step": 910100 }, { "epoch": 5.488054895623545, "grad_norm": 5.9319000244140625, "learning_rate": 3.3394296834946995e-05, "loss": 4.6055, "step": 910150 }, { "epoch": 5.488356387404879, "grad_norm": 3.930588722229004, "learning_rate": 3.338149755507403e-05, "loss": 4.4197, "step": 910200 }, { "epoch": 5.4886578791862135, "grad_norm": 0.5701434016227722, "learning_rate": 3.336870042139815e-05, "loss": 4.5447, "step": 910250 }, { "epoch": 5.488959370967548, "grad_norm": 3.789816379547119, "learning_rate": 3.335590543415494e-05, "loss": 4.6073, "step": 910300 }, { "epoch": 5.489260862748882, "grad_norm": 3.8859784603118896, "learning_rate": 3.3343112593579834e-05, "loss": 4.4937, "step": 910350 }, { "epoch": 5.489562354530215, "grad_norm": 3.9896790981292725, "learning_rate": 3.333032189990822e-05, "loss": 4.4412, "step": 910400 }, { "epoch": 5.4898638463115494, "grad_norm": 3.535137176513672, "learning_rate": 3.331753335337552e-05, "loss": 4.6792, "step": 910450 }, { "epoch": 5.490165338092884, "grad_norm": 4.807525157928467, "learning_rate": 3.3304746954217146e-05, "loss": 4.5172, "step": 910500 }, { "epoch": 5.490466829874218, "grad_norm": 6.132833480834961, "learning_rate": 3.329196270266835e-05, "loss": 4.8591, "step": 910550 }, { "epoch": 5.490768321655552, "grad_norm": 1.574721097946167, "learning_rate": 3.3279180598964404e-05, "loss": 4.404, "step": 910600 }, { "epoch": 5.491069813436885, "grad_norm": 4.531117916107178, "learning_rate": 3.326640064334062e-05, "loss": 4.4834, "step": 910650 }, { "epoch": 5.49137130521822, "grad_norm": 4.734782695770264, "learning_rate": 3.3253622836032154e-05, "loss": 4.5218, "step": 910700 }, { "epoch": 5.491672796999554, "grad_norm": 3.6301307678222656, "learning_rate": 3.324084717727411e-05, "loss": 4.3479, "step": 910750 }, { "epoch": 5.491974288780888, "grad_norm": 2.9844143390655518, "learning_rate": 3.322807366730168e-05, "loss": 4.8038, "step": 910800 }, { "epoch": 5.492275780562222, "grad_norm": 4.466570854187012, "learning_rate": 3.321530230634989e-05, "loss": 4.4309, "step": 910850 }, { "epoch": 5.4925772723435555, "grad_norm": 4.5126471519470215, "learning_rate": 3.320253309465379e-05, "loss": 4.5118, "step": 910900 }, { "epoch": 5.49287876412489, "grad_norm": 3.875229835510254, "learning_rate": 3.318976603244842e-05, "loss": 4.775, "step": 910950 }, { "epoch": 5.493180255906224, "grad_norm": 4.245510578155518, "learning_rate": 3.317700111996867e-05, "loss": 3.9844, "step": 911000 }, { "epoch": 5.493481747687558, "grad_norm": 3.615022897720337, "learning_rate": 3.316423835744951e-05, "loss": 4.4341, "step": 911050 }, { "epoch": 5.493783239468892, "grad_norm": 3.198887348175049, "learning_rate": 3.3151477745125844e-05, "loss": 4.6787, "step": 911100 }, { "epoch": 5.494084731250226, "grad_norm": 5.349935531616211, "learning_rate": 3.313871928323249e-05, "loss": 4.5654, "step": 911150 }, { "epoch": 5.49438622303156, "grad_norm": 3.584125280380249, "learning_rate": 3.3125962972004185e-05, "loss": 4.7028, "step": 911200 }, { "epoch": 5.494687714812894, "grad_norm": 3.005436420440674, "learning_rate": 3.311320881167578e-05, "loss": 4.7121, "step": 911250 }, { "epoch": 5.494989206594228, "grad_norm": 3.5474600791931152, "learning_rate": 3.310045680248192e-05, "loss": 4.4125, "step": 911300 }, { "epoch": 5.4952906983755625, "grad_norm": 3.2922606468200684, "learning_rate": 3.3087706944657345e-05, "loss": 4.5576, "step": 911350 }, { "epoch": 5.495592190156897, "grad_norm": 4.078487396240234, "learning_rate": 3.307495923843671e-05, "loss": 4.4304, "step": 911400 }, { "epoch": 5.49589368193823, "grad_norm": 4.342720031738281, "learning_rate": 3.306221368405455e-05, "loss": 4.518, "step": 911450 }, { "epoch": 5.496195173719564, "grad_norm": 3.7842700481414795, "learning_rate": 3.304947028174549e-05, "loss": 4.5876, "step": 911500 }, { "epoch": 5.496496665500898, "grad_norm": 3.434812068939209, "learning_rate": 3.303672903174407e-05, "loss": 4.8146, "step": 911550 }, { "epoch": 5.496798157282233, "grad_norm": 3.969470739364624, "learning_rate": 3.302398993428467e-05, "loss": 4.5587, "step": 911600 }, { "epoch": 5.497099649063567, "grad_norm": 3.3324451446533203, "learning_rate": 3.30112529896018e-05, "loss": 4.3418, "step": 911650 }, { "epoch": 5.4974011408449, "grad_norm": 2.1951842308044434, "learning_rate": 3.299851819792988e-05, "loss": 4.0982, "step": 911700 }, { "epoch": 5.497702632626234, "grad_norm": 4.268782138824463, "learning_rate": 3.2985785559503294e-05, "loss": 4.7531, "step": 911750 }, { "epoch": 5.498004124407569, "grad_norm": 3.7952888011932373, "learning_rate": 3.2973055074556295e-05, "loss": 4.4283, "step": 911800 }, { "epoch": 5.498305616188903, "grad_norm": 4.457744598388672, "learning_rate": 3.2960326743323265e-05, "loss": 4.6792, "step": 911850 }, { "epoch": 5.498607107970237, "grad_norm": 3.4341330528259277, "learning_rate": 3.294760056603835e-05, "loss": 4.193, "step": 911900 }, { "epoch": 5.498908599751571, "grad_norm": 5.786385536193848, "learning_rate": 3.293487654293583e-05, "loss": 4.4153, "step": 911950 }, { "epoch": 5.4992100915329045, "grad_norm": 3.1228647232055664, "learning_rate": 3.292215467424982e-05, "loss": 4.0277, "step": 912000 }, { "epoch": 5.4992100915329045, "eval_loss": 4.9485859870910645, "eval_runtime": 39.2376, "eval_samples_per_second": 13.049, "eval_steps_per_second": 6.524, "eval_tts_loss": 7.90543537814799, "step": 912000 }, { "epoch": 5.499511583314239, "grad_norm": 7.240533828735352, "learning_rate": 3.2909434960214506e-05, "loss": 4.488, "step": 912050 }, { "epoch": 5.499813075095573, "grad_norm": 5.483821392059326, "learning_rate": 3.289671740106391e-05, "loss": 4.4086, "step": 912100 }, { "epoch": 5.500114566876907, "grad_norm": 5.08900785446167, "learning_rate": 3.28840019970321e-05, "loss": 4.5375, "step": 912150 }, { "epoch": 5.500416058658241, "grad_norm": 3.4121553897857666, "learning_rate": 3.2871288748353174e-05, "loss": 4.7442, "step": 912200 }, { "epoch": 5.500717550439575, "grad_norm": 3.128535270690918, "learning_rate": 3.2858577655260955e-05, "loss": 4.7744, "step": 912250 }, { "epoch": 5.501019042220909, "grad_norm": 3.5813493728637695, "learning_rate": 3.2845868717989506e-05, "loss": 4.772, "step": 912300 }, { "epoch": 5.501320534002243, "grad_norm": 4.431572437286377, "learning_rate": 3.283316193677264e-05, "loss": 4.0719, "step": 912350 }, { "epoch": 5.501622025783577, "grad_norm": 4.616466522216797, "learning_rate": 3.282045731184419e-05, "loss": 4.4294, "step": 912400 }, { "epoch": 5.5019235175649115, "grad_norm": 3.122300148010254, "learning_rate": 3.2807754843438e-05, "loss": 4.5869, "step": 912450 }, { "epoch": 5.502225009346246, "grad_norm": 1.4532192945480347, "learning_rate": 3.2795054531787865e-05, "loss": 4.3593, "step": 912500 }, { "epoch": 5.502526501127579, "grad_norm": 4.135629653930664, "learning_rate": 3.2782356377127474e-05, "loss": 4.5363, "step": 912550 }, { "epoch": 5.502827992908913, "grad_norm": 1.595167636871338, "learning_rate": 3.2769660379690524e-05, "loss": 4.3803, "step": 912600 }, { "epoch": 5.503129484690247, "grad_norm": 3.37155818939209, "learning_rate": 3.275696653971071e-05, "loss": 4.5031, "step": 912650 }, { "epoch": 5.503430976471582, "grad_norm": 3.831838846206665, "learning_rate": 3.274427485742162e-05, "loss": 4.7268, "step": 912700 }, { "epoch": 5.503732468252915, "grad_norm": 5.148808479309082, "learning_rate": 3.2731585333056786e-05, "loss": 4.5522, "step": 912750 }, { "epoch": 5.504033960034249, "grad_norm": 3.078965902328491, "learning_rate": 3.2718897966849804e-05, "loss": 4.7049, "step": 912800 }, { "epoch": 5.504335451815583, "grad_norm": 1.2894729375839233, "learning_rate": 3.270621275903409e-05, "loss": 4.6004, "step": 912850 }, { "epoch": 5.504636943596918, "grad_norm": 3.159869909286499, "learning_rate": 3.2693529709843155e-05, "loss": 4.605, "step": 912900 }, { "epoch": 5.504938435378252, "grad_norm": 3.068582534790039, "learning_rate": 3.268084881951044e-05, "loss": 4.6401, "step": 912950 }, { "epoch": 5.505239927159586, "grad_norm": 2.738762378692627, "learning_rate": 3.2668170088269234e-05, "loss": 4.5317, "step": 913000 }, { "epoch": 5.505541418940919, "grad_norm": 4.054309844970703, "learning_rate": 3.265549351635292e-05, "loss": 3.9947, "step": 913050 }, { "epoch": 5.5058429107222535, "grad_norm": 3.748835325241089, "learning_rate": 3.2642819103994826e-05, "loss": 4.4272, "step": 913100 }, { "epoch": 5.506144402503588, "grad_norm": 2.055398464202881, "learning_rate": 3.263014685142817e-05, "loss": 4.107, "step": 913150 }, { "epoch": 5.506445894284922, "grad_norm": 5.821165561676025, "learning_rate": 3.2617476758886116e-05, "loss": 4.4197, "step": 913200 }, { "epoch": 5.506747386066256, "grad_norm": 3.0750465393066406, "learning_rate": 3.260480882660194e-05, "loss": 4.4509, "step": 913250 }, { "epoch": 5.5070488778475895, "grad_norm": 7.559163570404053, "learning_rate": 3.259214305480868e-05, "loss": 4.5654, "step": 913300 }, { "epoch": 5.507350369628924, "grad_norm": 5.1052751541137695, "learning_rate": 3.257947944373948e-05, "loss": 4.4997, "step": 913350 }, { "epoch": 5.507651861410258, "grad_norm": 4.711912155151367, "learning_rate": 3.256681799362743e-05, "loss": 4.741, "step": 913400 }, { "epoch": 5.507953353191592, "grad_norm": 3.626720905303955, "learning_rate": 3.2554158704705465e-05, "loss": 4.4127, "step": 913450 }, { "epoch": 5.508254844972926, "grad_norm": 1.582030177116394, "learning_rate": 3.2541501577206654e-05, "loss": 4.2586, "step": 913500 }, { "epoch": 5.5085563367542605, "grad_norm": 1.1548153162002563, "learning_rate": 3.252884661136388e-05, "loss": 4.4928, "step": 913550 }, { "epoch": 5.508857828535594, "grad_norm": 2.325118064880371, "learning_rate": 3.251619380740999e-05, "loss": 4.6166, "step": 913600 }, { "epoch": 5.509159320316928, "grad_norm": 3.4578351974487305, "learning_rate": 3.2503543165577915e-05, "loss": 4.0499, "step": 913650 }, { "epoch": 5.509460812098262, "grad_norm": 4.583946228027344, "learning_rate": 3.249089468610049e-05, "loss": 4.6561, "step": 913700 }, { "epoch": 5.509762303879596, "grad_norm": 3.587611198425293, "learning_rate": 3.24782483692104e-05, "loss": 4.6883, "step": 913750 }, { "epoch": 5.510063795660931, "grad_norm": 6.131679534912109, "learning_rate": 3.246560421514044e-05, "loss": 4.6529, "step": 913800 }, { "epoch": 5.510365287442264, "grad_norm": 5.528194427490234, "learning_rate": 3.2452962224123365e-05, "loss": 4.3617, "step": 913850 }, { "epoch": 5.510666779223598, "grad_norm": 4.091848850250244, "learning_rate": 3.2440322396391757e-05, "loss": 4.5474, "step": 913900 }, { "epoch": 5.510968271004932, "grad_norm": 3.4564332962036133, "learning_rate": 3.242768473217819e-05, "loss": 4.5128, "step": 913950 }, { "epoch": 5.511269762786267, "grad_norm": 3.3983402252197266, "learning_rate": 3.2415049231715364e-05, "loss": 4.6893, "step": 914000 }, { "epoch": 5.511571254567601, "grad_norm": 4.9464430809021, "learning_rate": 3.24024158952357e-05, "loss": 4.6557, "step": 914050 }, { "epoch": 5.511872746348935, "grad_norm": 3.714097499847412, "learning_rate": 3.2389784722971754e-05, "loss": 4.5887, "step": 914100 }, { "epoch": 5.512174238130268, "grad_norm": 3.0747733116149902, "learning_rate": 3.2377155715156046e-05, "loss": 4.608, "step": 914150 }, { "epoch": 5.5124757299116025, "grad_norm": 7.24098014831543, "learning_rate": 3.236452887202087e-05, "loss": 4.7027, "step": 914200 }, { "epoch": 5.512777221692937, "grad_norm": 3.7100846767425537, "learning_rate": 3.2351904193798653e-05, "loss": 4.5884, "step": 914250 }, { "epoch": 5.513078713474271, "grad_norm": 2.7589375972747803, "learning_rate": 3.233928168072181e-05, "loss": 4.3482, "step": 914300 }, { "epoch": 5.513380205255604, "grad_norm": 3.8578765392303467, "learning_rate": 3.232666133302255e-05, "loss": 4.7476, "step": 914350 }, { "epoch": 5.5136816970369384, "grad_norm": 4.210733890533447, "learning_rate": 3.231404315093314e-05, "loss": 4.9328, "step": 914400 }, { "epoch": 5.513983188818273, "grad_norm": 4.256622791290283, "learning_rate": 3.2301427134685865e-05, "loss": 4.4833, "step": 914450 }, { "epoch": 5.514284680599607, "grad_norm": 4.493930816650391, "learning_rate": 3.2288813284512794e-05, "loss": 4.3709, "step": 914500 }, { "epoch": 5.514586172380941, "grad_norm": 4.209929943084717, "learning_rate": 3.227620160064616e-05, "loss": 4.136, "step": 914550 }, { "epoch": 5.514887664162275, "grad_norm": 1.6144444942474365, "learning_rate": 3.226359208331805e-05, "loss": 4.3603, "step": 914600 }, { "epoch": 5.515189155943609, "grad_norm": 2.1306934356689453, "learning_rate": 3.2250984732760474e-05, "loss": 4.2387, "step": 914650 }, { "epoch": 5.515490647724943, "grad_norm": 3.6660315990448, "learning_rate": 3.2238379549205526e-05, "loss": 4.2755, "step": 914700 }, { "epoch": 5.515792139506277, "grad_norm": 3.7527191638946533, "learning_rate": 3.222577653288516e-05, "loss": 4.8925, "step": 914750 }, { "epoch": 5.516093631287611, "grad_norm": 1.9467811584472656, "learning_rate": 3.221317568403126e-05, "loss": 4.3538, "step": 914800 }, { "epoch": 5.516395123068945, "grad_norm": 3.7181806564331055, "learning_rate": 3.220057700287576e-05, "loss": 4.2759, "step": 914850 }, { "epoch": 5.516696614850279, "grad_norm": 5.964927673339844, "learning_rate": 3.2187980489650586e-05, "loss": 4.438, "step": 914900 }, { "epoch": 5.516998106631613, "grad_norm": 3.3966894149780273, "learning_rate": 3.217538614458746e-05, "loss": 3.8748, "step": 914950 }, { "epoch": 5.517299598412947, "grad_norm": 3.5164449214935303, "learning_rate": 3.216279396791821e-05, "loss": 4.5176, "step": 915000 }, { "epoch": 5.517299598412947, "eval_loss": 4.946063995361328, "eval_runtime": 39.0313, "eval_samples_per_second": 13.118, "eval_steps_per_second": 6.559, "eval_tts_loss": 7.844386587936475, "step": 915000 }, { "epoch": 5.517601090194281, "grad_norm": 1.3173208236694336, "learning_rate": 3.215020395987461e-05, "loss": 4.5963, "step": 915050 }, { "epoch": 5.517902581975616, "grad_norm": 3.8623106479644775, "learning_rate": 3.213761612068832e-05, "loss": 4.2814, "step": 915100 }, { "epoch": 5.51820407375695, "grad_norm": 3.0471699237823486, "learning_rate": 3.212503045059097e-05, "loss": 4.5343, "step": 915150 }, { "epoch": 5.518505565538283, "grad_norm": 4.269237518310547, "learning_rate": 3.2112446949814216e-05, "loss": 4.6988, "step": 915200 }, { "epoch": 5.518807057319617, "grad_norm": 4.859570503234863, "learning_rate": 3.209986561858969e-05, "loss": 4.5575, "step": 915250 }, { "epoch": 5.5191085491009515, "grad_norm": 2.7808234691619873, "learning_rate": 3.208728645714884e-05, "loss": 4.1297, "step": 915300 }, { "epoch": 5.519410040882286, "grad_norm": 3.4960527420043945, "learning_rate": 3.207470946572321e-05, "loss": 4.8944, "step": 915350 }, { "epoch": 5.51971153266362, "grad_norm": 4.4061970710754395, "learning_rate": 3.2062134644544306e-05, "loss": 4.6329, "step": 915400 }, { "epoch": 5.520013024444953, "grad_norm": 1.3293708562850952, "learning_rate": 3.204956199384347e-05, "loss": 4.1351, "step": 915450 }, { "epoch": 5.520314516226287, "grad_norm": 4.345183849334717, "learning_rate": 3.203699151385216e-05, "loss": 4.2179, "step": 915500 }, { "epoch": 5.520616008007622, "grad_norm": 3.6962451934814453, "learning_rate": 3.2024423204801664e-05, "loss": 4.4978, "step": 915550 }, { "epoch": 5.520917499788956, "grad_norm": 4.168375015258789, "learning_rate": 3.2011857066923274e-05, "loss": 4.2062, "step": 915600 }, { "epoch": 5.52121899157029, "grad_norm": 3.2502987384796143, "learning_rate": 3.199929310044827e-05, "loss": 4.5113, "step": 915650 }, { "epoch": 5.521520483351624, "grad_norm": 3.3373382091522217, "learning_rate": 3.198673130560793e-05, "loss": 4.5284, "step": 915700 }, { "epoch": 5.521821975132958, "grad_norm": 3.0794527530670166, "learning_rate": 3.197417168263334e-05, "loss": 3.9947, "step": 915750 }, { "epoch": 5.522123466914292, "grad_norm": 3.0116755962371826, "learning_rate": 3.196161423175569e-05, "loss": 4.7964, "step": 915800 }, { "epoch": 5.522424958695626, "grad_norm": 3.724486827850342, "learning_rate": 3.194905895320612e-05, "loss": 4.5922, "step": 915850 }, { "epoch": 5.52272645047696, "grad_norm": 3.383713960647583, "learning_rate": 3.1936505847215664e-05, "loss": 4.4864, "step": 915900 }, { "epoch": 5.523027942258294, "grad_norm": 3.661628007888794, "learning_rate": 3.192395491401528e-05, "loss": 3.9911, "step": 915950 }, { "epoch": 5.523329434039628, "grad_norm": 4.740444660186768, "learning_rate": 3.1911406153836034e-05, "loss": 4.1034, "step": 916000 }, { "epoch": 5.523630925820962, "grad_norm": 3.3825109004974365, "learning_rate": 3.18988595669088e-05, "loss": 4.8788, "step": 916050 }, { "epoch": 5.523932417602296, "grad_norm": 1.090095043182373, "learning_rate": 3.188631515346453e-05, "loss": 4.3702, "step": 916100 }, { "epoch": 5.52423390938363, "grad_norm": 5.392988681793213, "learning_rate": 3.1873772913734105e-05, "loss": 4.8162, "step": 916150 }, { "epoch": 5.5245354011649646, "grad_norm": 3.7915539741516113, "learning_rate": 3.1861232847948265e-05, "loss": 4.601, "step": 916200 }, { "epoch": 5.524836892946299, "grad_norm": 7.670810222625732, "learning_rate": 3.18486949563379e-05, "loss": 4.6224, "step": 916250 }, { "epoch": 5.525138384727632, "grad_norm": 2.33453106880188, "learning_rate": 3.183615923913368e-05, "loss": 4.1166, "step": 916300 }, { "epoch": 5.525439876508966, "grad_norm": 4.488497734069824, "learning_rate": 3.1823625696566285e-05, "loss": 4.4207, "step": 916350 }, { "epoch": 5.5257413682903005, "grad_norm": 2.125336170196533, "learning_rate": 3.18110943288664e-05, "loss": 4.5056, "step": 916400 }, { "epoch": 5.526042860071635, "grad_norm": 3.8015668392181396, "learning_rate": 3.179856513626471e-05, "loss": 4.3307, "step": 916450 }, { "epoch": 5.526344351852968, "grad_norm": 3.696709632873535, "learning_rate": 3.178603811899171e-05, "loss": 4.3214, "step": 916500 }, { "epoch": 5.526645843634302, "grad_norm": 3.573441505432129, "learning_rate": 3.1773513277277976e-05, "loss": 4.6225, "step": 916550 }, { "epoch": 5.526947335415636, "grad_norm": 4.76745080947876, "learning_rate": 3.176099061135404e-05, "loss": 4.5343, "step": 916600 }, { "epoch": 5.527248827196971, "grad_norm": 3.1779422760009766, "learning_rate": 3.17484701214503e-05, "loss": 4.5111, "step": 916650 }, { "epoch": 5.527550318978305, "grad_norm": 3.435119867324829, "learning_rate": 3.173595180779726e-05, "loss": 4.5443, "step": 916700 }, { "epoch": 5.527851810759639, "grad_norm": 4.797258377075195, "learning_rate": 3.172343567062526e-05, "loss": 4.3962, "step": 916750 }, { "epoch": 5.528153302540972, "grad_norm": 3.3724889755249023, "learning_rate": 3.171092171016459e-05, "loss": 4.5849, "step": 916800 }, { "epoch": 5.528454794322307, "grad_norm": 1.0267595052719116, "learning_rate": 3.16984099266456e-05, "loss": 4.3931, "step": 916850 }, { "epoch": 5.528756286103641, "grad_norm": 3.9635801315307617, "learning_rate": 3.1685900320298585e-05, "loss": 4.2321, "step": 916900 }, { "epoch": 5.529057777884975, "grad_norm": 1.6597272157669067, "learning_rate": 3.1673392891353694e-05, "loss": 4.6138, "step": 916950 }, { "epoch": 5.529359269666309, "grad_norm": 3.33617901802063, "learning_rate": 3.166088764004116e-05, "loss": 4.4229, "step": 917000 }, { "epoch": 5.5296607614476425, "grad_norm": 1.9689534902572632, "learning_rate": 3.1648384566591136e-05, "loss": 4.6986, "step": 917050 }, { "epoch": 5.529962253228977, "grad_norm": 6.268651485443115, "learning_rate": 3.16358836712337e-05, "loss": 4.5949, "step": 917100 }, { "epoch": 5.530263745010311, "grad_norm": 2.8044111728668213, "learning_rate": 3.162338495419887e-05, "loss": 4.4209, "step": 917150 }, { "epoch": 5.530565236791645, "grad_norm": 3.9364163875579834, "learning_rate": 3.1610888415716736e-05, "loss": 4.1902, "step": 917200 }, { "epoch": 5.530866728572979, "grad_norm": 1.1836793422698975, "learning_rate": 3.159839405601722e-05, "loss": 4.1356, "step": 917250 }, { "epoch": 5.5311682203543135, "grad_norm": 4.252340316772461, "learning_rate": 3.15859018753303e-05, "loss": 4.5119, "step": 917300 }, { "epoch": 5.531469712135647, "grad_norm": 3.115145206451416, "learning_rate": 3.15734118738859e-05, "loss": 3.9321, "step": 917350 }, { "epoch": 5.531771203916981, "grad_norm": 4.095522403717041, "learning_rate": 3.1560924051913806e-05, "loss": 4.8145, "step": 917400 }, { "epoch": 5.532072695698315, "grad_norm": 3.4146647453308105, "learning_rate": 3.1548438409643926e-05, "loss": 4.2563, "step": 917450 }, { "epoch": 5.5323741874796495, "grad_norm": 3.3356449604034424, "learning_rate": 3.1535954947305966e-05, "loss": 4.5851, "step": 917500 }, { "epoch": 5.532675679260984, "grad_norm": 4.209408760070801, "learning_rate": 3.1523473665129745e-05, "loss": 4.4566, "step": 917550 }, { "epoch": 5.532977171042317, "grad_norm": 5.931037902832031, "learning_rate": 3.1510994563344856e-05, "loss": 4.3412, "step": 917600 }, { "epoch": 5.533278662823651, "grad_norm": 6.368936061859131, "learning_rate": 3.149851764218108e-05, "loss": 4.3325, "step": 917650 }, { "epoch": 5.533580154604985, "grad_norm": 5.6848673820495605, "learning_rate": 3.148604290186793e-05, "loss": 4.3266, "step": 917700 }, { "epoch": 5.53388164638632, "grad_norm": 2.69252610206604, "learning_rate": 3.147357034263502e-05, "loss": 4.601, "step": 917750 }, { "epoch": 5.534183138167654, "grad_norm": 5.477636337280273, "learning_rate": 3.1461099964711946e-05, "loss": 4.5705, "step": 917800 }, { "epoch": 5.534484629948988, "grad_norm": 3.151930093765259, "learning_rate": 3.144863176832812e-05, "loss": 4.5052, "step": 917850 }, { "epoch": 5.534786121730321, "grad_norm": 3.0683670043945312, "learning_rate": 3.143616575371309e-05, "loss": 4.4384, "step": 917900 }, { "epoch": 5.535087613511656, "grad_norm": 3.7791357040405273, "learning_rate": 3.142370192109623e-05, "loss": 4.0865, "step": 917950 }, { "epoch": 5.53538910529299, "grad_norm": 3.623176336288452, "learning_rate": 3.141124027070689e-05, "loss": 4.7455, "step": 918000 }, { "epoch": 5.53538910529299, "eval_loss": 4.935760498046875, "eval_runtime": 39.1513, "eval_samples_per_second": 13.077, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.88633884294472, "step": 918000 }, { "epoch": 5.535690597074324, "grad_norm": 3.9062724113464355, "learning_rate": 3.1398780802774426e-05, "loss": 4.5223, "step": 918050 }, { "epoch": 5.535992088855657, "grad_norm": 3.6822805404663086, "learning_rate": 3.138632351752818e-05, "loss": 4.5226, "step": 918100 }, { "epoch": 5.5362935806369915, "grad_norm": 3.011244297027588, "learning_rate": 3.137386841519734e-05, "loss": 4.3455, "step": 918150 }, { "epoch": 5.536595072418326, "grad_norm": 3.779353380203247, "learning_rate": 3.136141549601118e-05, "loss": 4.9456, "step": 918200 }, { "epoch": 5.53689656419966, "grad_norm": 4.9241108894348145, "learning_rate": 3.134896476019889e-05, "loss": 4.4368, "step": 918250 }, { "epoch": 5.537198055980994, "grad_norm": 2.7228004932403564, "learning_rate": 3.133651620798959e-05, "loss": 4.6363, "step": 918300 }, { "epoch": 5.537499547762328, "grad_norm": 2.4625179767608643, "learning_rate": 3.132406983961232e-05, "loss": 4.4366, "step": 918350 }, { "epoch": 5.537801039543662, "grad_norm": 8.787477493286133, "learning_rate": 3.131162565529617e-05, "loss": 4.1806, "step": 918400 }, { "epoch": 5.538102531324996, "grad_norm": 4.539738655090332, "learning_rate": 3.129918365527023e-05, "loss": 4.6189, "step": 918450 }, { "epoch": 5.53840402310633, "grad_norm": 3.357917547225952, "learning_rate": 3.1286743839763376e-05, "loss": 4.6577, "step": 918500 }, { "epoch": 5.538705514887664, "grad_norm": 4.1478753089904785, "learning_rate": 3.1274306209004576e-05, "loss": 4.2295, "step": 918550 }, { "epoch": 5.5390070066689985, "grad_norm": 4.514820575714111, "learning_rate": 3.12618707632228e-05, "loss": 4.0465, "step": 918600 }, { "epoch": 5.539308498450332, "grad_norm": 5.2306108474731445, "learning_rate": 3.1249437502646794e-05, "loss": 4.3739, "step": 918650 }, { "epoch": 5.539609990231666, "grad_norm": 4.191529750823975, "learning_rate": 3.123700642750548e-05, "loss": 4.605, "step": 918700 }, { "epoch": 5.539911482013, "grad_norm": 3.3876214027404785, "learning_rate": 3.122457753802755e-05, "loss": 4.2578, "step": 918750 }, { "epoch": 5.540212973794334, "grad_norm": 4.907551288604736, "learning_rate": 3.121215083444176e-05, "loss": 4.5674, "step": 918800 }, { "epoch": 5.540514465575669, "grad_norm": 2.865424394607544, "learning_rate": 3.1199726316976795e-05, "loss": 4.5836, "step": 918850 }, { "epoch": 5.540815957357003, "grad_norm": 3.453199625015259, "learning_rate": 3.118730398586137e-05, "loss": 4.0388, "step": 918900 }, { "epoch": 5.541117449138336, "grad_norm": 1.3889286518096924, "learning_rate": 3.117488384132402e-05, "loss": 4.5704, "step": 918950 }, { "epoch": 5.54141894091967, "grad_norm": 4.955084800720215, "learning_rate": 3.1162465883593365e-05, "loss": 4.5389, "step": 919000 }, { "epoch": 5.541720432701005, "grad_norm": 2.572057008743286, "learning_rate": 3.115005011289798e-05, "loss": 4.6559, "step": 919050 }, { "epoch": 5.542021924482339, "grad_norm": 2.5393548011779785, "learning_rate": 3.113763652946631e-05, "loss": 4.2118, "step": 919100 }, { "epoch": 5.542323416263673, "grad_norm": 3.6440610885620117, "learning_rate": 3.112522513352677e-05, "loss": 4.8806, "step": 919150 }, { "epoch": 5.542624908045006, "grad_norm": 2.990229606628418, "learning_rate": 3.1112815925307864e-05, "loss": 4.4341, "step": 919200 }, { "epoch": 5.5429263998263405, "grad_norm": 4.547486782073975, "learning_rate": 3.1100408905037865e-05, "loss": 4.4349, "step": 919250 }, { "epoch": 5.543227891607675, "grad_norm": 4.10889196395874, "learning_rate": 3.1088004072945185e-05, "loss": 4.6042, "step": 919300 }, { "epoch": 5.543529383389009, "grad_norm": 3.319211721420288, "learning_rate": 3.107560142925812e-05, "loss": 4.5554, "step": 919350 }, { "epoch": 5.543830875170343, "grad_norm": 4.808568000793457, "learning_rate": 3.106320097420485e-05, "loss": 4.4849, "step": 919400 }, { "epoch": 5.544132366951677, "grad_norm": 4.066752910614014, "learning_rate": 3.105080270801368e-05, "loss": 4.4362, "step": 919450 }, { "epoch": 5.544433858733011, "grad_norm": 4.46519660949707, "learning_rate": 3.1038406630912756e-05, "loss": 4.2822, "step": 919500 }, { "epoch": 5.544735350514345, "grad_norm": 4.665613651275635, "learning_rate": 3.1026012743130116e-05, "loss": 3.8861, "step": 919550 }, { "epoch": 5.545036842295679, "grad_norm": 3.482351541519165, "learning_rate": 3.101362104489396e-05, "loss": 4.618, "step": 919600 }, { "epoch": 5.545338334077013, "grad_norm": 3.6751673221588135, "learning_rate": 3.100123153643233e-05, "loss": 4.8454, "step": 919650 }, { "epoch": 5.5456398258583475, "grad_norm": 4.294198036193848, "learning_rate": 3.0988844217973166e-05, "loss": 4.6702, "step": 919700 }, { "epoch": 5.545941317639681, "grad_norm": 1.6000328063964844, "learning_rate": 3.0976459089744494e-05, "loss": 3.9301, "step": 919750 }, { "epoch": 5.546242809421015, "grad_norm": 5.006657123565674, "learning_rate": 3.096407615197427e-05, "loss": 4.524, "step": 919800 }, { "epoch": 5.546544301202349, "grad_norm": 4.966272830963135, "learning_rate": 3.095169540489033e-05, "loss": 4.5064, "step": 919850 }, { "epoch": 5.546845792983683, "grad_norm": 8.504027366638184, "learning_rate": 3.0939316848720577e-05, "loss": 4.4987, "step": 919900 }, { "epoch": 5.547147284765018, "grad_norm": 3.4988250732421875, "learning_rate": 3.092694048369277e-05, "loss": 4.1272, "step": 919950 }, { "epoch": 5.547448776546352, "grad_norm": 4.711808204650879, "learning_rate": 3.091456631003468e-05, "loss": 4.3928, "step": 920000 }, { "epoch": 5.547750268327685, "grad_norm": 4.806146621704102, "learning_rate": 3.090219432797403e-05, "loss": 4.7375, "step": 920050 }, { "epoch": 5.548051760109019, "grad_norm": 2.962117910385132, "learning_rate": 3.088982453773859e-05, "loss": 4.4121, "step": 920100 }, { "epoch": 5.5483532518903536, "grad_norm": 3.060317277908325, "learning_rate": 3.087745693955592e-05, "loss": 4.5287, "step": 920150 }, { "epoch": 5.548654743671688, "grad_norm": 4.364900588989258, "learning_rate": 3.086509153365363e-05, "loss": 4.2867, "step": 920200 }, { "epoch": 5.548956235453021, "grad_norm": 3.005204200744629, "learning_rate": 3.0852728320259376e-05, "loss": 4.5062, "step": 920250 }, { "epoch": 5.549257727234355, "grad_norm": 4.194184303283691, "learning_rate": 3.0840367299600615e-05, "loss": 4.6115, "step": 920300 }, { "epoch": 5.5495592190156895, "grad_norm": 3.906924247741699, "learning_rate": 3.0828008471904815e-05, "loss": 4.5438, "step": 920350 }, { "epoch": 5.549860710797024, "grad_norm": 3.5081825256347656, "learning_rate": 3.0815651837399475e-05, "loss": 4.6069, "step": 920400 }, { "epoch": 5.550162202578358, "grad_norm": 2.8914074897766113, "learning_rate": 3.0803297396311946e-05, "loss": 4.4, "step": 920450 }, { "epoch": 5.550463694359692, "grad_norm": 3.5693273544311523, "learning_rate": 3.0790945148869625e-05, "loss": 4.6819, "step": 920500 }, { "epoch": 5.550765186141025, "grad_norm": 3.0133142471313477, "learning_rate": 3.077859509529988e-05, "loss": 4.5776, "step": 920550 }, { "epoch": 5.55106667792236, "grad_norm": 4.177707672119141, "learning_rate": 3.076624723582992e-05, "loss": 4.4496, "step": 920600 }, { "epoch": 5.551368169703694, "grad_norm": 2.1785337924957275, "learning_rate": 3.075390157068705e-05, "loss": 4.4588, "step": 920650 }, { "epoch": 5.551669661485028, "grad_norm": 3.4133105278015137, "learning_rate": 3.0741558100098467e-05, "loss": 4.7093, "step": 920700 }, { "epoch": 5.551971153266362, "grad_norm": 3.775916337966919, "learning_rate": 3.072921682429126e-05, "loss": 4.3715, "step": 920750 }, { "epoch": 5.552272645047696, "grad_norm": 4.1272735595703125, "learning_rate": 3.071687774349261e-05, "loss": 4.4953, "step": 920800 }, { "epoch": 5.55257413682903, "grad_norm": 4.088614463806152, "learning_rate": 3.0704540857929647e-05, "loss": 4.37, "step": 920850 }, { "epoch": 5.552875628610364, "grad_norm": 2.9423258304595947, "learning_rate": 3.069220616782932e-05, "loss": 4.0808, "step": 920900 }, { "epoch": 5.553177120391698, "grad_norm": 0.6944441199302673, "learning_rate": 3.067987367341867e-05, "loss": 4.3561, "step": 920950 }, { "epoch": 5.553478612173032, "grad_norm": 4.4562907218933105, "learning_rate": 3.0667543374924716e-05, "loss": 4.1899, "step": 921000 }, { "epoch": 5.553478612173032, "eval_loss": 4.9335479736328125, "eval_runtime": 39.1788, "eval_samples_per_second": 13.068, "eval_steps_per_second": 6.534, "eval_tts_loss": 7.861535921662461, "step": 921000 }, { "epoch": 5.553780103954367, "grad_norm": 5.808689117431641, "learning_rate": 3.06552152725743e-05, "loss": 4.6746, "step": 921050 }, { "epoch": 5.5540815957357, "grad_norm": 2.9738287925720215, "learning_rate": 3.0642889366594356e-05, "loss": 4.3617, "step": 921100 }, { "epoch": 5.554383087517034, "grad_norm": 3.941957950592041, "learning_rate": 3.063056565721171e-05, "loss": 4.0628, "step": 921150 }, { "epoch": 5.554684579298368, "grad_norm": 3.808372735977173, "learning_rate": 3.061824414465311e-05, "loss": 4.2826, "step": 921200 }, { "epoch": 5.5549860710797025, "grad_norm": 3.03708553314209, "learning_rate": 3.0605924829145356e-05, "loss": 4.5635, "step": 921250 }, { "epoch": 5.555287562861037, "grad_norm": 6.004734039306641, "learning_rate": 3.059360771091522e-05, "loss": 4.3111, "step": 921300 }, { "epoch": 5.55558905464237, "grad_norm": 3.715740919113159, "learning_rate": 3.0581292790189274e-05, "loss": 4.3423, "step": 921350 }, { "epoch": 5.555890546423704, "grad_norm": 3.129056453704834, "learning_rate": 3.056898006719423e-05, "loss": 4.5243, "step": 921400 }, { "epoch": 5.5561920382050385, "grad_norm": 2.2197742462158203, "learning_rate": 3.055666954215672e-05, "loss": 4.7153, "step": 921450 }, { "epoch": 5.556493529986373, "grad_norm": 3.55482816696167, "learning_rate": 3.0544361215303244e-05, "loss": 4.2246, "step": 921500 }, { "epoch": 5.556795021767707, "grad_norm": 3.537465810775757, "learning_rate": 3.053205508686027e-05, "loss": 4.121, "step": 921550 }, { "epoch": 5.557096513549041, "grad_norm": 4.17453145980835, "learning_rate": 3.051975115705439e-05, "loss": 4.7174, "step": 921600 }, { "epoch": 5.557398005330374, "grad_norm": 3.230431079864502, "learning_rate": 3.050744942611194e-05, "loss": 4.135, "step": 921650 }, { "epoch": 5.557699497111709, "grad_norm": 4.44349479675293, "learning_rate": 3.0495149894259357e-05, "loss": 4.5128, "step": 921700 }, { "epoch": 5.558000988893043, "grad_norm": 5.918879985809326, "learning_rate": 3.0482852561723002e-05, "loss": 4.685, "step": 921750 }, { "epoch": 5.558302480674377, "grad_norm": 3.5076348781585693, "learning_rate": 3.0470557428729214e-05, "loss": 4.5627, "step": 921800 }, { "epoch": 5.55860397245571, "grad_norm": 4.019539833068848, "learning_rate": 3.0458264495504243e-05, "loss": 4.2712, "step": 921850 }, { "epoch": 5.558905464237045, "grad_norm": 4.996609210968018, "learning_rate": 3.044597376227429e-05, "loss": 4.2734, "step": 921900 }, { "epoch": 5.559206956018379, "grad_norm": 2.819617748260498, "learning_rate": 3.043368522926561e-05, "loss": 4.2696, "step": 921950 }, { "epoch": 5.559508447799713, "grad_norm": 4.6230788230896, "learning_rate": 3.0421398896704268e-05, "loss": 4.446, "step": 922000 }, { "epoch": 5.559809939581047, "grad_norm": 2.0545096397399902, "learning_rate": 3.0409114764816455e-05, "loss": 4.3807, "step": 922050 }, { "epoch": 5.560111431362381, "grad_norm": 4.26593017578125, "learning_rate": 3.0396832833828257e-05, "loss": 4.4725, "step": 922100 }, { "epoch": 5.560412923143715, "grad_norm": 4.368889808654785, "learning_rate": 3.0384553103965602e-05, "loss": 4.3518, "step": 922150 }, { "epoch": 5.560714414925049, "grad_norm": 4.932823181152344, "learning_rate": 3.037227557545457e-05, "loss": 4.6035, "step": 922200 }, { "epoch": 5.561015906706383, "grad_norm": 2.0499720573425293, "learning_rate": 3.036000024852112e-05, "loss": 4.4914, "step": 922250 }, { "epoch": 5.561317398487717, "grad_norm": 6.041503429412842, "learning_rate": 3.0347727123391146e-05, "loss": 4.707, "step": 922300 }, { "epoch": 5.5616188902690515, "grad_norm": 2.208461284637451, "learning_rate": 3.0335456200290448e-05, "loss": 4.6499, "step": 922350 }, { "epoch": 5.561920382050385, "grad_norm": 4.362411022186279, "learning_rate": 3.0323187479444943e-05, "loss": 4.5443, "step": 922400 }, { "epoch": 5.562221873831719, "grad_norm": 3.4260406494140625, "learning_rate": 3.0310920961080353e-05, "loss": 4.4775, "step": 922450 }, { "epoch": 5.562523365613053, "grad_norm": 3.0864315032958984, "learning_rate": 3.0298656645422448e-05, "loss": 4.6903, "step": 922500 }, { "epoch": 5.5628248573943875, "grad_norm": 3.603902816772461, "learning_rate": 3.0286394532696993e-05, "loss": 4.7289, "step": 922550 }, { "epoch": 5.563126349175722, "grad_norm": 4.365411758422852, "learning_rate": 3.0274134623129564e-05, "loss": 4.6858, "step": 922600 }, { "epoch": 5.563427840957056, "grad_norm": 2.184048891067505, "learning_rate": 3.0261876916945878e-05, "loss": 4.2535, "step": 922650 }, { "epoch": 5.563729332738389, "grad_norm": 3.97475528717041, "learning_rate": 3.024962141437145e-05, "loss": 4.3503, "step": 922700 }, { "epoch": 5.564030824519723, "grad_norm": 3.258999824523926, "learning_rate": 3.023736811563182e-05, "loss": 3.8381, "step": 922750 }, { "epoch": 5.564332316301058, "grad_norm": 3.4671998023986816, "learning_rate": 3.022511702095253e-05, "loss": 4.1989, "step": 922800 }, { "epoch": 5.564633808082392, "grad_norm": 3.280561923980713, "learning_rate": 3.0212868130559052e-05, "loss": 4.4657, "step": 922850 }, { "epoch": 5.564935299863726, "grad_norm": 4.366532802581787, "learning_rate": 3.020062144467677e-05, "loss": 4.5067, "step": 922900 }, { "epoch": 5.565236791645059, "grad_norm": 4.320262432098389, "learning_rate": 3.018837696353108e-05, "loss": 4.3966, "step": 922950 }, { "epoch": 5.565538283426394, "grad_norm": 3.3899223804473877, "learning_rate": 3.017613468734737e-05, "loss": 3.8923, "step": 923000 }, { "epoch": 5.565839775207728, "grad_norm": 4.790628910064697, "learning_rate": 3.0163894616350857e-05, "loss": 4.4584, "step": 923050 }, { "epoch": 5.566141266989062, "grad_norm": 3.8062996864318848, "learning_rate": 3.01516567507669e-05, "loss": 4.7312, "step": 923100 }, { "epoch": 5.566442758770396, "grad_norm": 2.7767560482025146, "learning_rate": 3.0139421090820664e-05, "loss": 4.6884, "step": 923150 }, { "epoch": 5.56674425055173, "grad_norm": 2.0469939708709717, "learning_rate": 3.012718763673729e-05, "loss": 4.6384, "step": 923200 }, { "epoch": 5.567045742333064, "grad_norm": 3.2056984901428223, "learning_rate": 3.0114956388741958e-05, "loss": 4.926, "step": 923250 }, { "epoch": 5.567347234114398, "grad_norm": 3.859023332595825, "learning_rate": 3.0102727347059806e-05, "loss": 4.3882, "step": 923300 }, { "epoch": 5.567648725895732, "grad_norm": 3.5096750259399414, "learning_rate": 3.0090500511915826e-05, "loss": 4.746, "step": 923350 }, { "epoch": 5.567950217677066, "grad_norm": 3.809393882751465, "learning_rate": 3.0078275883535068e-05, "loss": 4.2621, "step": 923400 }, { "epoch": 5.5682517094584005, "grad_norm": 0.9454516172409058, "learning_rate": 3.0066053462142536e-05, "loss": 4.4397, "step": 923450 }, { "epoch": 5.568553201239734, "grad_norm": 3.3672518730163574, "learning_rate": 3.0053833247963133e-05, "loss": 3.902, "step": 923500 }, { "epoch": 5.568854693021068, "grad_norm": 1.4234912395477295, "learning_rate": 3.004161524122171e-05, "loss": 4.0885, "step": 923550 }, { "epoch": 5.569156184802402, "grad_norm": 2.5759847164154053, "learning_rate": 3.0029399442143228e-05, "loss": 4.4349, "step": 923600 }, { "epoch": 5.5694576765837365, "grad_norm": 4.355066299438477, "learning_rate": 3.001718585095238e-05, "loss": 4.2144, "step": 923650 }, { "epoch": 5.569759168365071, "grad_norm": 4.330026626586914, "learning_rate": 3.000497446787401e-05, "loss": 4.4896, "step": 923700 }, { "epoch": 5.570060660146405, "grad_norm": 3.171405792236328, "learning_rate": 2.9992765293132888e-05, "loss": 4.4719, "step": 923750 }, { "epoch": 5.570362151927738, "grad_norm": 5.177535057067871, "learning_rate": 2.9980558326953614e-05, "loss": 4.45, "step": 923800 }, { "epoch": 5.570663643709072, "grad_norm": 6.154521465301514, "learning_rate": 2.9968353569560943e-05, "loss": 4.5609, "step": 923850 }, { "epoch": 5.570965135490407, "grad_norm": 3.5359604358673096, "learning_rate": 2.9956151021179414e-05, "loss": 4.1836, "step": 923900 }, { "epoch": 5.571266627271741, "grad_norm": 3.743924856185913, "learning_rate": 2.994395068203358e-05, "loss": 4.7689, "step": 923950 }, { "epoch": 5.571568119053074, "grad_norm": 3.264021635055542, "learning_rate": 2.9931752552347992e-05, "loss": 4.7231, "step": 924000 }, { "epoch": 5.571568119053074, "eval_loss": 4.929631233215332, "eval_runtime": 39.1086, "eval_samples_per_second": 13.092, "eval_steps_per_second": 6.546, "eval_tts_loss": 7.884825188461226, "step": 924000 }, { "epoch": 5.571869610834408, "grad_norm": 1.1923657655715942, "learning_rate": 2.991955663234721e-05, "loss": 4.4679, "step": 924050 }, { "epoch": 5.5721711026157426, "grad_norm": 3.4239566326141357, "learning_rate": 2.990736292225558e-05, "loss": 4.6308, "step": 924100 }, { "epoch": 5.572472594397077, "grad_norm": 3.833536148071289, "learning_rate": 2.989517142229754e-05, "loss": 4.6686, "step": 924150 }, { "epoch": 5.572774086178411, "grad_norm": 1.736838698387146, "learning_rate": 2.9882982132697536e-05, "loss": 4.2215, "step": 924200 }, { "epoch": 5.573075577959745, "grad_norm": 4.396650314331055, "learning_rate": 2.9870795053679763e-05, "loss": 4.9441, "step": 924250 }, { "epoch": 5.5733770697410785, "grad_norm": 3.903748035430908, "learning_rate": 2.9858610185468628e-05, "loss": 4.4353, "step": 924300 }, { "epoch": 5.573678561522413, "grad_norm": 4.129205226898193, "learning_rate": 2.984642752828832e-05, "loss": 4.0623, "step": 924350 }, { "epoch": 5.573980053303747, "grad_norm": 4.1361613273620605, "learning_rate": 2.9834247082363005e-05, "loss": 4.4438, "step": 924400 }, { "epoch": 5.574281545085081, "grad_norm": 3.7678637504577637, "learning_rate": 2.982206884791689e-05, "loss": 4.7383, "step": 924450 }, { "epoch": 5.574583036866415, "grad_norm": 9.341135025024414, "learning_rate": 2.9809892825174133e-05, "loss": 4.3289, "step": 924500 }, { "epoch": 5.574884528647749, "grad_norm": 4.273016929626465, "learning_rate": 2.979771901435873e-05, "loss": 4.5033, "step": 924550 }, { "epoch": 5.575186020429083, "grad_norm": 4.56472635269165, "learning_rate": 2.978554741569479e-05, "loss": 4.2039, "step": 924600 }, { "epoch": 5.575487512210417, "grad_norm": 1.2416292428970337, "learning_rate": 2.9773378029406312e-05, "loss": 4.7528, "step": 924650 }, { "epoch": 5.575789003991751, "grad_norm": 5.310318946838379, "learning_rate": 2.9761210855717244e-05, "loss": 4.1895, "step": 924700 }, { "epoch": 5.5760904957730855, "grad_norm": 3.290574312210083, "learning_rate": 2.9749045894851463e-05, "loss": 4.5541, "step": 924750 }, { "epoch": 5.57639198755442, "grad_norm": 3.4432859420776367, "learning_rate": 2.973688314703293e-05, "loss": 4.6685, "step": 924800 }, { "epoch": 5.576693479335753, "grad_norm": 3.411407947540283, "learning_rate": 2.9724722612485398e-05, "loss": 4.4323, "step": 924850 }, { "epoch": 5.576994971117087, "grad_norm": 4.974469184875488, "learning_rate": 2.97125642914327e-05, "loss": 4.8653, "step": 924900 }, { "epoch": 5.577296462898421, "grad_norm": 2.187635898590088, "learning_rate": 2.9700408184098647e-05, "loss": 4.7171, "step": 924950 }, { "epoch": 5.577597954679756, "grad_norm": 5.031187057495117, "learning_rate": 2.9688254290706853e-05, "loss": 4.3904, "step": 925000 }, { "epoch": 5.57789944646109, "grad_norm": 4.042912483215332, "learning_rate": 2.9676102611481078e-05, "loss": 4.6427, "step": 925050 }, { "epoch": 5.578200938242423, "grad_norm": 3.8996543884277344, "learning_rate": 2.966395314664494e-05, "loss": 4.5118, "step": 925100 }, { "epoch": 5.578502430023757, "grad_norm": 3.6069211959838867, "learning_rate": 2.9651805896421965e-05, "loss": 4.4138, "step": 925150 }, { "epoch": 5.5788039218050915, "grad_norm": 3.445436716079712, "learning_rate": 2.963966086103575e-05, "loss": 4.7552, "step": 925200 }, { "epoch": 5.579105413586426, "grad_norm": 4.232401371002197, "learning_rate": 2.9627518040709823e-05, "loss": 4.4174, "step": 925250 }, { "epoch": 5.57940690536776, "grad_norm": 2.932161808013916, "learning_rate": 2.9615377435667688e-05, "loss": 4.8301, "step": 925300 }, { "epoch": 5.579708397149094, "grad_norm": 3.5853271484375, "learning_rate": 2.9603239046132692e-05, "loss": 4.4971, "step": 925350 }, { "epoch": 5.5800098889304275, "grad_norm": 6.647004127502441, "learning_rate": 2.9591102872328266e-05, "loss": 4.9763, "step": 925400 }, { "epoch": 5.580311380711762, "grad_norm": 3.3186066150665283, "learning_rate": 2.957896891447779e-05, "loss": 4.5702, "step": 925450 }, { "epoch": 5.580612872493096, "grad_norm": 3.1854448318481445, "learning_rate": 2.9566837172804558e-05, "loss": 4.4177, "step": 925500 }, { "epoch": 5.58091436427443, "grad_norm": 3.495488405227661, "learning_rate": 2.955470764753177e-05, "loss": 4.2668, "step": 925550 }, { "epoch": 5.581215856055763, "grad_norm": 5.364174842834473, "learning_rate": 2.9542580338882737e-05, "loss": 4.2578, "step": 925600 }, { "epoch": 5.581517347837098, "grad_norm": 5.316311359405518, "learning_rate": 2.9530455247080576e-05, "loss": 4.6073, "step": 925650 }, { "epoch": 5.581818839618432, "grad_norm": 3.5887856483459473, "learning_rate": 2.9518332372348473e-05, "loss": 4.3274, "step": 925700 }, { "epoch": 5.582120331399766, "grad_norm": 4.079892158508301, "learning_rate": 2.9506211714909555e-05, "loss": 4.5511, "step": 925750 }, { "epoch": 5.5824218231811, "grad_norm": 3.361633777618408, "learning_rate": 2.9494093274986824e-05, "loss": 4.3823, "step": 925800 }, { "epoch": 5.5827233149624345, "grad_norm": 2.920114517211914, "learning_rate": 2.9481977052803384e-05, "loss": 4.5884, "step": 925850 }, { "epoch": 5.583024806743769, "grad_norm": 3.1610560417175293, "learning_rate": 2.946986304858216e-05, "loss": 4.5221, "step": 925900 }, { "epoch": 5.583326298525102, "grad_norm": 3.2052338123321533, "learning_rate": 2.9457751262546055e-05, "loss": 4.3357, "step": 925950 }, { "epoch": 5.583627790306436, "grad_norm": 3.824474334716797, "learning_rate": 2.944564169491802e-05, "loss": 4.5149, "step": 926000 }, { "epoch": 5.58392928208777, "grad_norm": 3.521064519882202, "learning_rate": 2.943353434592095e-05, "loss": 4.4162, "step": 926050 }, { "epoch": 5.584230773869105, "grad_norm": 5.106335639953613, "learning_rate": 2.942142921577758e-05, "loss": 4.4187, "step": 926100 }, { "epoch": 5.584532265650438, "grad_norm": 2.817807912826538, "learning_rate": 2.9409326304710736e-05, "loss": 4.2863, "step": 926150 }, { "epoch": 5.584833757431772, "grad_norm": 4.540045261383057, "learning_rate": 2.939722561294318e-05, "loss": 4.5739, "step": 926200 }, { "epoch": 5.585135249213106, "grad_norm": 3.0643820762634277, "learning_rate": 2.9385127140697578e-05, "loss": 4.7975, "step": 926250 }, { "epoch": 5.5854367409944405, "grad_norm": 3.67272686958313, "learning_rate": 2.9373030888196546e-05, "loss": 4.3553, "step": 926300 }, { "epoch": 5.585738232775775, "grad_norm": 2.286738395690918, "learning_rate": 2.936093685566279e-05, "loss": 4.5766, "step": 926350 }, { "epoch": 5.586039724557109, "grad_norm": 3.858682155609131, "learning_rate": 2.9348845043318768e-05, "loss": 4.2809, "step": 926400 }, { "epoch": 5.586341216338442, "grad_norm": 5.103832244873047, "learning_rate": 2.9336755451387083e-05, "loss": 4.2773, "step": 926450 }, { "epoch": 5.5866427081197765, "grad_norm": 2.3345961570739746, "learning_rate": 2.9324668080090275e-05, "loss": 4.6242, "step": 926500 }, { "epoch": 5.586944199901111, "grad_norm": 4.419332027435303, "learning_rate": 2.931258292965068e-05, "loss": 4.4662, "step": 926550 }, { "epoch": 5.587245691682445, "grad_norm": 2.0186734199523926, "learning_rate": 2.930050000029077e-05, "loss": 4.5693, "step": 926600 }, { "epoch": 5.587547183463779, "grad_norm": 4.790592670440674, "learning_rate": 2.928841929223294e-05, "loss": 4.3698, "step": 926650 }, { "epoch": 5.587848675245112, "grad_norm": 3.1269724369049072, "learning_rate": 2.9276340805699504e-05, "loss": 4.1949, "step": 926700 }, { "epoch": 5.588150167026447, "grad_norm": 6.499690532684326, "learning_rate": 2.926426454091269e-05, "loss": 4.4584, "step": 926750 }, { "epoch": 5.588451658807781, "grad_norm": 3.7822771072387695, "learning_rate": 2.9252190498094818e-05, "loss": 4.3916, "step": 926800 }, { "epoch": 5.588753150589115, "grad_norm": 3.687466621398926, "learning_rate": 2.9240118677468045e-05, "loss": 4.7156, "step": 926850 }, { "epoch": 5.589054642370449, "grad_norm": 3.446244716644287, "learning_rate": 2.922804907925454e-05, "loss": 3.8609, "step": 926900 }, { "epoch": 5.5893561341517835, "grad_norm": 4.261351108551025, "learning_rate": 2.9215981703676477e-05, "loss": 4.417, "step": 926950 }, { "epoch": 5.589657625933117, "grad_norm": 4.268089771270752, "learning_rate": 2.9203916550955864e-05, "loss": 4.3537, "step": 927000 }, { "epoch": 5.589657625933117, "eval_loss": 4.93143367767334, "eval_runtime": 38.98, "eval_samples_per_second": 13.135, "eval_steps_per_second": 6.567, "eval_tts_loss": 7.941791993620463, "step": 927000 }, { "epoch": 5.589959117714451, "grad_norm": 3.3007609844207764, "learning_rate": 2.919185362131482e-05, "loss": 4.2626, "step": 927050 }, { "epoch": 5.590260609495785, "grad_norm": 5.446077823638916, "learning_rate": 2.917979291497532e-05, "loss": 4.6661, "step": 927100 }, { "epoch": 5.590562101277119, "grad_norm": 3.716921806335449, "learning_rate": 2.916773443215925e-05, "loss": 4.5505, "step": 927150 }, { "epoch": 5.590863593058454, "grad_norm": 3.760333776473999, "learning_rate": 2.915567817308862e-05, "loss": 4.2751, "step": 927200 }, { "epoch": 5.591165084839787, "grad_norm": 3.4959068298339844, "learning_rate": 2.9143624137985298e-05, "loss": 4.7213, "step": 927250 }, { "epoch": 5.591466576621121, "grad_norm": 3.0775930881500244, "learning_rate": 2.9131572327071064e-05, "loss": 4.4048, "step": 927300 }, { "epoch": 5.591768068402455, "grad_norm": 2.790064811706543, "learning_rate": 2.9119522740567765e-05, "loss": 4.2919, "step": 927350 }, { "epoch": 5.5920695601837895, "grad_norm": 2.6018731594085693, "learning_rate": 2.9107475378697176e-05, "loss": 4.5529, "step": 927400 }, { "epoch": 5.592371051965124, "grad_norm": 3.3437023162841797, "learning_rate": 2.909543024168097e-05, "loss": 4.5305, "step": 927450 }, { "epoch": 5.592672543746458, "grad_norm": 3.3966362476348877, "learning_rate": 2.908338732974081e-05, "loss": 3.9086, "step": 927500 }, { "epoch": 5.592974035527791, "grad_norm": 1.7793071269989014, "learning_rate": 2.9071346643098374e-05, "loss": 4.1173, "step": 927550 }, { "epoch": 5.5932755273091255, "grad_norm": 3.874514102935791, "learning_rate": 2.9059308181975193e-05, "loss": 4.4053, "step": 927600 }, { "epoch": 5.59357701909046, "grad_norm": 4.677847385406494, "learning_rate": 2.9047271946592854e-05, "loss": 4.4781, "step": 927650 }, { "epoch": 5.593878510871794, "grad_norm": 3.5445876121520996, "learning_rate": 2.903523793717291e-05, "loss": 4.6517, "step": 927700 }, { "epoch": 5.594180002653127, "grad_norm": 4.1613898277282715, "learning_rate": 2.902320615393674e-05, "loss": 4.4679, "step": 927750 }, { "epoch": 5.594481494434461, "grad_norm": 3.8414976596832275, "learning_rate": 2.9011176597105812e-05, "loss": 4.6541, "step": 927800 }, { "epoch": 5.594782986215796, "grad_norm": 4.2219648361206055, "learning_rate": 2.8999149266901568e-05, "loss": 4.2518, "step": 927850 }, { "epoch": 5.59508447799713, "grad_norm": 3.1638879776000977, "learning_rate": 2.8987124163545313e-05, "loss": 4.5582, "step": 927900 }, { "epoch": 5.595385969778464, "grad_norm": 3.2350659370422363, "learning_rate": 2.8975101287258284e-05, "loss": 4.2091, "step": 927950 }, { "epoch": 5.595687461559798, "grad_norm": 1.5870246887207031, "learning_rate": 2.896308063826186e-05, "loss": 4.4123, "step": 928000 }, { "epoch": 5.5959889533411316, "grad_norm": 1.2904415130615234, "learning_rate": 2.895106221677716e-05, "loss": 4.0953, "step": 928050 }, { "epoch": 5.596290445122466, "grad_norm": 3.2505767345428467, "learning_rate": 2.893904602302541e-05, "loss": 4.2767, "step": 928100 }, { "epoch": 5.5965919369038, "grad_norm": 3.055290699005127, "learning_rate": 2.8927032057227796e-05, "loss": 4.3548, "step": 928150 }, { "epoch": 5.596893428685134, "grad_norm": 5.577789306640625, "learning_rate": 2.8915020319605343e-05, "loss": 4.5049, "step": 928200 }, { "epoch": 5.597194920466468, "grad_norm": 4.038674831390381, "learning_rate": 2.8903010810379174e-05, "loss": 4.4652, "step": 928250 }, { "epoch": 5.597496412247802, "grad_norm": 5.501264572143555, "learning_rate": 2.889100352977028e-05, "loss": 4.5858, "step": 928300 }, { "epoch": 5.597797904029136, "grad_norm": 3.402705669403076, "learning_rate": 2.8878998477999598e-05, "loss": 4.4466, "step": 928350 }, { "epoch": 5.59809939581047, "grad_norm": 3.7829787731170654, "learning_rate": 2.8866995655288106e-05, "loss": 4.595, "step": 928400 }, { "epoch": 5.598400887591804, "grad_norm": 2.7974448204040527, "learning_rate": 2.8854995061856723e-05, "loss": 4.2226, "step": 928450 }, { "epoch": 5.5987023793731385, "grad_norm": 4.142935276031494, "learning_rate": 2.8842996697926225e-05, "loss": 4.1162, "step": 928500 }, { "epoch": 5.599003871154473, "grad_norm": 3.863124132156372, "learning_rate": 2.883100056371748e-05, "loss": 4.328, "step": 928550 }, { "epoch": 5.599305362935806, "grad_norm": 3.181777000427246, "learning_rate": 2.8819006659451304e-05, "loss": 4.3537, "step": 928600 }, { "epoch": 5.59960685471714, "grad_norm": 4.5132246017456055, "learning_rate": 2.880701498534833e-05, "loss": 4.226, "step": 928650 }, { "epoch": 5.5999083464984745, "grad_norm": 5.243364334106445, "learning_rate": 2.879502554162932e-05, "loss": 4.5184, "step": 928700 }, { "epoch": 5.600209838279809, "grad_norm": 2.77669095993042, "learning_rate": 2.8783038328514874e-05, "loss": 4.6163, "step": 928750 }, { "epoch": 5.600511330061143, "grad_norm": 2.919637680053711, "learning_rate": 2.8771053346225653e-05, "loss": 4.3972, "step": 928800 }, { "epoch": 5.600812821842476, "grad_norm": 2.6052560806274414, "learning_rate": 2.8759070594982148e-05, "loss": 4.6678, "step": 928850 }, { "epoch": 5.60111431362381, "grad_norm": 4.431750297546387, "learning_rate": 2.874709007500493e-05, "loss": 4.2636, "step": 928900 }, { "epoch": 5.601415805405145, "grad_norm": 3.524651527404785, "learning_rate": 2.8735111786514526e-05, "loss": 4.5334, "step": 928950 }, { "epoch": 5.601717297186479, "grad_norm": 6.273021697998047, "learning_rate": 2.8723135729731307e-05, "loss": 4.7686, "step": 929000 }, { "epoch": 5.602018788967813, "grad_norm": 1.661818504333496, "learning_rate": 2.871116190487573e-05, "loss": 4.0329, "step": 929050 }, { "epoch": 5.602320280749147, "grad_norm": 3.5151233673095703, "learning_rate": 2.8699190312168118e-05, "loss": 4.4498, "step": 929100 }, { "epoch": 5.6026217725304805, "grad_norm": 5.050721168518066, "learning_rate": 2.8687220951828782e-05, "loss": 4.0936, "step": 929150 }, { "epoch": 5.602923264311815, "grad_norm": 3.304959297180176, "learning_rate": 2.8675253824078027e-05, "loss": 4.3253, "step": 929200 }, { "epoch": 5.603224756093149, "grad_norm": 5.5087409019470215, "learning_rate": 2.866328892913611e-05, "loss": 4.405, "step": 929250 }, { "epoch": 5.603526247874483, "grad_norm": 2.8901853561401367, "learning_rate": 2.865132626722317e-05, "loss": 4.5465, "step": 929300 }, { "epoch": 5.6038277396558165, "grad_norm": 5.733102321624756, "learning_rate": 2.8639365838559382e-05, "loss": 4.5833, "step": 929350 }, { "epoch": 5.604129231437151, "grad_norm": 4.68875789642334, "learning_rate": 2.8627407643364934e-05, "loss": 4.6386, "step": 929400 }, { "epoch": 5.604430723218485, "grad_norm": 3.9357893466949463, "learning_rate": 2.8615451681859825e-05, "loss": 4.3762, "step": 929450 }, { "epoch": 5.604732214999819, "grad_norm": 3.3248345851898193, "learning_rate": 2.860349795426407e-05, "loss": 4.474, "step": 929500 }, { "epoch": 5.605033706781153, "grad_norm": 4.655608177185059, "learning_rate": 2.8591546460797732e-05, "loss": 4.6465, "step": 929550 }, { "epoch": 5.6053351985624875, "grad_norm": 4.971400737762451, "learning_rate": 2.8579597201680698e-05, "loss": 4.331, "step": 929600 }, { "epoch": 5.605636690343822, "grad_norm": 3.99483323097229, "learning_rate": 2.856765017713288e-05, "loss": 4.7311, "step": 929650 }, { "epoch": 5.605938182125155, "grad_norm": 4.592703342437744, "learning_rate": 2.855570538737421e-05, "loss": 4.4521, "step": 929700 }, { "epoch": 5.606239673906489, "grad_norm": 2.975003480911255, "learning_rate": 2.8543762832624438e-05, "loss": 4.6883, "step": 929750 }, { "epoch": 5.6065411656878235, "grad_norm": 3.397695541381836, "learning_rate": 2.8531822513103385e-05, "loss": 4.5399, "step": 929800 }, { "epoch": 5.606842657469158, "grad_norm": 1.3663970232009888, "learning_rate": 2.8519884429030825e-05, "loss": 3.9957, "step": 929850 }, { "epoch": 5.607144149250491, "grad_norm": 1.1607636213302612, "learning_rate": 2.850794858062645e-05, "loss": 4.2629, "step": 929900 }, { "epoch": 5.607445641031825, "grad_norm": 3.2621543407440186, "learning_rate": 2.8496014968109847e-05, "loss": 4.2433, "step": 929950 }, { "epoch": 5.607747132813159, "grad_norm": 3.6013026237487793, "learning_rate": 2.848408359170073e-05, "loss": 4.8549, "step": 930000 }, { "epoch": 5.607747132813159, "eval_loss": 4.921924591064453, "eval_runtime": 39.2238, "eval_samples_per_second": 13.053, "eval_steps_per_second": 6.527, "eval_tts_loss": 7.899590526893285, "step": 930000 }, { "epoch": 5.608048624594494, "grad_norm": 2.5944058895111084, "learning_rate": 2.8472154451618617e-05, "loss": 4.6541, "step": 930050 }, { "epoch": 5.608350116375828, "grad_norm": 3.6993680000305176, "learning_rate": 2.8460227548083053e-05, "loss": 4.415, "step": 930100 }, { "epoch": 5.608651608157162, "grad_norm": 7.34164571762085, "learning_rate": 2.8448302881313613e-05, "loss": 4.6807, "step": 930150 }, { "epoch": 5.608953099938495, "grad_norm": 3.2281012535095215, "learning_rate": 2.8436380451529635e-05, "loss": 4.3581, "step": 930200 }, { "epoch": 5.6092545917198295, "grad_norm": 3.4770991802215576, "learning_rate": 2.8424460258950643e-05, "loss": 3.9912, "step": 930250 }, { "epoch": 5.609556083501164, "grad_norm": 3.680346727371216, "learning_rate": 2.8412542303795964e-05, "loss": 4.5638, "step": 930300 }, { "epoch": 5.609857575282498, "grad_norm": 3.502990961074829, "learning_rate": 2.840062658628489e-05, "loss": 4.7657, "step": 930350 }, { "epoch": 5.610159067063832, "grad_norm": 4.760255336761475, "learning_rate": 2.8388713106636763e-05, "loss": 4.2227, "step": 930400 }, { "epoch": 5.6104605588451655, "grad_norm": 3.6051645278930664, "learning_rate": 2.837680186507085e-05, "loss": 4.6673, "step": 930450 }, { "epoch": 5.6107620506265, "grad_norm": 3.71985125541687, "learning_rate": 2.8364892861806288e-05, "loss": 4.6808, "step": 930500 }, { "epoch": 5.611063542407834, "grad_norm": 3.374415397644043, "learning_rate": 2.8352986097062308e-05, "loss": 4.385, "step": 930550 }, { "epoch": 5.611365034189168, "grad_norm": 5.2903289794921875, "learning_rate": 2.834108157105804e-05, "loss": 4.364, "step": 930600 }, { "epoch": 5.611666525970502, "grad_norm": 3.644343852996826, "learning_rate": 2.8329179284012555e-05, "loss": 4.4974, "step": 930650 }, { "epoch": 5.6119680177518365, "grad_norm": 4.0646653175354, "learning_rate": 2.8317279236144864e-05, "loss": 4.5376, "step": 930700 }, { "epoch": 5.61226950953317, "grad_norm": 4.3025922775268555, "learning_rate": 2.8305381427674023e-05, "loss": 4.4321, "step": 930750 }, { "epoch": 5.612571001314504, "grad_norm": 2.8795478343963623, "learning_rate": 2.8293485858818925e-05, "loss": 4.3934, "step": 930800 }, { "epoch": 5.612872493095838, "grad_norm": 3.564952850341797, "learning_rate": 2.8281592529798548e-05, "loss": 4.4543, "step": 930850 }, { "epoch": 5.6131739848771725, "grad_norm": 3.54404616355896, "learning_rate": 2.8269701440831793e-05, "loss": 4.1653, "step": 930900 }, { "epoch": 5.613475476658507, "grad_norm": 4.426295280456543, "learning_rate": 2.8257812592137442e-05, "loss": 4.4644, "step": 930950 }, { "epoch": 5.61377696843984, "grad_norm": 3.9967262744903564, "learning_rate": 2.8245925983934297e-05, "loss": 4.3706, "step": 931000 }, { "epoch": 5.614078460221174, "grad_norm": 3.507265090942383, "learning_rate": 2.8234041616441173e-05, "loss": 4.5295, "step": 931050 }, { "epoch": 5.614379952002508, "grad_norm": 1.0475324392318726, "learning_rate": 2.8222159489876738e-05, "loss": 4.3482, "step": 931100 }, { "epoch": 5.614681443783843, "grad_norm": 2.476231813430786, "learning_rate": 2.821027960445964e-05, "loss": 4.4869, "step": 931150 }, { "epoch": 5.614982935565177, "grad_norm": 4.117085933685303, "learning_rate": 2.8198401960408585e-05, "loss": 4.0442, "step": 931200 }, { "epoch": 5.615284427346511, "grad_norm": 3.5715529918670654, "learning_rate": 2.818652655794208e-05, "loss": 4.3785, "step": 931250 }, { "epoch": 5.615585919127844, "grad_norm": 4.559660911560059, "learning_rate": 2.8174653397278718e-05, "loss": 4.1169, "step": 931300 }, { "epoch": 5.6158874109091785, "grad_norm": 2.30100154876709, "learning_rate": 2.8162782478637043e-05, "loss": 4.4907, "step": 931350 }, { "epoch": 5.616188902690513, "grad_norm": 3.3527109622955322, "learning_rate": 2.8150913802235442e-05, "loss": 4.3966, "step": 931400 }, { "epoch": 5.616490394471847, "grad_norm": 4.2308759689331055, "learning_rate": 2.8139047368292427e-05, "loss": 4.2171, "step": 931450 }, { "epoch": 5.61679188625318, "grad_norm": 5.549124717712402, "learning_rate": 2.8127183177026343e-05, "loss": 4.2727, "step": 931500 }, { "epoch": 5.6170933780345145, "grad_norm": 5.605222702026367, "learning_rate": 2.811532122865549e-05, "loss": 4.5228, "step": 931550 }, { "epoch": 5.617394869815849, "grad_norm": 1.3902431726455688, "learning_rate": 2.8103461523398202e-05, "loss": 4.4528, "step": 931600 }, { "epoch": 5.617696361597183, "grad_norm": 3.1361923217773438, "learning_rate": 2.80916040614728e-05, "loss": 4.5981, "step": 931650 }, { "epoch": 5.617997853378517, "grad_norm": 3.2645599842071533, "learning_rate": 2.807974884309741e-05, "loss": 4.739, "step": 931700 }, { "epoch": 5.618299345159851, "grad_norm": 3.5760371685028076, "learning_rate": 2.806789586849026e-05, "loss": 4.0801, "step": 931750 }, { "epoch": 5.618600836941185, "grad_norm": 6.148419380187988, "learning_rate": 2.8056045137869503e-05, "loss": 4.7447, "step": 931800 }, { "epoch": 5.618902328722519, "grad_norm": 4.837556838989258, "learning_rate": 2.804419665145322e-05, "loss": 4.3153, "step": 931850 }, { "epoch": 5.619203820503853, "grad_norm": 4.929124355316162, "learning_rate": 2.803235040945942e-05, "loss": 4.417, "step": 931900 }, { "epoch": 5.619505312285187, "grad_norm": 3.7578125, "learning_rate": 2.8020506412106157e-05, "loss": 4.2762, "step": 931950 }, { "epoch": 5.619806804066521, "grad_norm": 3.445178985595703, "learning_rate": 2.800866465961143e-05, "loss": 4.1617, "step": 932000 }, { "epoch": 5.620108295847855, "grad_norm": 2.778047561645508, "learning_rate": 2.799682515219311e-05, "loss": 4.4342, "step": 932050 }, { "epoch": 5.620409787629189, "grad_norm": 5.485171794891357, "learning_rate": 2.7984987890069104e-05, "loss": 4.3948, "step": 932100 }, { "epoch": 5.620711279410523, "grad_norm": 4.881966590881348, "learning_rate": 2.797315287345731e-05, "loss": 4.2901, "step": 932150 }, { "epoch": 5.621012771191857, "grad_norm": 1.9378710985183716, "learning_rate": 2.7961320102575467e-05, "loss": 4.1736, "step": 932200 }, { "epoch": 5.621314262973192, "grad_norm": 5.985702037811279, "learning_rate": 2.7949489577641403e-05, "loss": 4.2074, "step": 932250 }, { "epoch": 5.621615754754526, "grad_norm": 4.079751968383789, "learning_rate": 2.793766129887281e-05, "loss": 4.6247, "step": 932300 }, { "epoch": 5.621917246535859, "grad_norm": 4.327616214752197, "learning_rate": 2.7925835266487328e-05, "loss": 4.2105, "step": 932350 }, { "epoch": 5.622218738317193, "grad_norm": 4.593883514404297, "learning_rate": 2.7914011480702637e-05, "loss": 4.4171, "step": 932400 }, { "epoch": 5.6225202300985275, "grad_norm": 2.601412773132324, "learning_rate": 2.7902189941736376e-05, "loss": 4.4263, "step": 932450 }, { "epoch": 5.622821721879862, "grad_norm": 3.6446754932403564, "learning_rate": 2.7890370649806036e-05, "loss": 4.5581, "step": 932500 }, { "epoch": 5.623123213661196, "grad_norm": 4.2090983390808105, "learning_rate": 2.787855360512915e-05, "loss": 4.3705, "step": 932550 }, { "epoch": 5.623424705442529, "grad_norm": 2.901416063308716, "learning_rate": 2.7866738807923257e-05, "loss": 4.6911, "step": 932600 }, { "epoch": 5.6237261972238635, "grad_norm": 3.586110830307007, "learning_rate": 2.7854926258405745e-05, "loss": 4.088, "step": 932650 }, { "epoch": 5.624027689005198, "grad_norm": 2.6409342288970947, "learning_rate": 2.7843115956793966e-05, "loss": 4.6337, "step": 932700 }, { "epoch": 5.624329180786532, "grad_norm": 2.331068992614746, "learning_rate": 2.783130790330534e-05, "loss": 4.0291, "step": 932750 }, { "epoch": 5.624630672567866, "grad_norm": 4.7809014320373535, "learning_rate": 2.7819502098157108e-05, "loss": 4.2599, "step": 932800 }, { "epoch": 5.6249321643492, "grad_norm": 4.895135402679443, "learning_rate": 2.7807698541566587e-05, "loss": 4.7078, "step": 932850 }, { "epoch": 5.625233656130534, "grad_norm": 0.8235400319099426, "learning_rate": 2.7795897233751028e-05, "loss": 4.3141, "step": 932900 }, { "epoch": 5.625535147911868, "grad_norm": 3.9072227478027344, "learning_rate": 2.7784098174927565e-05, "loss": 4.5375, "step": 932950 }, { "epoch": 5.625836639693202, "grad_norm": 1.646905541419983, "learning_rate": 2.777230136531339e-05, "loss": 4.235, "step": 933000 }, { "epoch": 5.625836639693202, "eval_loss": 4.921823501586914, "eval_runtime": 39.0103, "eval_samples_per_second": 13.125, "eval_steps_per_second": 6.562, "eval_tts_loss": 7.9250209300660375, "step": 933000 }, { "epoch": 5.626138131474536, "grad_norm": 4.772911548614502, "learning_rate": 2.776050680512557e-05, "loss": 4.623, "step": 933050 }, { "epoch": 5.6264396232558695, "grad_norm": 3.89752197265625, "learning_rate": 2.7748714494581158e-05, "loss": 4.2005, "step": 933100 }, { "epoch": 5.626741115037204, "grad_norm": 5.297937393188477, "learning_rate": 2.773692443389718e-05, "loss": 4.3144, "step": 933150 }, { "epoch": 5.627042606818538, "grad_norm": 3.5621161460876465, "learning_rate": 2.7725136623290673e-05, "loss": 4.5467, "step": 933200 }, { "epoch": 5.627344098599872, "grad_norm": 4.140651702880859, "learning_rate": 2.77133510629785e-05, "loss": 4.5219, "step": 933250 }, { "epoch": 5.627645590381206, "grad_norm": 3.687885046005249, "learning_rate": 2.7701567753177572e-05, "loss": 4.3003, "step": 933300 }, { "epoch": 5.627947082162541, "grad_norm": 2.6079118251800537, "learning_rate": 2.7689786694104817e-05, "loss": 4.1227, "step": 933350 }, { "epoch": 5.628248573943875, "grad_norm": 5.1890106201171875, "learning_rate": 2.7678007885976938e-05, "loss": 4.4592, "step": 933400 }, { "epoch": 5.628550065725208, "grad_norm": 5.283238887786865, "learning_rate": 2.7666231329010803e-05, "loss": 4.5524, "step": 933450 }, { "epoch": 5.628851557506542, "grad_norm": 3.9052324295043945, "learning_rate": 2.76544570234231e-05, "loss": 4.2551, "step": 933500 }, { "epoch": 5.6291530492878765, "grad_norm": 4.628576278686523, "learning_rate": 2.764268496943049e-05, "loss": 4.6173, "step": 933550 }, { "epoch": 5.629454541069211, "grad_norm": 5.481235504150391, "learning_rate": 2.7630915167249635e-05, "loss": 4.36, "step": 933600 }, { "epoch": 5.629756032850544, "grad_norm": 2.138251304626465, "learning_rate": 2.761914761709721e-05, "loss": 4.3385, "step": 933650 }, { "epoch": 5.630057524631878, "grad_norm": 3.6044187545776367, "learning_rate": 2.7607382319189675e-05, "loss": 4.1963, "step": 933700 }, { "epoch": 5.6303590164132125, "grad_norm": 6.300692081451416, "learning_rate": 2.759561927374361e-05, "loss": 4.3194, "step": 933750 }, { "epoch": 5.630660508194547, "grad_norm": 3.8008148670196533, "learning_rate": 2.7583858480975533e-05, "loss": 4.0096, "step": 933800 }, { "epoch": 5.630961999975881, "grad_norm": 3.506411075592041, "learning_rate": 2.757209994110183e-05, "loss": 4.3439, "step": 933850 }, { "epoch": 5.631263491757215, "grad_norm": 4.700475215911865, "learning_rate": 2.7560343654338897e-05, "loss": 4.4149, "step": 933900 }, { "epoch": 5.631564983538548, "grad_norm": 4.882603645324707, "learning_rate": 2.7548589620903132e-05, "loss": 4.1842, "step": 933950 }, { "epoch": 5.631866475319883, "grad_norm": 5.386880397796631, "learning_rate": 2.753683784101079e-05, "loss": 4.6927, "step": 934000 }, { "epoch": 5.632167967101217, "grad_norm": 3.262742757797241, "learning_rate": 2.7525088314878192e-05, "loss": 4.2537, "step": 934050 }, { "epoch": 5.632469458882551, "grad_norm": 4.276177406311035, "learning_rate": 2.75133410427216e-05, "loss": 4.3246, "step": 934100 }, { "epoch": 5.632770950663885, "grad_norm": 5.102374076843262, "learning_rate": 2.7501596024757123e-05, "loss": 4.1026, "step": 934150 }, { "epoch": 5.6330724424452185, "grad_norm": 4.262080192565918, "learning_rate": 2.748985326120099e-05, "loss": 4.4564, "step": 934200 }, { "epoch": 5.633373934226553, "grad_norm": 4.122506618499756, "learning_rate": 2.7478112752269242e-05, "loss": 4.6752, "step": 934250 }, { "epoch": 5.633675426007887, "grad_norm": 3.817234992980957, "learning_rate": 2.7466374498178027e-05, "loss": 4.3984, "step": 934300 }, { "epoch": 5.633976917789221, "grad_norm": 5.540943145751953, "learning_rate": 2.7454638499143272e-05, "loss": 3.9255, "step": 934350 }, { "epoch": 5.634278409570555, "grad_norm": 4.331048965454102, "learning_rate": 2.7442904755381067e-05, "loss": 4.3024, "step": 934400 }, { "epoch": 5.63457990135189, "grad_norm": 5.2793169021606445, "learning_rate": 2.7431173267107258e-05, "loss": 4.5802, "step": 934450 }, { "epoch": 5.634881393133223, "grad_norm": 4.8471879959106445, "learning_rate": 2.741944403453779e-05, "loss": 4.4744, "step": 934500 }, { "epoch": 5.635182884914557, "grad_norm": 3.5925779342651367, "learning_rate": 2.7407717057888558e-05, "loss": 4.4211, "step": 934550 }, { "epoch": 5.635484376695891, "grad_norm": 4.899122714996338, "learning_rate": 2.7395992337375318e-05, "loss": 4.4357, "step": 934600 }, { "epoch": 5.6357858684772255, "grad_norm": 3.743265151977539, "learning_rate": 2.73842698732139e-05, "loss": 4.4773, "step": 934650 }, { "epoch": 5.63608736025856, "grad_norm": 2.6258251667022705, "learning_rate": 2.7372549665620015e-05, "loss": 4.6946, "step": 934700 }, { "epoch": 5.636388852039893, "grad_norm": 3.4521689414978027, "learning_rate": 2.7360831714809307e-05, "loss": 4.3962, "step": 934750 }, { "epoch": 5.636690343821227, "grad_norm": 6.163872718811035, "learning_rate": 2.7349116020997485e-05, "loss": 4.359, "step": 934800 }, { "epoch": 5.6369918356025615, "grad_norm": 3.063248872756958, "learning_rate": 2.7337402584400196e-05, "loss": 4.183, "step": 934850 }, { "epoch": 5.637293327383896, "grad_norm": 3.130894184112549, "learning_rate": 2.732569140523292e-05, "loss": 4.4688, "step": 934900 }, { "epoch": 5.63759481916523, "grad_norm": 3.28401255607605, "learning_rate": 2.7313982483711226e-05, "loss": 4.353, "step": 934950 }, { "epoch": 5.637896310946564, "grad_norm": 4.045835018157959, "learning_rate": 2.7302275820050634e-05, "loss": 4.188, "step": 935000 }, { "epoch": 5.638197802727897, "grad_norm": 4.1778059005737305, "learning_rate": 2.7290571414466567e-05, "loss": 4.6006, "step": 935050 }, { "epoch": 5.638499294509232, "grad_norm": 4.768836975097656, "learning_rate": 2.7278869267174387e-05, "loss": 4.4453, "step": 935100 }, { "epoch": 5.638800786290566, "grad_norm": 4.229702472686768, "learning_rate": 2.7267169378389503e-05, "loss": 4.5322, "step": 935150 }, { "epoch": 5.6391022780719, "grad_norm": 3.719046115875244, "learning_rate": 2.7255471748327196e-05, "loss": 4.867, "step": 935200 }, { "epoch": 5.639403769853233, "grad_norm": 3.690725088119507, "learning_rate": 2.7243776377202758e-05, "loss": 4.3911, "step": 935250 }, { "epoch": 5.6397052616345675, "grad_norm": 4.39017391204834, "learning_rate": 2.7232083265231437e-05, "loss": 4.2124, "step": 935300 }, { "epoch": 5.640006753415902, "grad_norm": 6.672234535217285, "learning_rate": 2.7220392412628457e-05, "loss": 4.6062, "step": 935350 }, { "epoch": 5.640308245197236, "grad_norm": 5.33302116394043, "learning_rate": 2.7208703819608894e-05, "loss": 4.1304, "step": 935400 }, { "epoch": 5.64060973697857, "grad_norm": 4.585775852203369, "learning_rate": 2.7197017486387952e-05, "loss": 4.3158, "step": 935450 }, { "epoch": 5.640911228759904, "grad_norm": 3.91744065284729, "learning_rate": 2.7185333413180648e-05, "loss": 4.7161, "step": 935500 }, { "epoch": 5.641212720541238, "grad_norm": 5.783289909362793, "learning_rate": 2.7173651600201983e-05, "loss": 4.2262, "step": 935550 }, { "epoch": 5.641514212322572, "grad_norm": 4.587335109710693, "learning_rate": 2.7161972047666984e-05, "loss": 4.3888, "step": 935600 }, { "epoch": 5.641815704103906, "grad_norm": 4.812171936035156, "learning_rate": 2.715029475579063e-05, "loss": 4.3717, "step": 935650 }, { "epoch": 5.64211719588524, "grad_norm": 2.6740269660949707, "learning_rate": 2.713861972478773e-05, "loss": 4.1723, "step": 935700 }, { "epoch": 5.6424186876665745, "grad_norm": 4.2135419845581055, "learning_rate": 2.7126946954873213e-05, "loss": 4.6482, "step": 935750 }, { "epoch": 5.642720179447908, "grad_norm": 3.4580748081207275, "learning_rate": 2.7115276446261923e-05, "loss": 4.486, "step": 935800 }, { "epoch": 5.643021671229242, "grad_norm": 3.268303394317627, "learning_rate": 2.7103608199168608e-05, "loss": 4.2623, "step": 935850 }, { "epoch": 5.643323163010576, "grad_norm": 3.4807074069976807, "learning_rate": 2.709194221380796e-05, "loss": 4.4047, "step": 935900 }, { "epoch": 5.6436246547919104, "grad_norm": 5.069052219390869, "learning_rate": 2.7080278490394758e-05, "loss": 4.6758, "step": 935950 }, { "epoch": 5.643926146573245, "grad_norm": 5.303681373596191, "learning_rate": 2.706861702914358e-05, "loss": 4.555, "step": 936000 }, { "epoch": 5.643926146573245, "eval_loss": 4.9123969078063965, "eval_runtime": 38.8783, "eval_samples_per_second": 13.169, "eval_steps_per_second": 6.585, "eval_tts_loss": 7.956236333215077, "step": 936000 }, { "epoch": 5.644227638354579, "grad_norm": 2.643815279006958, "learning_rate": 2.7056957830269055e-05, "loss": 4.4554, "step": 936050 }, { "epoch": 5.644529130135912, "grad_norm": 5.230113983154297, "learning_rate": 2.7045300893985827e-05, "loss": 4.7006, "step": 936100 }, { "epoch": 5.644830621917246, "grad_norm": 4.834632396697998, "learning_rate": 2.7033646220508325e-05, "loss": 4.5635, "step": 936150 }, { "epoch": 5.645132113698581, "grad_norm": 4.246316432952881, "learning_rate": 2.702199381005113e-05, "loss": 4.342, "step": 936200 }, { "epoch": 5.645433605479915, "grad_norm": 2.686471462249756, "learning_rate": 2.7010343662828633e-05, "loss": 4.4293, "step": 936250 }, { "epoch": 5.645735097261249, "grad_norm": 5.1103057861328125, "learning_rate": 2.6998695779055195e-05, "loss": 4.4802, "step": 936300 }, { "epoch": 5.646036589042582, "grad_norm": 7.051990032196045, "learning_rate": 2.698705015894525e-05, "loss": 4.0247, "step": 936350 }, { "epoch": 5.6463380808239165, "grad_norm": 5.190579891204834, "learning_rate": 2.6975406802713123e-05, "loss": 4.8569, "step": 936400 }, { "epoch": 5.646639572605251, "grad_norm": 3.1096811294555664, "learning_rate": 2.6963765710573028e-05, "loss": 4.4139, "step": 936450 }, { "epoch": 5.646941064386585, "grad_norm": 3.405594825744629, "learning_rate": 2.6952126882739255e-05, "loss": 4.7196, "step": 936500 }, { "epoch": 5.647242556167919, "grad_norm": 1.7275681495666504, "learning_rate": 2.694049031942602e-05, "loss": 4.2713, "step": 936550 }, { "epoch": 5.647544047949253, "grad_norm": 8.045683860778809, "learning_rate": 2.6928856020847407e-05, "loss": 4.5904, "step": 936600 }, { "epoch": 5.647845539730587, "grad_norm": 4.426901340484619, "learning_rate": 2.69172239872176e-05, "loss": 4.4857, "step": 936650 }, { "epoch": 5.648147031511921, "grad_norm": 4.68492317199707, "learning_rate": 2.690559421875064e-05, "loss": 4.6281, "step": 936700 }, { "epoch": 5.648448523293255, "grad_norm": 6.164378643035889, "learning_rate": 2.6893966715660526e-05, "loss": 4.7805, "step": 936750 }, { "epoch": 5.648750015074589, "grad_norm": 3.2431740760803223, "learning_rate": 2.6882341478161263e-05, "loss": 4.5955, "step": 936800 }, { "epoch": 5.649051506855923, "grad_norm": 4.144867420196533, "learning_rate": 2.6870718506466843e-05, "loss": 4.8757, "step": 936850 }, { "epoch": 5.649352998637257, "grad_norm": 4.599395275115967, "learning_rate": 2.6859097800791095e-05, "loss": 4.5445, "step": 936900 }, { "epoch": 5.649654490418591, "grad_norm": 5.5592474937438965, "learning_rate": 2.6847479361347917e-05, "loss": 4.2671, "step": 936950 }, { "epoch": 5.649955982199925, "grad_norm": 3.360689163208008, "learning_rate": 2.6835863188351187e-05, "loss": 4.8668, "step": 937000 }, { "epoch": 5.650257473981259, "grad_norm": 1.7794976234436035, "learning_rate": 2.6824249282014616e-05, "loss": 4.1789, "step": 937050 }, { "epoch": 5.650558965762594, "grad_norm": 2.856374740600586, "learning_rate": 2.6812637642551932e-05, "loss": 4.4639, "step": 937100 }, { "epoch": 5.650860457543928, "grad_norm": 3.606821060180664, "learning_rate": 2.6801028270176882e-05, "loss": 4.5331, "step": 937150 }, { "epoch": 5.651161949325261, "grad_norm": 3.5061733722686768, "learning_rate": 2.6789421165103048e-05, "loss": 4.6768, "step": 937200 }, { "epoch": 5.651463441106595, "grad_norm": 5.416250705718994, "learning_rate": 2.6777816327544104e-05, "loss": 4.6496, "step": 937250 }, { "epoch": 5.65176493288793, "grad_norm": 4.9130778312683105, "learning_rate": 2.6766213757713618e-05, "loss": 4.7082, "step": 937300 }, { "epoch": 5.652066424669264, "grad_norm": 6.568314075469971, "learning_rate": 2.675461345582508e-05, "loss": 4.8819, "step": 937350 }, { "epoch": 5.652367916450597, "grad_norm": 4.4004693031311035, "learning_rate": 2.674301542209204e-05, "loss": 4.6613, "step": 937400 }, { "epoch": 5.652669408231931, "grad_norm": 3.3808677196502686, "learning_rate": 2.673141965672791e-05, "loss": 4.7861, "step": 937450 }, { "epoch": 5.6529709000132655, "grad_norm": 2.019369602203369, "learning_rate": 2.6719826159946033e-05, "loss": 4.3698, "step": 937500 }, { "epoch": 5.6532723917946, "grad_norm": 4.130560874938965, "learning_rate": 2.6708234931959844e-05, "loss": 4.1467, "step": 937550 }, { "epoch": 5.653573883575934, "grad_norm": 3.4101269245147705, "learning_rate": 2.6696645972982668e-05, "loss": 4.7973, "step": 937600 }, { "epoch": 5.653875375357268, "grad_norm": 6.018741607666016, "learning_rate": 2.6685059283227717e-05, "loss": 4.6042, "step": 937650 }, { "epoch": 5.6541768671386015, "grad_norm": 4.2228779792785645, "learning_rate": 2.667347486290829e-05, "loss": 4.8993, "step": 937700 }, { "epoch": 5.654478358919936, "grad_norm": 4.570728778839111, "learning_rate": 2.6661892712237582e-05, "loss": 4.3432, "step": 937750 }, { "epoch": 5.65477985070127, "grad_norm": 3.5152270793914795, "learning_rate": 2.6650312831428687e-05, "loss": 4.0201, "step": 937800 }, { "epoch": 5.655081342482604, "grad_norm": 3.501054048538208, "learning_rate": 2.6638735220694806e-05, "loss": 4.4818, "step": 937850 }, { "epoch": 5.655382834263938, "grad_norm": 3.6647534370422363, "learning_rate": 2.662715988024895e-05, "loss": 4.6468, "step": 937900 }, { "epoch": 5.655684326045272, "grad_norm": 3.8741836547851562, "learning_rate": 2.6615586810304113e-05, "loss": 4.8379, "step": 937950 }, { "epoch": 5.655985817826606, "grad_norm": 5.1905646324157715, "learning_rate": 2.6604016011073325e-05, "loss": 4.6512, "step": 938000 }, { "epoch": 5.65628730960794, "grad_norm": 2.4341695308685303, "learning_rate": 2.6592447482769567e-05, "loss": 4.2939, "step": 938050 }, { "epoch": 5.656588801389274, "grad_norm": 3.5847907066345215, "learning_rate": 2.658088122560567e-05, "loss": 4.1017, "step": 938100 }, { "epoch": 5.656890293170608, "grad_norm": 3.2804908752441406, "learning_rate": 2.6569317239794512e-05, "loss": 4.7056, "step": 938150 }, { "epoch": 5.657191784951943, "grad_norm": 4.568192005157471, "learning_rate": 2.6557755525548973e-05, "loss": 4.5252, "step": 938200 }, { "epoch": 5.657493276733276, "grad_norm": 0.961000919342041, "learning_rate": 2.6546196083081785e-05, "loss": 4.3893, "step": 938250 }, { "epoch": 5.65779476851461, "grad_norm": 5.155702114105225, "learning_rate": 2.6534638912605622e-05, "loss": 4.3423, "step": 938300 }, { "epoch": 5.658096260295944, "grad_norm": 4.530178070068359, "learning_rate": 2.6523084014333285e-05, "loss": 4.1693, "step": 938350 }, { "epoch": 5.658397752077279, "grad_norm": 3.172532558441162, "learning_rate": 2.6511531388477336e-05, "loss": 4.5188, "step": 938400 }, { "epoch": 5.658699243858613, "grad_norm": 4.686961650848389, "learning_rate": 2.649998103525044e-05, "loss": 4.7654, "step": 938450 }, { "epoch": 5.659000735639946, "grad_norm": 4.281839370727539, "learning_rate": 2.6488432954865157e-05, "loss": 4.9409, "step": 938500 }, { "epoch": 5.65930222742128, "grad_norm": 3.6817626953125, "learning_rate": 2.6476887147533988e-05, "loss": 4.8506, "step": 938550 }, { "epoch": 5.6596037192026145, "grad_norm": 7.764769554138184, "learning_rate": 2.6465343613469477e-05, "loss": 4.9032, "step": 938600 }, { "epoch": 5.659905210983949, "grad_norm": 4.237313747406006, "learning_rate": 2.6453802352883968e-05, "loss": 4.8196, "step": 938650 }, { "epoch": 5.660206702765283, "grad_norm": 6.225255489349365, "learning_rate": 2.644226336598996e-05, "loss": 4.2253, "step": 938700 }, { "epoch": 5.660508194546617, "grad_norm": 6.711450576782227, "learning_rate": 2.643072665299974e-05, "loss": 4.6471, "step": 938750 }, { "epoch": 5.6608096863279505, "grad_norm": 6.750197410583496, "learning_rate": 2.6419192214125644e-05, "loss": 4.3628, "step": 938800 }, { "epoch": 5.661111178109285, "grad_norm": 5.947877407073975, "learning_rate": 2.640766004957999e-05, "loss": 4.2556, "step": 938850 }, { "epoch": 5.661412669890619, "grad_norm": 3.669050693511963, "learning_rate": 2.6396130159574948e-05, "loss": 4.6003, "step": 938900 }, { "epoch": 5.661714161671953, "grad_norm": 3.6720409393310547, "learning_rate": 2.6384602544322735e-05, "loss": 4.5142, "step": 938950 }, { "epoch": 5.662015653453286, "grad_norm": 3.514871835708618, "learning_rate": 2.6373077204035543e-05, "loss": 4.5515, "step": 939000 }, { "epoch": 5.662015653453286, "eval_loss": 4.9086456298828125, "eval_runtime": 38.8639, "eval_samples_per_second": 13.174, "eval_steps_per_second": 6.587, "eval_tts_loss": 7.855476577402807, "step": 939000 }, { "epoch": 5.662317145234621, "grad_norm": 4.981222629547119, "learning_rate": 2.6361554138925444e-05, "loss": 4.7253, "step": 939050 }, { "epoch": 5.662618637015955, "grad_norm": 3.7231497764587402, "learning_rate": 2.6350033349204458e-05, "loss": 4.6447, "step": 939100 }, { "epoch": 5.662920128797289, "grad_norm": 2.016679048538208, "learning_rate": 2.6338514835084694e-05, "loss": 4.3626, "step": 939150 }, { "epoch": 5.663221620578623, "grad_norm": 4.128783702850342, "learning_rate": 2.6326998596778037e-05, "loss": 4.4359, "step": 939200 }, { "epoch": 5.663523112359957, "grad_norm": 4.665727615356445, "learning_rate": 2.63154846344965e-05, "loss": 4.6114, "step": 939250 }, { "epoch": 5.663824604141291, "grad_norm": 3.944793462753296, "learning_rate": 2.6303972948451974e-05, "loss": 4.6344, "step": 939300 }, { "epoch": 5.664126095922625, "grad_norm": 3.9695520401000977, "learning_rate": 2.6292463538856277e-05, "loss": 4.5474, "step": 939350 }, { "epoch": 5.664427587703959, "grad_norm": 4.5076446533203125, "learning_rate": 2.6280956405921293e-05, "loss": 4.411, "step": 939400 }, { "epoch": 5.664729079485293, "grad_norm": 2.7592556476593018, "learning_rate": 2.6269451549858735e-05, "loss": 4.1504, "step": 939450 }, { "epoch": 5.665030571266628, "grad_norm": 3.304349660873413, "learning_rate": 2.6257948970880327e-05, "loss": 4.1621, "step": 939500 }, { "epoch": 5.665332063047961, "grad_norm": 3.9739432334899902, "learning_rate": 2.624644866919777e-05, "loss": 4.3543, "step": 939550 }, { "epoch": 5.665633554829295, "grad_norm": 3.0781567096710205, "learning_rate": 2.623495064502275e-05, "loss": 4.4341, "step": 939600 }, { "epoch": 5.665935046610629, "grad_norm": 5.080848217010498, "learning_rate": 2.622345489856681e-05, "loss": 3.9073, "step": 939650 }, { "epoch": 5.6662365383919635, "grad_norm": 3.7085299491882324, "learning_rate": 2.6211961430041544e-05, "loss": 4.7234, "step": 939700 }, { "epoch": 5.666538030173298, "grad_norm": 2.923008441925049, "learning_rate": 2.6200470239658522e-05, "loss": 4.6749, "step": 939750 }, { "epoch": 5.666839521954632, "grad_norm": 6.158195972442627, "learning_rate": 2.6188981327629153e-05, "loss": 4.402, "step": 939800 }, { "epoch": 5.667141013735965, "grad_norm": 3.9323182106018066, "learning_rate": 2.6177494694164874e-05, "loss": 4.1395, "step": 939850 }, { "epoch": 5.6674425055172994, "grad_norm": 3.6965384483337402, "learning_rate": 2.6166010339477128e-05, "loss": 4.7152, "step": 939900 }, { "epoch": 5.667743997298634, "grad_norm": 3.874079465866089, "learning_rate": 2.6154528263777215e-05, "loss": 4.4424, "step": 939950 }, { "epoch": 5.668045489079968, "grad_norm": 4.122894287109375, "learning_rate": 2.614304846727646e-05, "loss": 4.7429, "step": 940000 }, { "epoch": 5.668346980861302, "grad_norm": 6.174907207489014, "learning_rate": 2.613157095018619e-05, "loss": 4.5231, "step": 940050 }, { "epoch": 5.668648472642635, "grad_norm": 4.002656936645508, "learning_rate": 2.6120095712717554e-05, "loss": 4.7369, "step": 940100 }, { "epoch": 5.66894996442397, "grad_norm": 3.842039108276367, "learning_rate": 2.610862275508176e-05, "loss": 4.5982, "step": 940150 }, { "epoch": 5.669251456205304, "grad_norm": 4.228862285614014, "learning_rate": 2.6097152077490003e-05, "loss": 4.4741, "step": 940200 }, { "epoch": 5.669552947986638, "grad_norm": 2.9788222312927246, "learning_rate": 2.6085683680153345e-05, "loss": 4.4567, "step": 940250 }, { "epoch": 5.669854439767972, "grad_norm": 3.8695755004882812, "learning_rate": 2.6074217563282807e-05, "loss": 4.3731, "step": 940300 }, { "epoch": 5.670155931549306, "grad_norm": 3.683368682861328, "learning_rate": 2.6062753727089465e-05, "loss": 4.2382, "step": 940350 }, { "epoch": 5.67045742333064, "grad_norm": 3.318497657775879, "learning_rate": 2.605129217178425e-05, "loss": 4.3299, "step": 940400 }, { "epoch": 5.670758915111974, "grad_norm": 3.5569207668304443, "learning_rate": 2.6039832897578112e-05, "loss": 4.3404, "step": 940450 }, { "epoch": 5.671060406893308, "grad_norm": 4.5724663734436035, "learning_rate": 2.6028375904681997e-05, "loss": 4.5656, "step": 940500 }, { "epoch": 5.671361898674642, "grad_norm": 4.437685012817383, "learning_rate": 2.6016921193306656e-05, "loss": 4.367, "step": 940550 }, { "epoch": 5.671663390455976, "grad_norm": 3.019685745239258, "learning_rate": 2.6005468763663e-05, "loss": 4.5703, "step": 940600 }, { "epoch": 5.67196488223731, "grad_norm": 3.0341451168060303, "learning_rate": 2.599401861596173e-05, "loss": 4.0684, "step": 940650 }, { "epoch": 5.672266374018644, "grad_norm": 3.2396278381347656, "learning_rate": 2.5982570750413544e-05, "loss": 4.4565, "step": 940700 }, { "epoch": 5.672567865799978, "grad_norm": 4.807519435882568, "learning_rate": 2.5971125167229157e-05, "loss": 4.5827, "step": 940750 }, { "epoch": 5.6728693575813125, "grad_norm": 3.6970179080963135, "learning_rate": 2.5959681866619266e-05, "loss": 4.0588, "step": 940800 }, { "epoch": 5.673170849362647, "grad_norm": 7.271252155303955, "learning_rate": 2.5948240848794365e-05, "loss": 4.5256, "step": 940850 }, { "epoch": 5.673472341143981, "grad_norm": 3.221006393432617, "learning_rate": 2.5936802113965062e-05, "loss": 4.1387, "step": 940900 }, { "epoch": 5.673773832925314, "grad_norm": 3.699021577835083, "learning_rate": 2.592536566234191e-05, "loss": 4.6699, "step": 940950 }, { "epoch": 5.674075324706648, "grad_norm": 3.495826005935669, "learning_rate": 2.591393149413529e-05, "loss": 4.6036, "step": 941000 }, { "epoch": 5.674376816487983, "grad_norm": 4.10322904586792, "learning_rate": 2.5902499609555733e-05, "loss": 4.3902, "step": 941050 }, { "epoch": 5.674678308269317, "grad_norm": 3.5449094772338867, "learning_rate": 2.5891070008813562e-05, "loss": 4.3005, "step": 941100 }, { "epoch": 5.67497980005065, "grad_norm": 4.1083502769470215, "learning_rate": 2.5879642692119112e-05, "loss": 4.2135, "step": 941150 }, { "epoch": 5.675281291831984, "grad_norm": 5.28841495513916, "learning_rate": 2.5868217659682694e-05, "loss": 4.3653, "step": 941200 }, { "epoch": 5.675582783613319, "grad_norm": 1.2386609315872192, "learning_rate": 2.5856794911714622e-05, "loss": 4.3987, "step": 941250 }, { "epoch": 5.675884275394653, "grad_norm": 3.1964454650878906, "learning_rate": 2.584537444842505e-05, "loss": 4.3569, "step": 941300 }, { "epoch": 5.676185767175987, "grad_norm": 4.481339931488037, "learning_rate": 2.5833956270024186e-05, "loss": 4.7804, "step": 941350 }, { "epoch": 5.676487258957321, "grad_norm": 3.9960012435913086, "learning_rate": 2.5822540376722184e-05, "loss": 4.4792, "step": 941400 }, { "epoch": 5.6767887507386545, "grad_norm": 3.2547550201416016, "learning_rate": 2.5811126768729124e-05, "loss": 4.5631, "step": 941450 }, { "epoch": 5.677090242519989, "grad_norm": 3.3024349212646484, "learning_rate": 2.579971544625502e-05, "loss": 3.9907, "step": 941500 }, { "epoch": 5.677391734301323, "grad_norm": 4.2919535636901855, "learning_rate": 2.5788306409509925e-05, "loss": 4.5891, "step": 941550 }, { "epoch": 5.677693226082657, "grad_norm": 1.0230029821395874, "learning_rate": 2.5776899658703782e-05, "loss": 4.3528, "step": 941600 }, { "epoch": 5.677994717863991, "grad_norm": 3.2548255920410156, "learning_rate": 2.576549519404651e-05, "loss": 4.199, "step": 941650 }, { "epoch": 5.678296209645325, "grad_norm": 6.287138938903809, "learning_rate": 2.575409301574804e-05, "loss": 4.9925, "step": 941700 }, { "epoch": 5.678597701426659, "grad_norm": 4.074394226074219, "learning_rate": 2.5742693124018137e-05, "loss": 4.5907, "step": 941750 }, { "epoch": 5.678899193207993, "grad_norm": 3.1042346954345703, "learning_rate": 2.57312955190667e-05, "loss": 4.3848, "step": 941800 }, { "epoch": 5.679200684989327, "grad_norm": 7.748051643371582, "learning_rate": 2.5719900201103417e-05, "loss": 4.4686, "step": 941850 }, { "epoch": 5.6795021767706615, "grad_norm": 3.2393577098846436, "learning_rate": 2.570850717033798e-05, "loss": 3.7901, "step": 941900 }, { "epoch": 5.679803668551996, "grad_norm": 3.24604868888855, "learning_rate": 2.5697116426980085e-05, "loss": 4.623, "step": 941950 }, { "epoch": 5.680105160333329, "grad_norm": 4.97883415222168, "learning_rate": 2.5685727971239373e-05, "loss": 4.3645, "step": 942000 }, { "epoch": 5.680105160333329, "eval_loss": 4.910106658935547, "eval_runtime": 38.9343, "eval_samples_per_second": 13.15, "eval_steps_per_second": 6.575, "eval_tts_loss": 7.925142382183957, "step": 942000 }, { "epoch": 5.680406652114663, "grad_norm": 3.659421443939209, "learning_rate": 2.5674341803325486e-05, "loss": 4.4701, "step": 942050 }, { "epoch": 5.680708143895997, "grad_norm": 2.3680219650268555, "learning_rate": 2.5662957923447862e-05, "loss": 4.6678, "step": 942100 }, { "epoch": 5.681009635677332, "grad_norm": 4.1008806228637695, "learning_rate": 2.5651576331816065e-05, "loss": 4.4737, "step": 942150 }, { "epoch": 5.681311127458666, "grad_norm": 5.577685832977295, "learning_rate": 2.564019702863959e-05, "loss": 4.0677, "step": 942200 }, { "epoch": 5.681612619239999, "grad_norm": 6.429434299468994, "learning_rate": 2.562882001412781e-05, "loss": 4.4506, "step": 942250 }, { "epoch": 5.681914111021333, "grad_norm": 6.4051103591918945, "learning_rate": 2.5617445288490084e-05, "loss": 4.3318, "step": 942300 }, { "epoch": 5.682215602802668, "grad_norm": 4.607983589172363, "learning_rate": 2.56060728519358e-05, "loss": 4.5789, "step": 942350 }, { "epoch": 5.682517094584002, "grad_norm": 2.430546998977661, "learning_rate": 2.5594702704674203e-05, "loss": 4.2051, "step": 942400 }, { "epoch": 5.682818586365336, "grad_norm": 3.521651029586792, "learning_rate": 2.5583334846914545e-05, "loss": 4.6459, "step": 942450 }, { "epoch": 5.68312007814667, "grad_norm": 5.300507545471191, "learning_rate": 2.5571969278866123e-05, "loss": 4.5976, "step": 942500 }, { "epoch": 5.6834215699280035, "grad_norm": 4.2264556884765625, "learning_rate": 2.556060600073797e-05, "loss": 4.4396, "step": 942550 }, { "epoch": 5.683723061709338, "grad_norm": 4.925900936126709, "learning_rate": 2.5549245012739337e-05, "loss": 4.2452, "step": 942600 }, { "epoch": 5.684024553490672, "grad_norm": 3.897351026535034, "learning_rate": 2.553788631507924e-05, "loss": 4.7111, "step": 942650 }, { "epoch": 5.684326045272006, "grad_norm": 3.84659743309021, "learning_rate": 2.5526529907966698e-05, "loss": 4.5487, "step": 942700 }, { "epoch": 5.6846275370533395, "grad_norm": 4.94359016418457, "learning_rate": 2.5515175791610735e-05, "loss": 4.5518, "step": 942750 }, { "epoch": 5.684929028834674, "grad_norm": 1.949009895324707, "learning_rate": 2.5503823966220343e-05, "loss": 4.5885, "step": 942800 }, { "epoch": 5.685230520616008, "grad_norm": 6.194468975067139, "learning_rate": 2.5492474432004383e-05, "loss": 4.2873, "step": 942850 }, { "epoch": 5.685532012397342, "grad_norm": 3.2027409076690674, "learning_rate": 2.548112718917174e-05, "loss": 4.4209, "step": 942900 }, { "epoch": 5.685833504178676, "grad_norm": 2.991481065750122, "learning_rate": 2.5469782237931296e-05, "loss": 4.1808, "step": 942950 }, { "epoch": 5.6861349959600105, "grad_norm": 3.904672861099243, "learning_rate": 2.5458439578491783e-05, "loss": 4.9406, "step": 943000 }, { "epoch": 5.686436487741344, "grad_norm": 5.471789360046387, "learning_rate": 2.5447099211061938e-05, "loss": 4.3715, "step": 943050 }, { "epoch": 5.686737979522678, "grad_norm": 3.6640467643737793, "learning_rate": 2.5435761135850523e-05, "loss": 4.2883, "step": 943100 }, { "epoch": 5.687039471304012, "grad_norm": 4.123707294464111, "learning_rate": 2.542442535306611e-05, "loss": 4.7916, "step": 943150 }, { "epoch": 5.687340963085346, "grad_norm": 3.0821757316589355, "learning_rate": 2.5413091862917374e-05, "loss": 4.6468, "step": 943200 }, { "epoch": 5.687642454866681, "grad_norm": 4.303806781768799, "learning_rate": 2.5401760665612935e-05, "loss": 4.6073, "step": 943250 }, { "epoch": 5.687943946648014, "grad_norm": 3.230987310409546, "learning_rate": 2.539043176136123e-05, "loss": 4.3256, "step": 943300 }, { "epoch": 5.688245438429348, "grad_norm": 1.5787867307662964, "learning_rate": 2.5379105150370804e-05, "loss": 4.4921, "step": 943350 }, { "epoch": 5.688546930210682, "grad_norm": 3.548128128051758, "learning_rate": 2.5367780832850143e-05, "loss": 4.3717, "step": 943400 }, { "epoch": 5.688848421992017, "grad_norm": 2.5056910514831543, "learning_rate": 2.5356458809007614e-05, "loss": 4.6038, "step": 943450 }, { "epoch": 5.689149913773351, "grad_norm": 2.700597047805786, "learning_rate": 2.5345139079051548e-05, "loss": 4.6925, "step": 943500 }, { "epoch": 5.689451405554685, "grad_norm": 3.62754225730896, "learning_rate": 2.5333821643190344e-05, "loss": 4.9167, "step": 943550 }, { "epoch": 5.689752897336018, "grad_norm": 3.74194598197937, "learning_rate": 2.5322506501632207e-05, "loss": 4.675, "step": 943600 }, { "epoch": 5.6900543891173525, "grad_norm": 3.1005866527557373, "learning_rate": 2.5311193654585415e-05, "loss": 4.1953, "step": 943650 }, { "epoch": 5.690355880898687, "grad_norm": 5.015761375427246, "learning_rate": 2.5299883102258207e-05, "loss": 4.4286, "step": 943700 }, { "epoch": 5.690657372680021, "grad_norm": 2.6833412647247314, "learning_rate": 2.5288574844858643e-05, "loss": 4.6462, "step": 943750 }, { "epoch": 5.690958864461355, "grad_norm": 4.643792629241943, "learning_rate": 2.527726888259493e-05, "loss": 4.8394, "step": 943800 }, { "epoch": 5.6912603562426884, "grad_norm": 3.574129343032837, "learning_rate": 2.52659652156751e-05, "loss": 4.3473, "step": 943850 }, { "epoch": 5.691561848024023, "grad_norm": 5.828909397125244, "learning_rate": 2.5254663844307134e-05, "loss": 4.6744, "step": 943900 }, { "epoch": 5.691863339805357, "grad_norm": 4.4760565757751465, "learning_rate": 2.5243364768699047e-05, "loss": 4.4704, "step": 943950 }, { "epoch": 5.692164831586691, "grad_norm": 3.7872867584228516, "learning_rate": 2.5232067989058846e-05, "loss": 4.3823, "step": 944000 }, { "epoch": 5.692466323368025, "grad_norm": 3.0731425285339355, "learning_rate": 2.5220773505594326e-05, "loss": 4.3294, "step": 944050 }, { "epoch": 5.6927678151493595, "grad_norm": 4.274466037750244, "learning_rate": 2.5209481318513426e-05, "loss": 4.2712, "step": 944100 }, { "epoch": 5.693069306930693, "grad_norm": 4.938294410705566, "learning_rate": 2.5198191428023957e-05, "loss": 4.5551, "step": 944150 }, { "epoch": 5.693370798712027, "grad_norm": 3.404205799102783, "learning_rate": 2.5186903834333673e-05, "loss": 4.7172, "step": 944200 }, { "epoch": 5.693672290493361, "grad_norm": 5.294496536254883, "learning_rate": 2.5175618537650273e-05, "loss": 4.345, "step": 944250 }, { "epoch": 5.693973782274695, "grad_norm": 5.308791160583496, "learning_rate": 2.5164335538181507e-05, "loss": 4.4206, "step": 944300 }, { "epoch": 5.694275274056029, "grad_norm": 1.5335959196090698, "learning_rate": 2.515305483613498e-05, "loss": 4.1763, "step": 944350 }, { "epoch": 5.694576765837363, "grad_norm": 1.9515459537506104, "learning_rate": 2.5141776431718302e-05, "loss": 4.2255, "step": 944400 }, { "epoch": 5.694878257618697, "grad_norm": 4.2310099601745605, "learning_rate": 2.513050032513908e-05, "loss": 4.4497, "step": 944450 }, { "epoch": 5.695179749400031, "grad_norm": 4.276172161102295, "learning_rate": 2.5119226516604764e-05, "loss": 4.2011, "step": 944500 }, { "epoch": 5.695481241181366, "grad_norm": 8.16921329498291, "learning_rate": 2.5107955006322865e-05, "loss": 4.6757, "step": 944550 }, { "epoch": 5.6957827329627, "grad_norm": 3.394514560699463, "learning_rate": 2.509668579450086e-05, "loss": 4.6039, "step": 944600 }, { "epoch": 5.696084224744034, "grad_norm": 4.617472171783447, "learning_rate": 2.5085418881346125e-05, "loss": 4.5339, "step": 944650 }, { "epoch": 5.696385716525367, "grad_norm": 4.104163646697998, "learning_rate": 2.5074154267065933e-05, "loss": 4.27, "step": 944700 }, { "epoch": 5.6966872083067015, "grad_norm": 3.3906567096710205, "learning_rate": 2.5062891951867714e-05, "loss": 4.5399, "step": 944750 }, { "epoch": 5.696988700088036, "grad_norm": 3.334766149520874, "learning_rate": 2.505163193595862e-05, "loss": 4.6917, "step": 944800 }, { "epoch": 5.69729019186937, "grad_norm": 3.1732895374298096, "learning_rate": 2.5040374219545923e-05, "loss": 4.2756, "step": 944850 }, { "epoch": 5.697591683650703, "grad_norm": 4.46401834487915, "learning_rate": 2.502911880283687e-05, "loss": 4.3968, "step": 944900 }, { "epoch": 5.697893175432037, "grad_norm": 4.070868015289307, "learning_rate": 2.501786568603849e-05, "loss": 4.5683, "step": 944950 }, { "epoch": 5.698194667213372, "grad_norm": 2.5854978561401367, "learning_rate": 2.5006614869357978e-05, "loss": 4.4717, "step": 945000 }, { "epoch": 5.698194667213372, "eval_loss": 4.900127410888672, "eval_runtime": 38.9716, "eval_samples_per_second": 13.138, "eval_steps_per_second": 6.569, "eval_tts_loss": 7.925825114564469, "step": 945000 }, { "epoch": 5.698496158994706, "grad_norm": 3.9945666790008545, "learning_rate": 2.4995366353002327e-05, "loss": 4.4725, "step": 945050 }, { "epoch": 5.69879765077604, "grad_norm": 3.5658915042877197, "learning_rate": 2.4984120137178543e-05, "loss": 4.479, "step": 945100 }, { "epoch": 5.699099142557374, "grad_norm": 4.494343280792236, "learning_rate": 2.4972876222093625e-05, "loss": 4.2797, "step": 945150 }, { "epoch": 5.699400634338708, "grad_norm": 4.469644546508789, "learning_rate": 2.496163460795452e-05, "loss": 4.5122, "step": 945200 }, { "epoch": 5.699702126120042, "grad_norm": 5.292475700378418, "learning_rate": 2.4950395294968074e-05, "loss": 4.4076, "step": 945250 }, { "epoch": 5.700003617901376, "grad_norm": 2.3061211109161377, "learning_rate": 2.4939158283341142e-05, "loss": 4.3111, "step": 945300 }, { "epoch": 5.70030510968271, "grad_norm": 4.229901313781738, "learning_rate": 2.492792357328055e-05, "loss": 4.7953, "step": 945350 }, { "epoch": 5.700606601464044, "grad_norm": 3.6378822326660156, "learning_rate": 2.4916691164993013e-05, "loss": 4.0571, "step": 945400 }, { "epoch": 5.700908093245378, "grad_norm": 4.492855548858643, "learning_rate": 2.4905461058685316e-05, "loss": 4.5339, "step": 945450 }, { "epoch": 5.701209585026712, "grad_norm": 4.756678104400635, "learning_rate": 2.4894233254564043e-05, "loss": 4.4906, "step": 945500 }, { "epoch": 5.701511076808046, "grad_norm": 4.661296844482422, "learning_rate": 2.48830077528359e-05, "loss": 4.7283, "step": 945550 }, { "epoch": 5.70181256858938, "grad_norm": 5.907868385314941, "learning_rate": 2.487178455370743e-05, "loss": 4.6138, "step": 945600 }, { "epoch": 5.7021140603707146, "grad_norm": 3.7288506031036377, "learning_rate": 2.4860563657385173e-05, "loss": 4.5894, "step": 945650 }, { "epoch": 5.702415552152049, "grad_norm": 5.235634803771973, "learning_rate": 2.4849345064075714e-05, "loss": 4.5019, "step": 945700 }, { "epoch": 5.702717043933382, "grad_norm": 1.9204355478286743, "learning_rate": 2.483812877398542e-05, "loss": 4.2062, "step": 945750 }, { "epoch": 5.703018535714716, "grad_norm": 3.4138002395629883, "learning_rate": 2.4826914787320797e-05, "loss": 4.5582, "step": 945800 }, { "epoch": 5.7033200274960505, "grad_norm": 4.812062740325928, "learning_rate": 2.4815703104288155e-05, "loss": 4.585, "step": 945850 }, { "epoch": 5.703621519277385, "grad_norm": 4.10906982421875, "learning_rate": 2.4804493725093817e-05, "loss": 4.1124, "step": 945900 }, { "epoch": 5.703923011058719, "grad_norm": 5.526692867279053, "learning_rate": 2.4793286649944115e-05, "loss": 4.7356, "step": 945950 }, { "epoch": 5.704224502840052, "grad_norm": 4.34367561340332, "learning_rate": 2.4782081879045323e-05, "loss": 4.6233, "step": 946000 }, { "epoch": 5.704525994621386, "grad_norm": 3.8064522743225098, "learning_rate": 2.477087941260359e-05, "loss": 4.3428, "step": 946050 }, { "epoch": 5.704827486402721, "grad_norm": 3.6789910793304443, "learning_rate": 2.4759679250825116e-05, "loss": 4.0451, "step": 946100 }, { "epoch": 5.705128978184055, "grad_norm": 3.1738717555999756, "learning_rate": 2.4748481393916037e-05, "loss": 4.3311, "step": 946150 }, { "epoch": 5.705430469965389, "grad_norm": 3.6912453174591064, "learning_rate": 2.4737285842082423e-05, "loss": 4.4517, "step": 946200 }, { "epoch": 5.705731961746723, "grad_norm": 3.8827333450317383, "learning_rate": 2.4726092595530273e-05, "loss": 4.4139, "step": 946250 }, { "epoch": 5.706033453528057, "grad_norm": 5.120059967041016, "learning_rate": 2.471490165446564e-05, "loss": 4.6768, "step": 946300 }, { "epoch": 5.706334945309391, "grad_norm": 3.9109766483306885, "learning_rate": 2.470371301909441e-05, "loss": 4.4878, "step": 946350 }, { "epoch": 5.706636437090725, "grad_norm": 1.8339207172393799, "learning_rate": 2.4692526689622556e-05, "loss": 4.5361, "step": 946400 }, { "epoch": 5.706937928872059, "grad_norm": 2.696744441986084, "learning_rate": 2.468134266625594e-05, "loss": 4.4488, "step": 946450 }, { "epoch": 5.7072394206533925, "grad_norm": 2.396772861480713, "learning_rate": 2.4670160949200347e-05, "loss": 4.6223, "step": 946500 }, { "epoch": 5.707540912434727, "grad_norm": 3.19270658493042, "learning_rate": 2.4658981538661614e-05, "loss": 4.4191, "step": 946550 }, { "epoch": 5.707842404216061, "grad_norm": 4.493042469024658, "learning_rate": 2.464780443484543e-05, "loss": 4.2187, "step": 946600 }, { "epoch": 5.708143895997395, "grad_norm": 3.3310468196868896, "learning_rate": 2.4636629637957545e-05, "loss": 4.2425, "step": 946650 }, { "epoch": 5.708445387778729, "grad_norm": 3.25711989402771, "learning_rate": 2.4625457148203558e-05, "loss": 4.1823, "step": 946700 }, { "epoch": 5.7087468795600635, "grad_norm": 2.6030499935150146, "learning_rate": 2.4614286965789142e-05, "loss": 4.3422, "step": 946750 }, { "epoch": 5.709048371341397, "grad_norm": 3.9321510791778564, "learning_rate": 2.4603119090919815e-05, "loss": 4.7804, "step": 946800 }, { "epoch": 5.709349863122731, "grad_norm": 3.606156826019287, "learning_rate": 2.459195352380111e-05, "loss": 4.1565, "step": 946850 }, { "epoch": 5.709651354904065, "grad_norm": 2.7812535762786865, "learning_rate": 2.4580790264638585e-05, "loss": 4.3015, "step": 946900 }, { "epoch": 5.7099528466853995, "grad_norm": 8.066612243652344, "learning_rate": 2.45696293136376e-05, "loss": 4.3297, "step": 946950 }, { "epoch": 5.710254338466734, "grad_norm": 3.2612032890319824, "learning_rate": 2.4558470671003617e-05, "loss": 4.3495, "step": 947000 }, { "epoch": 5.710555830248067, "grad_norm": 4.357885837554932, "learning_rate": 2.4547314336941964e-05, "loss": 4.5756, "step": 947050 }, { "epoch": 5.710857322029401, "grad_norm": 3.841374158859253, "learning_rate": 2.453616031165793e-05, "loss": 4.1275, "step": 947100 }, { "epoch": 5.711158813810735, "grad_norm": 5.285566329956055, "learning_rate": 2.452500859535682e-05, "loss": 4.7462, "step": 947150 }, { "epoch": 5.71146030559207, "grad_norm": 5.517605781555176, "learning_rate": 2.45138591882439e-05, "loss": 4.5583, "step": 947200 }, { "epoch": 5.711761797373404, "grad_norm": 4.054987907409668, "learning_rate": 2.4502712090524272e-05, "loss": 4.8458, "step": 947250 }, { "epoch": 5.712063289154738, "grad_norm": 8.96916675567627, "learning_rate": 2.449156730240316e-05, "loss": 4.3369, "step": 947300 }, { "epoch": 5.712364780936071, "grad_norm": 3.4909586906433105, "learning_rate": 2.448042482408566e-05, "loss": 4.5224, "step": 947350 }, { "epoch": 5.712666272717406, "grad_norm": 4.883467197418213, "learning_rate": 2.4469284655776827e-05, "loss": 4.6723, "step": 947400 }, { "epoch": 5.71296776449874, "grad_norm": 5.791742324829102, "learning_rate": 2.4458146797681617e-05, "loss": 4.4114, "step": 947450 }, { "epoch": 5.713269256280074, "grad_norm": 4.691702842712402, "learning_rate": 2.4447011250005093e-05, "loss": 4.02, "step": 947500 }, { "epoch": 5.713570748061408, "grad_norm": 3.5519602298736572, "learning_rate": 2.443587801295213e-05, "loss": 4.3418, "step": 947550 }, { "epoch": 5.7138722398427415, "grad_norm": 3.764474630355835, "learning_rate": 2.4424747086727648e-05, "loss": 4.4062, "step": 947600 }, { "epoch": 5.714173731624076, "grad_norm": 6.03237771987915, "learning_rate": 2.441361847153651e-05, "loss": 4.6619, "step": 947650 }, { "epoch": 5.71447522340541, "grad_norm": 4.787864685058594, "learning_rate": 2.4402492167583477e-05, "loss": 4.47, "step": 947700 }, { "epoch": 5.714776715186744, "grad_norm": 3.75272798538208, "learning_rate": 2.439136817507333e-05, "loss": 4.5759, "step": 947750 }, { "epoch": 5.715078206968078, "grad_norm": 3.480440616607666, "learning_rate": 2.4380246494210836e-05, "loss": 4.5623, "step": 947800 }, { "epoch": 5.7153796987494125, "grad_norm": 3.8512868881225586, "learning_rate": 2.436912712520064e-05, "loss": 4.4016, "step": 947850 }, { "epoch": 5.715681190530746, "grad_norm": 3.393981456756592, "learning_rate": 2.4358010068247335e-05, "loss": 4.5528, "step": 947900 }, { "epoch": 5.71598268231208, "grad_norm": 3.0649123191833496, "learning_rate": 2.43468953235556e-05, "loss": 4.1057, "step": 947950 }, { "epoch": 5.716284174093414, "grad_norm": 2.353224754333496, "learning_rate": 2.4335782891329896e-05, "loss": 4.3116, "step": 948000 }, { "epoch": 5.716284174093414, "eval_loss": 4.896012306213379, "eval_runtime": 38.9889, "eval_samples_per_second": 13.132, "eval_steps_per_second": 6.566, "eval_tts_loss": 7.926066337825444, "step": 948000 }, { "epoch": 5.7165856658747485, "grad_norm": 4.2121262550354, "learning_rate": 2.4324672771774766e-05, "loss": 4.4306, "step": 948050 }, { "epoch": 5.716887157656082, "grad_norm": 3.9814651012420654, "learning_rate": 2.431356496509473e-05, "loss": 4.4945, "step": 948100 }, { "epoch": 5.717188649437416, "grad_norm": 1.9849944114685059, "learning_rate": 2.4302459471494135e-05, "loss": 4.2627, "step": 948150 }, { "epoch": 5.71749014121875, "grad_norm": 3.3355116844177246, "learning_rate": 2.429135629117742e-05, "loss": 4.4274, "step": 948200 }, { "epoch": 5.717791633000084, "grad_norm": 6.202104091644287, "learning_rate": 2.428025542434891e-05, "loss": 4.5221, "step": 948250 }, { "epoch": 5.718093124781419, "grad_norm": 4.279924392700195, "learning_rate": 2.4269156871212835e-05, "loss": 4.3676, "step": 948300 }, { "epoch": 5.718394616562753, "grad_norm": 4.153377056121826, "learning_rate": 2.425806063197352e-05, "loss": 4.033, "step": 948350 }, { "epoch": 5.718696108344087, "grad_norm": 4.338130474090576, "learning_rate": 2.424696670683518e-05, "loss": 4.6894, "step": 948400 }, { "epoch": 5.71899760012542, "grad_norm": 4.586724281311035, "learning_rate": 2.4235875096001923e-05, "loss": 4.5529, "step": 948450 }, { "epoch": 5.719299091906755, "grad_norm": 2.4583797454833984, "learning_rate": 2.4224785799677898e-05, "loss": 4.2291, "step": 948500 }, { "epoch": 5.719600583688089, "grad_norm": 4.984464168548584, "learning_rate": 2.4213698818067263e-05, "loss": 4.375, "step": 948550 }, { "epoch": 5.719902075469423, "grad_norm": 3.307307720184326, "learning_rate": 2.4202614151373965e-05, "loss": 4.8999, "step": 948600 }, { "epoch": 5.720203567250756, "grad_norm": 3.579089641571045, "learning_rate": 2.4191531799802016e-05, "loss": 4.2469, "step": 948650 }, { "epoch": 5.7205050590320905, "grad_norm": 4.793903827667236, "learning_rate": 2.4180451763555413e-05, "loss": 4.3119, "step": 948700 }, { "epoch": 5.720806550813425, "grad_norm": 3.8614020347595215, "learning_rate": 2.4169374042838012e-05, "loss": 4.4063, "step": 948750 }, { "epoch": 5.721108042594759, "grad_norm": 3.274977684020996, "learning_rate": 2.4158298637853702e-05, "loss": 4.2118, "step": 948800 }, { "epoch": 5.721409534376093, "grad_norm": 3.731009006500244, "learning_rate": 2.414722554880633e-05, "loss": 4.4465, "step": 948850 }, { "epoch": 5.721711026157427, "grad_norm": 3.6447253227233887, "learning_rate": 2.413615477589969e-05, "loss": 4.4752, "step": 948900 }, { "epoch": 5.722012517938761, "grad_norm": 0.9255989193916321, "learning_rate": 2.4125086319337485e-05, "loss": 4.0705, "step": 948950 }, { "epoch": 5.722314009720095, "grad_norm": 2.5854949951171875, "learning_rate": 2.4114020179323463e-05, "loss": 4.5836, "step": 949000 }, { "epoch": 5.722615501501429, "grad_norm": 3.327789783477783, "learning_rate": 2.410295635606125e-05, "loss": 3.9179, "step": 949050 }, { "epoch": 5.722916993282763, "grad_norm": 5.057155132293701, "learning_rate": 2.409189484975443e-05, "loss": 4.9388, "step": 949100 }, { "epoch": 5.7232184850640975, "grad_norm": 3.2865517139434814, "learning_rate": 2.408083566060659e-05, "loss": 4.5417, "step": 949150 }, { "epoch": 5.723519976845431, "grad_norm": 2.3821375370025635, "learning_rate": 2.406977878882132e-05, "loss": 4.5685, "step": 949200 }, { "epoch": 5.723821468626765, "grad_norm": 3.289504289627075, "learning_rate": 2.4058724234602024e-05, "loss": 4.7194, "step": 949250 }, { "epoch": 5.724122960408099, "grad_norm": 6.018074989318848, "learning_rate": 2.4047671998152184e-05, "loss": 4.32, "step": 949300 }, { "epoch": 5.724424452189433, "grad_norm": 4.976550102233887, "learning_rate": 2.403662207967523e-05, "loss": 4.5606, "step": 949350 }, { "epoch": 5.724725943970768, "grad_norm": 5.240113258361816, "learning_rate": 2.4025574479374486e-05, "loss": 4.1193, "step": 949400 }, { "epoch": 5.725027435752102, "grad_norm": 3.580859422683716, "learning_rate": 2.4014529197453237e-05, "loss": 4.3967, "step": 949450 }, { "epoch": 5.725328927533435, "grad_norm": 3.8822903633117676, "learning_rate": 2.400348623411481e-05, "loss": 3.9223, "step": 949500 }, { "epoch": 5.725630419314769, "grad_norm": 4.705540180206299, "learning_rate": 2.3992445589562383e-05, "loss": 4.5646, "step": 949550 }, { "epoch": 5.7259319110961036, "grad_norm": 4.5010504722595215, "learning_rate": 2.3981407263999187e-05, "loss": 4.2169, "step": 949600 }, { "epoch": 5.726233402877438, "grad_norm": 4.601189136505127, "learning_rate": 2.397037125762836e-05, "loss": 4.1931, "step": 949650 }, { "epoch": 5.726534894658772, "grad_norm": 3.4035768508911133, "learning_rate": 2.3959337570652982e-05, "loss": 4.9254, "step": 949700 }, { "epoch": 5.726836386440105, "grad_norm": 3.7184295654296875, "learning_rate": 2.3948306203276163e-05, "loss": 4.2583, "step": 949750 }, { "epoch": 5.7271378782214395, "grad_norm": 5.081122875213623, "learning_rate": 2.3937277155700864e-05, "loss": 4.5023, "step": 949800 }, { "epoch": 5.727439370002774, "grad_norm": 4.914637088775635, "learning_rate": 2.392625042813005e-05, "loss": 4.3771, "step": 949850 }, { "epoch": 5.727740861784108, "grad_norm": 3.9718496799468994, "learning_rate": 2.39152260207667e-05, "loss": 4.3008, "step": 949900 }, { "epoch": 5.728042353565442, "grad_norm": 3.95338773727417, "learning_rate": 2.3904203933813694e-05, "loss": 4.459, "step": 949950 }, { "epoch": 5.728343845346776, "grad_norm": 5.531321048736572, "learning_rate": 2.389318416747384e-05, "loss": 4.9974, "step": 950000 }, { "epoch": 5.72864533712811, "grad_norm": 6.243918418884277, "learning_rate": 2.388216672194996e-05, "loss": 4.3249, "step": 950050 }, { "epoch": 5.728946828909444, "grad_norm": 4.154842853546143, "learning_rate": 2.387115159744487e-05, "loss": 4.6173, "step": 950100 }, { "epoch": 5.729248320690778, "grad_norm": 3.148599863052368, "learning_rate": 2.3860138794161198e-05, "loss": 4.4278, "step": 950150 }, { "epoch": 5.729549812472112, "grad_norm": 4.727330207824707, "learning_rate": 2.3849128312301696e-05, "loss": 4.3892, "step": 950200 }, { "epoch": 5.729851304253446, "grad_norm": 4.7020039558410645, "learning_rate": 2.3838120152068953e-05, "loss": 4.4324, "step": 950250 }, { "epoch": 5.73015279603478, "grad_norm": 4.970787525177002, "learning_rate": 2.3827114313665536e-05, "loss": 4.775, "step": 950300 }, { "epoch": 5.730454287816114, "grad_norm": 3.8948488235473633, "learning_rate": 2.381611079729402e-05, "loss": 4.3978, "step": 950350 }, { "epoch": 5.730755779597448, "grad_norm": 2.5401453971862793, "learning_rate": 2.3805109603156957e-05, "loss": 4.4001, "step": 950400 }, { "epoch": 5.731057271378782, "grad_norm": 4.921814441680908, "learning_rate": 2.3794110731456705e-05, "loss": 4.254, "step": 950450 }, { "epoch": 5.731358763160117, "grad_norm": 2.5671491622924805, "learning_rate": 2.3783114182395753e-05, "loss": 4.406, "step": 950500 }, { "epoch": 5.73166025494145, "grad_norm": 3.497046947479248, "learning_rate": 2.37721199561765e-05, "loss": 4.4484, "step": 950550 }, { "epoch": 5.731961746722784, "grad_norm": 3.1662871837615967, "learning_rate": 2.376112805300124e-05, "loss": 4.1482, "step": 950600 }, { "epoch": 5.732263238504118, "grad_norm": 1.8130433559417725, "learning_rate": 2.375013847307224e-05, "loss": 4.2927, "step": 950650 }, { "epoch": 5.7325647302854525, "grad_norm": 4.269565105438232, "learning_rate": 2.373915121659179e-05, "loss": 4.367, "step": 950700 }, { "epoch": 5.732866222066787, "grad_norm": 4.822585105895996, "learning_rate": 2.3728166283762046e-05, "loss": 4.5777, "step": 950750 }, { "epoch": 5.73316771384812, "grad_norm": 3.7296462059020996, "learning_rate": 2.371718367478521e-05, "loss": 4.629, "step": 950800 }, { "epoch": 5.733469205629454, "grad_norm": 3.405031681060791, "learning_rate": 2.3706203389863426e-05, "loss": 4.3009, "step": 950850 }, { "epoch": 5.7337706974107885, "grad_norm": 4.004791259765625, "learning_rate": 2.369522542919871e-05, "loss": 4.7737, "step": 950900 }, { "epoch": 5.734072189192123, "grad_norm": 3.065814733505249, "learning_rate": 2.3684249792993153e-05, "loss": 4.1601, "step": 950950 }, { "epoch": 5.734373680973457, "grad_norm": 1.8820744752883911, "learning_rate": 2.367327648144868e-05, "loss": 4.4133, "step": 951000 }, { "epoch": 5.734373680973457, "eval_loss": 4.893610954284668, "eval_runtime": 39.192, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 7.944883708725986, "step": 951000 }, { "epoch": 5.734675172754791, "grad_norm": 5.611132621765137, "learning_rate": 2.3662305494767302e-05, "loss": 4.3271, "step": 951050 }, { "epoch": 5.734976664536124, "grad_norm": 3.621784210205078, "learning_rate": 2.365133683315087e-05, "loss": 4.6031, "step": 951100 }, { "epoch": 5.735278156317459, "grad_norm": 2.4401886463165283, "learning_rate": 2.3640370496801313e-05, "loss": 4.243, "step": 951150 }, { "epoch": 5.735579648098793, "grad_norm": 4.533651351928711, "learning_rate": 2.3629406485920377e-05, "loss": 4.0879, "step": 951200 }, { "epoch": 5.735881139880127, "grad_norm": 5.514577865600586, "learning_rate": 2.361844480070986e-05, "loss": 4.5488, "step": 951250 }, { "epoch": 5.736182631661461, "grad_norm": 3.324179172515869, "learning_rate": 2.360748544137156e-05, "loss": 4.4057, "step": 951300 }, { "epoch": 5.736484123442795, "grad_norm": 3.9917330741882324, "learning_rate": 2.359652840810709e-05, "loss": 4.4574, "step": 951350 }, { "epoch": 5.736785615224129, "grad_norm": 3.776254415512085, "learning_rate": 2.3585573701118142e-05, "loss": 4.3597, "step": 951400 }, { "epoch": 5.737087107005463, "grad_norm": 3.601813554763794, "learning_rate": 2.3574621320606325e-05, "loss": 4.7367, "step": 951450 }, { "epoch": 5.737388598786797, "grad_norm": 6.578322887420654, "learning_rate": 2.356367126677314e-05, "loss": 4.5405, "step": 951500 }, { "epoch": 5.737690090568131, "grad_norm": 5.1435675621032715, "learning_rate": 2.3552723539820157e-05, "loss": 4.5957, "step": 951550 }, { "epoch": 5.737991582349466, "grad_norm": 1.479783058166504, "learning_rate": 2.354177813994887e-05, "loss": 4.4297, "step": 951600 }, { "epoch": 5.738293074130799, "grad_norm": 6.033353328704834, "learning_rate": 2.3530835067360665e-05, "loss": 4.3244, "step": 951650 }, { "epoch": 5.738594565912133, "grad_norm": 5.761683464050293, "learning_rate": 2.3519894322256966e-05, "loss": 4.8019, "step": 951700 }, { "epoch": 5.738896057693467, "grad_norm": 4.355960845947266, "learning_rate": 2.3508955904839126e-05, "loss": 4.5146, "step": 951750 }, { "epoch": 5.7391975494748015, "grad_norm": 6.360668182373047, "learning_rate": 2.3498019815308473e-05, "loss": 4.825, "step": 951800 }, { "epoch": 5.739499041256135, "grad_norm": 3.3011250495910645, "learning_rate": 2.3487086053866172e-05, "loss": 4.4365, "step": 951850 }, { "epoch": 5.739800533037469, "grad_norm": 3.3820433616638184, "learning_rate": 2.3476154620713565e-05, "loss": 4.365, "step": 951900 }, { "epoch": 5.740102024818803, "grad_norm": 2.8745651245117188, "learning_rate": 2.346522551605173e-05, "loss": 4.5872, "step": 951950 }, { "epoch": 5.7404035166001375, "grad_norm": 4.067492961883545, "learning_rate": 2.3454298740081846e-05, "loss": 4.6789, "step": 952000 }, { "epoch": 5.740705008381472, "grad_norm": 5.316725730895996, "learning_rate": 2.3443374293005023e-05, "loss": 4.7324, "step": 952050 }, { "epoch": 5.741006500162806, "grad_norm": 3.4130139350891113, "learning_rate": 2.343245217502227e-05, "loss": 4.5382, "step": 952100 }, { "epoch": 5.74130799194414, "grad_norm": 3.522702217102051, "learning_rate": 2.3421532386334624e-05, "loss": 4.2836, "step": 952150 }, { "epoch": 5.741609483725473, "grad_norm": 6.071360111236572, "learning_rate": 2.3410614927143058e-05, "loss": 4.6678, "step": 952200 }, { "epoch": 5.741910975506808, "grad_norm": 3.3174121379852295, "learning_rate": 2.3399699797648473e-05, "loss": 4.5625, "step": 952250 }, { "epoch": 5.742212467288142, "grad_norm": 3.98115611076355, "learning_rate": 2.3388786998051716e-05, "loss": 4.4715, "step": 952300 }, { "epoch": 5.742513959069476, "grad_norm": 3.116806983947754, "learning_rate": 2.337787652855364e-05, "loss": 4.6415, "step": 952350 }, { "epoch": 5.742815450850809, "grad_norm": 2.983785629272461, "learning_rate": 2.3366968389355083e-05, "loss": 4.4062, "step": 952400 }, { "epoch": 5.743116942632144, "grad_norm": 3.9594507217407227, "learning_rate": 2.3356062580656736e-05, "loss": 4.7847, "step": 952450 }, { "epoch": 5.743418434413478, "grad_norm": 5.425678253173828, "learning_rate": 2.3345159102659305e-05, "loss": 4.0747, "step": 952500 }, { "epoch": 5.743719926194812, "grad_norm": 5.346273899078369, "learning_rate": 2.333425795556353e-05, "loss": 4.0984, "step": 952550 }, { "epoch": 5.744021417976146, "grad_norm": 7.560357570648193, "learning_rate": 2.3323359139569965e-05, "loss": 4.7738, "step": 952600 }, { "epoch": 5.74432290975748, "grad_norm": 3.305506467819214, "learning_rate": 2.3312462654879166e-05, "loss": 4.8372, "step": 952650 }, { "epoch": 5.744624401538814, "grad_norm": 3.8906595706939697, "learning_rate": 2.3301568501691726e-05, "loss": 4.4307, "step": 952700 }, { "epoch": 5.744925893320148, "grad_norm": 4.484404563903809, "learning_rate": 2.3290676680208082e-05, "loss": 4.5678, "step": 952750 }, { "epoch": 5.745227385101482, "grad_norm": 4.247237205505371, "learning_rate": 2.3279787190628724e-05, "loss": 4.4629, "step": 952800 }, { "epoch": 5.745528876882816, "grad_norm": 4.782346725463867, "learning_rate": 2.326890003315406e-05, "loss": 4.4176, "step": 952850 }, { "epoch": 5.7458303686641505, "grad_norm": 9.571340560913086, "learning_rate": 2.3258015207984394e-05, "loss": 4.4711, "step": 952900 }, { "epoch": 5.746131860445484, "grad_norm": 5.11588191986084, "learning_rate": 2.324713271532013e-05, "loss": 3.9665, "step": 952950 }, { "epoch": 5.746433352226818, "grad_norm": 3.2861745357513428, "learning_rate": 2.3236252555361518e-05, "loss": 4.5969, "step": 953000 }, { "epoch": 5.746734844008152, "grad_norm": 4.424827575683594, "learning_rate": 2.3225374728308722e-05, "loss": 4.6186, "step": 953050 }, { "epoch": 5.7470363357894865, "grad_norm": 4.133995532989502, "learning_rate": 2.3214499234362e-05, "loss": 4.0995, "step": 953100 }, { "epoch": 5.747337827570821, "grad_norm": 5.458670616149902, "learning_rate": 2.320362607372151e-05, "loss": 4.1434, "step": 953150 }, { "epoch": 5.747639319352155, "grad_norm": 3.435723066329956, "learning_rate": 2.319275524658731e-05, "loss": 4.5541, "step": 953200 }, { "epoch": 5.747940811133488, "grad_norm": 7.6011576652526855, "learning_rate": 2.3181886753159485e-05, "loss": 3.972, "step": 953250 }, { "epoch": 5.748242302914822, "grad_norm": 4.216514587402344, "learning_rate": 2.3171020593638078e-05, "loss": 4.6385, "step": 953300 }, { "epoch": 5.748543794696157, "grad_norm": 4.449357509613037, "learning_rate": 2.3160156768223015e-05, "loss": 4.7006, "step": 953350 }, { "epoch": 5.748845286477491, "grad_norm": 3.9877426624298096, "learning_rate": 2.3149295277114298e-05, "loss": 4.2202, "step": 953400 }, { "epoch": 5.749146778258825, "grad_norm": 0.6151512265205383, "learning_rate": 2.3138436120511772e-05, "loss": 4.5237, "step": 953450 }, { "epoch": 5.749448270040158, "grad_norm": 3.1734845638275146, "learning_rate": 2.312757929861524e-05, "loss": 4.4802, "step": 953500 }, { "epoch": 5.7497497618214926, "grad_norm": 2.7978692054748535, "learning_rate": 2.3116724811624575e-05, "loss": 4.536, "step": 953550 }, { "epoch": 5.750051253602827, "grad_norm": 4.909815788269043, "learning_rate": 2.3105872659739538e-05, "loss": 4.3802, "step": 953600 }, { "epoch": 5.750352745384161, "grad_norm": 3.671715259552002, "learning_rate": 2.30950228431598e-05, "loss": 4.3254, "step": 953650 }, { "epoch": 5.750654237165495, "grad_norm": 3.3575549125671387, "learning_rate": 2.308417536208505e-05, "loss": 4.4762, "step": 953700 }, { "epoch": 5.750955728946829, "grad_norm": 5.529618740081787, "learning_rate": 2.307333021671498e-05, "loss": 4.5403, "step": 953750 }, { "epoch": 5.751257220728163, "grad_norm": 4.217447280883789, "learning_rate": 2.306248740724913e-05, "loss": 4.1343, "step": 953800 }, { "epoch": 5.751558712509497, "grad_norm": 1.7535816431045532, "learning_rate": 2.3051646933887025e-05, "loss": 4.7223, "step": 953850 }, { "epoch": 5.751860204290831, "grad_norm": 4.774880409240723, "learning_rate": 2.304080879682822e-05, "loss": 4.1611, "step": 953900 }, { "epoch": 5.752161696072165, "grad_norm": 3.665318012237549, "learning_rate": 2.3029972996272107e-05, "loss": 4.4193, "step": 953950 }, { "epoch": 5.752463187853499, "grad_norm": 2.922964334487915, "learning_rate": 2.3019139532418156e-05, "loss": 4.3017, "step": 954000 }, { "epoch": 5.752463187853499, "eval_loss": 4.8864312171936035, "eval_runtime": 38.8515, "eval_samples_per_second": 13.178, "eval_steps_per_second": 6.589, "eval_tts_loss": 7.930820156784777, "step": 954000 }, { "epoch": 5.752764679634833, "grad_norm": 4.233187198638916, "learning_rate": 2.3008308405465764e-05, "loss": 3.9575, "step": 954050 }, { "epoch": 5.753066171416167, "grad_norm": 4.190262794494629, "learning_rate": 2.2997479615614185e-05, "loss": 4.155, "step": 954100 }, { "epoch": 5.753367663197501, "grad_norm": 5.264037609100342, "learning_rate": 2.298665316306279e-05, "loss": 4.7686, "step": 954150 }, { "epoch": 5.7536691549788355, "grad_norm": 3.0360753536224365, "learning_rate": 2.297582904801079e-05, "loss": 4.3453, "step": 954200 }, { "epoch": 5.75397064676017, "grad_norm": 3.568105936050415, "learning_rate": 2.2965007270657344e-05, "loss": 4.2117, "step": 954250 }, { "epoch": 5.754272138541503, "grad_norm": 4.2682318687438965, "learning_rate": 2.2954187831201653e-05, "loss": 4.5195, "step": 954300 }, { "epoch": 5.754573630322837, "grad_norm": 4.608780384063721, "learning_rate": 2.294337072984288e-05, "loss": 4.4936, "step": 954350 }, { "epoch": 5.754875122104171, "grad_norm": 2.4614717960357666, "learning_rate": 2.293255596678e-05, "loss": 4.2121, "step": 954400 }, { "epoch": 5.755176613885506, "grad_norm": 3.8434243202209473, "learning_rate": 2.292174354221208e-05, "loss": 4.2238, "step": 954450 }, { "epoch": 5.75547810566684, "grad_norm": 2.493828535079956, "learning_rate": 2.2910933456338165e-05, "loss": 4.1602, "step": 954500 }, { "epoch": 5.755779597448173, "grad_norm": 4.951934337615967, "learning_rate": 2.290012570935713e-05, "loss": 4.2915, "step": 954550 }, { "epoch": 5.756081089229507, "grad_norm": 2.2671632766723633, "learning_rate": 2.2889320301467916e-05, "loss": 4.0846, "step": 954600 }, { "epoch": 5.7563825810108415, "grad_norm": 1.4026981592178345, "learning_rate": 2.2878517232869382e-05, "loss": 4.2811, "step": 954650 }, { "epoch": 5.756684072792176, "grad_norm": 3.610894203186035, "learning_rate": 2.286771650376028e-05, "loss": 4.3486, "step": 954700 }, { "epoch": 5.75698556457351, "grad_norm": 5.089946746826172, "learning_rate": 2.285691811433944e-05, "loss": 4.457, "step": 954750 }, { "epoch": 5.757287056354844, "grad_norm": 3.3314006328582764, "learning_rate": 2.28461220648056e-05, "loss": 4.1965, "step": 954800 }, { "epoch": 5.7575885481361775, "grad_norm": 3.175642251968384, "learning_rate": 2.28353283553574e-05, "loss": 4.9295, "step": 954850 }, { "epoch": 5.757890039917512, "grad_norm": 4.048881530761719, "learning_rate": 2.2824536986193505e-05, "loss": 4.064, "step": 954900 }, { "epoch": 5.758191531698846, "grad_norm": 4.794000625610352, "learning_rate": 2.2813747957512546e-05, "loss": 4.7922, "step": 954950 }, { "epoch": 5.75849302348018, "grad_norm": 3.7749929428100586, "learning_rate": 2.280296126951307e-05, "loss": 4.5066, "step": 955000 }, { "epoch": 5.758794515261514, "grad_norm": 2.1889679431915283, "learning_rate": 2.279217692239352e-05, "loss": 4.351, "step": 955050 }, { "epoch": 5.759096007042848, "grad_norm": 3.7850866317749023, "learning_rate": 2.2781394916352462e-05, "loss": 4.6168, "step": 955100 }, { "epoch": 5.759397498824182, "grad_norm": 3.4963202476501465, "learning_rate": 2.2770615251588233e-05, "loss": 4.6324, "step": 955150 }, { "epoch": 5.759698990605516, "grad_norm": 2.751697301864624, "learning_rate": 2.275983792829928e-05, "loss": 4.83, "step": 955200 }, { "epoch": 5.76000048238685, "grad_norm": 3.304877281188965, "learning_rate": 2.274906294668395e-05, "loss": 4.1264, "step": 955250 }, { "epoch": 5.7603019741681845, "grad_norm": 5.6017303466796875, "learning_rate": 2.27382903069405e-05, "loss": 4.4523, "step": 955300 }, { "epoch": 5.760603465949519, "grad_norm": 4.677584648132324, "learning_rate": 2.2727520009267218e-05, "loss": 4.8438, "step": 955350 }, { "epoch": 5.760904957730852, "grad_norm": 4.3699259757995605, "learning_rate": 2.2716752053862318e-05, "loss": 4.7078, "step": 955400 }, { "epoch": 5.761206449512186, "grad_norm": 4.195272922515869, "learning_rate": 2.2705986440923923e-05, "loss": 4.3266, "step": 955450 }, { "epoch": 5.76150794129352, "grad_norm": 3.5051960945129395, "learning_rate": 2.269522317065017e-05, "loss": 4.6914, "step": 955500 }, { "epoch": 5.761809433074855, "grad_norm": 8.867666244506836, "learning_rate": 2.2684462243239177e-05, "loss": 4.4461, "step": 955550 }, { "epoch": 5.762110924856188, "grad_norm": 2.488569498062134, "learning_rate": 2.267370365888899e-05, "loss": 4.6851, "step": 955600 }, { "epoch": 5.762412416637522, "grad_norm": 2.894864559173584, "learning_rate": 2.2662947417797563e-05, "loss": 4.1743, "step": 955650 }, { "epoch": 5.762713908418856, "grad_norm": 5.251177787780762, "learning_rate": 2.265219352016286e-05, "loss": 4.6312, "step": 955700 }, { "epoch": 5.7630154002001905, "grad_norm": 4.1009602546691895, "learning_rate": 2.2641441966182845e-05, "loss": 4.9503, "step": 955750 }, { "epoch": 5.763316891981525, "grad_norm": 5.463018417358398, "learning_rate": 2.2630692756055325e-05, "loss": 4.5479, "step": 955800 }, { "epoch": 5.763618383762859, "grad_norm": 2.032444715499878, "learning_rate": 2.2619945889978107e-05, "loss": 4.5272, "step": 955850 }, { "epoch": 5.763919875544193, "grad_norm": 4.645201206207275, "learning_rate": 2.2609201368149053e-05, "loss": 4.4899, "step": 955900 }, { "epoch": 5.7642213673255265, "grad_norm": 3.796358823776245, "learning_rate": 2.2598459190765804e-05, "loss": 4.2875, "step": 955950 }, { "epoch": 5.764522859106861, "grad_norm": 3.083192825317383, "learning_rate": 2.25877193580261e-05, "loss": 4.1134, "step": 956000 }, { "epoch": 5.764824350888195, "grad_norm": 4.675045013427734, "learning_rate": 2.2576981870127635e-05, "loss": 4.5185, "step": 956050 }, { "epoch": 5.765125842669529, "grad_norm": 2.3603456020355225, "learning_rate": 2.2566246727267933e-05, "loss": 4.4971, "step": 956100 }, { "epoch": 5.765427334450862, "grad_norm": 5.048830509185791, "learning_rate": 2.2555513929644653e-05, "loss": 4.8538, "step": 956150 }, { "epoch": 5.765728826232197, "grad_norm": 4.058323383331299, "learning_rate": 2.2544783477455252e-05, "loss": 4.5212, "step": 956200 }, { "epoch": 5.766030318013531, "grad_norm": 1.0662606954574585, "learning_rate": 2.2534055370897193e-05, "loss": 4.5032, "step": 956250 }, { "epoch": 5.766331809794865, "grad_norm": 4.294166088104248, "learning_rate": 2.2523329610167943e-05, "loss": 4.3373, "step": 956300 }, { "epoch": 5.766633301576199, "grad_norm": 3.960566520690918, "learning_rate": 2.2512606195464923e-05, "loss": 4.6536, "step": 956350 }, { "epoch": 5.7669347933575335, "grad_norm": 5.752710342407227, "learning_rate": 2.2501885126985415e-05, "loss": 4.5271, "step": 956400 }, { "epoch": 5.767236285138867, "grad_norm": 2.118769884109497, "learning_rate": 2.2491166404926763e-05, "loss": 4.2167, "step": 956450 }, { "epoch": 5.767537776920201, "grad_norm": 3.1239778995513916, "learning_rate": 2.248045002948628e-05, "loss": 4.1934, "step": 956500 }, { "epoch": 5.767839268701535, "grad_norm": 4.06709623336792, "learning_rate": 2.2469736000861105e-05, "loss": 4.1733, "step": 956550 }, { "epoch": 5.768140760482869, "grad_norm": 5.313866138458252, "learning_rate": 2.245902431924843e-05, "loss": 4.3716, "step": 956600 }, { "epoch": 5.768442252264204, "grad_norm": 4.138159275054932, "learning_rate": 2.2448314984845433e-05, "loss": 4.3857, "step": 956650 }, { "epoch": 5.768743744045537, "grad_norm": 5.065803050994873, "learning_rate": 2.2437607997849138e-05, "loss": 4.6127, "step": 956700 }, { "epoch": 5.769045235826871, "grad_norm": 2.73323655128479, "learning_rate": 2.242690335845662e-05, "loss": 4.5073, "step": 956750 }, { "epoch": 5.769346727608205, "grad_norm": 5.266098499298096, "learning_rate": 2.2416201066864927e-05, "loss": 4.2557, "step": 956800 }, { "epoch": 5.7696482193895395, "grad_norm": 4.956908702850342, "learning_rate": 2.240550112327094e-05, "loss": 3.9116, "step": 956850 }, { "epoch": 5.769949711170874, "grad_norm": 4.054678916931152, "learning_rate": 2.239480352787163e-05, "loss": 4.6031, "step": 956900 }, { "epoch": 5.770251202952208, "grad_norm": 3.4433228969573975, "learning_rate": 2.2384108280863884e-05, "loss": 4.7113, "step": 956950 }, { "epoch": 5.770552694733541, "grad_norm": 5.385453224182129, "learning_rate": 2.237341538244451e-05, "loss": 4.5606, "step": 957000 }, { "epoch": 5.770552694733541, "eval_loss": 4.889510631561279, "eval_runtime": 38.8477, "eval_samples_per_second": 13.18, "eval_steps_per_second": 6.59, "eval_tts_loss": 7.916560272490912, "step": 957000 }, { "epoch": 5.7708541865148755, "grad_norm": 5.252657413482666, "learning_rate": 2.236272483281027e-05, "loss": 3.9004, "step": 957050 }, { "epoch": 5.77115567829621, "grad_norm": 1.9429031610488892, "learning_rate": 2.2352036632157955e-05, "loss": 4.4552, "step": 957100 }, { "epoch": 5.771457170077544, "grad_norm": 1.903401494026184, "learning_rate": 2.234135078068421e-05, "loss": 4.2902, "step": 957150 }, { "epoch": 5.771758661858878, "grad_norm": 4.349494934082031, "learning_rate": 2.2330667278585724e-05, "loss": 4.8256, "step": 957200 }, { "epoch": 5.772060153640211, "grad_norm": 3.5811891555786133, "learning_rate": 2.231998612605914e-05, "loss": 4.5049, "step": 957250 }, { "epoch": 5.772361645421546, "grad_norm": 4.989658355712891, "learning_rate": 2.2309307323300972e-05, "loss": 4.4648, "step": 957300 }, { "epoch": 5.77266313720288, "grad_norm": 5.188292503356934, "learning_rate": 2.2298630870507827e-05, "loss": 4.5384, "step": 957350 }, { "epoch": 5.772964628984214, "grad_norm": 3.042372703552246, "learning_rate": 2.228795676787613e-05, "loss": 4.4205, "step": 957400 }, { "epoch": 5.773266120765548, "grad_norm": 2.935417652130127, "learning_rate": 2.2277285015602293e-05, "loss": 4.4547, "step": 957450 }, { "epoch": 5.773567612546882, "grad_norm": 3.5372960567474365, "learning_rate": 2.2266615613882753e-05, "loss": 4.3637, "step": 957500 }, { "epoch": 5.773869104328216, "grad_norm": 4.974874973297119, "learning_rate": 2.2255948562913912e-05, "loss": 4.4589, "step": 957550 }, { "epoch": 5.77417059610955, "grad_norm": 5.938009262084961, "learning_rate": 2.2245283862892005e-05, "loss": 4.8446, "step": 957600 }, { "epoch": 5.774472087890884, "grad_norm": 3.314620018005371, "learning_rate": 2.2234621514013328e-05, "loss": 4.3221, "step": 957650 }, { "epoch": 5.774773579672218, "grad_norm": 4.742870330810547, "learning_rate": 2.222396151647414e-05, "loss": 4.6846, "step": 957700 }, { "epoch": 5.775075071453552, "grad_norm": 3.851487636566162, "learning_rate": 2.2213303870470574e-05, "loss": 4.0962, "step": 957750 }, { "epoch": 5.775376563234886, "grad_norm": 5.603748798370361, "learning_rate": 2.2202648576198812e-05, "loss": 4.3924, "step": 957800 }, { "epoch": 5.77567805501622, "grad_norm": 5.233419895172119, "learning_rate": 2.2191995633854937e-05, "loss": 4.2247, "step": 957850 }, { "epoch": 5.775979546797554, "grad_norm": 5.437860488891602, "learning_rate": 2.218134504363494e-05, "loss": 4.5482, "step": 957900 }, { "epoch": 5.7762810385788885, "grad_norm": 7.515443801879883, "learning_rate": 2.21706968057349e-05, "loss": 4.505, "step": 957950 }, { "epoch": 5.776582530360223, "grad_norm": 3.4917330741882324, "learning_rate": 2.2160050920350775e-05, "loss": 4.489, "step": 958000 }, { "epoch": 5.776884022141556, "grad_norm": 5.9801130294799805, "learning_rate": 2.214940738767846e-05, "loss": 4.1357, "step": 958050 }, { "epoch": 5.77718551392289, "grad_norm": 2.1389567852020264, "learning_rate": 2.2138766207913845e-05, "loss": 4.2897, "step": 958100 }, { "epoch": 5.7774870057042245, "grad_norm": 6.116377353668213, "learning_rate": 2.2128127381252792e-05, "loss": 4.7296, "step": 958150 }, { "epoch": 5.777788497485559, "grad_norm": 3.9707510471343994, "learning_rate": 2.2117490907891078e-05, "loss": 4.6435, "step": 958200 }, { "epoch": 5.778089989266893, "grad_norm": 4.500662326812744, "learning_rate": 2.2106856788024414e-05, "loss": 4.0826, "step": 958250 }, { "epoch": 5.778391481048226, "grad_norm": 4.056954860687256, "learning_rate": 2.2096225021848573e-05, "loss": 4.3234, "step": 958300 }, { "epoch": 5.77869297282956, "grad_norm": 2.2653980255126953, "learning_rate": 2.2085595609559153e-05, "loss": 4.1747, "step": 958350 }, { "epoch": 5.778994464610895, "grad_norm": 4.345104217529297, "learning_rate": 2.2074968551351794e-05, "loss": 4.3904, "step": 958400 }, { "epoch": 5.779295956392229, "grad_norm": 3.74507737159729, "learning_rate": 2.206434384742211e-05, "loss": 4.466, "step": 958450 }, { "epoch": 5.779597448173563, "grad_norm": 5.046630859375, "learning_rate": 2.2053721497965577e-05, "loss": 4.306, "step": 958500 }, { "epoch": 5.779898939954897, "grad_norm": 5.400970935821533, "learning_rate": 2.2043101503177752e-05, "loss": 4.6276, "step": 958550 }, { "epoch": 5.7802004317362305, "grad_norm": 3.7722837924957275, "learning_rate": 2.203248386325403e-05, "loss": 4.6231, "step": 958600 }, { "epoch": 5.780501923517565, "grad_norm": 3.6375417709350586, "learning_rate": 2.202186857838979e-05, "loss": 4.3325, "step": 958650 }, { "epoch": 5.780803415298899, "grad_norm": 3.99568772315979, "learning_rate": 2.2011255648780425e-05, "loss": 4.0721, "step": 958700 }, { "epoch": 5.781104907080233, "grad_norm": 3.8950610160827637, "learning_rate": 2.2000645074621276e-05, "loss": 4.3305, "step": 958750 }, { "epoch": 5.781406398861567, "grad_norm": 3.0264291763305664, "learning_rate": 2.1990036856107573e-05, "loss": 4.0863, "step": 958800 }, { "epoch": 5.781707890642901, "grad_norm": 3.1137707233428955, "learning_rate": 2.1979430993434542e-05, "loss": 4.5765, "step": 958850 }, { "epoch": 5.782009382424235, "grad_norm": 3.4623069763183594, "learning_rate": 2.1968827486797396e-05, "loss": 4.3855, "step": 958900 }, { "epoch": 5.782310874205569, "grad_norm": 4.088412761688232, "learning_rate": 2.195822633639131e-05, "loss": 4.1671, "step": 958950 }, { "epoch": 5.782612365986903, "grad_norm": 7.205889701843262, "learning_rate": 2.1947627542411345e-05, "loss": 4.3325, "step": 959000 }, { "epoch": 5.7829138577682375, "grad_norm": 3.937803268432617, "learning_rate": 2.193703110505251e-05, "loss": 4.7606, "step": 959050 }, { "epoch": 5.783215349549572, "grad_norm": 3.2524800300598145, "learning_rate": 2.1926437024509903e-05, "loss": 4.5766, "step": 959100 }, { "epoch": 5.783516841330905, "grad_norm": 6.618635654449463, "learning_rate": 2.1915845300978414e-05, "loss": 4.5743, "step": 959150 }, { "epoch": 5.783818333112239, "grad_norm": 4.872941970825195, "learning_rate": 2.1905255934653005e-05, "loss": 3.8838, "step": 959200 }, { "epoch": 5.7841198248935735, "grad_norm": 4.979904651641846, "learning_rate": 2.1894668925728588e-05, "loss": 4.1182, "step": 959250 }, { "epoch": 5.784421316674908, "grad_norm": 3.2720773220062256, "learning_rate": 2.1884084274399954e-05, "loss": 4.3644, "step": 959300 }, { "epoch": 5.784722808456241, "grad_norm": 5.615810871124268, "learning_rate": 2.1873501980861936e-05, "loss": 4.6259, "step": 959350 }, { "epoch": 5.785024300237575, "grad_norm": 4.346898555755615, "learning_rate": 2.186292204530929e-05, "loss": 4.6189, "step": 959400 }, { "epoch": 5.785325792018909, "grad_norm": 3.891305446624756, "learning_rate": 2.1852344467936644e-05, "loss": 4.1942, "step": 959450 }, { "epoch": 5.785627283800244, "grad_norm": 4.16539192199707, "learning_rate": 2.1841769248938728e-05, "loss": 4.2456, "step": 959500 }, { "epoch": 5.785928775581578, "grad_norm": 4.563016891479492, "learning_rate": 2.1831196388510202e-05, "loss": 4.4963, "step": 959550 }, { "epoch": 5.786230267362912, "grad_norm": 1.367227554321289, "learning_rate": 2.182062588684556e-05, "loss": 4.0344, "step": 959600 }, { "epoch": 5.786531759144246, "grad_norm": 4.283793926239014, "learning_rate": 2.181005774413938e-05, "loss": 4.5872, "step": 959650 }, { "epoch": 5.7868332509255795, "grad_norm": 3.3906707763671875, "learning_rate": 2.1799491960586173e-05, "loss": 4.4163, "step": 959700 }, { "epoch": 5.787134742706914, "grad_norm": 4.791022300720215, "learning_rate": 2.178892853638037e-05, "loss": 4.3023, "step": 959750 }, { "epoch": 5.787436234488248, "grad_norm": 3.5491244792938232, "learning_rate": 2.177836747171634e-05, "loss": 4.0254, "step": 959800 }, { "epoch": 5.787737726269582, "grad_norm": 3.9088528156280518, "learning_rate": 2.1767808766788502e-05, "loss": 4.6611, "step": 959850 }, { "epoch": 5.7880392180509155, "grad_norm": 6.434267520904541, "learning_rate": 2.1757252421791116e-05, "loss": 4.6916, "step": 959900 }, { "epoch": 5.78834070983225, "grad_norm": 5.489491939544678, "learning_rate": 2.174669843691849e-05, "loss": 4.0414, "step": 959950 }, { "epoch": 5.788642201613584, "grad_norm": 3.410922050476074, "learning_rate": 2.173614681236489e-05, "loss": 4.3923, "step": 960000 }, { "epoch": 5.788642201613584, "eval_loss": 4.886587142944336, "eval_runtime": 39.0585, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.92512951981746, "step": 960000 }, { "epoch": 5.788943693394918, "grad_norm": 6.253307342529297, "learning_rate": 2.1725597548324426e-05, "loss": 3.8984, "step": 960050 }, { "epoch": 5.789245185176252, "grad_norm": 3.7649338245391846, "learning_rate": 2.171505064499127e-05, "loss": 4.2844, "step": 960100 }, { "epoch": 5.7895466769575865, "grad_norm": 1.1891745328903198, "learning_rate": 2.1704506102559595e-05, "loss": 4.2237, "step": 960150 }, { "epoch": 5.78984816873892, "grad_norm": 5.244391441345215, "learning_rate": 2.1693963921223385e-05, "loss": 4.5279, "step": 960200 }, { "epoch": 5.790149660520254, "grad_norm": 7.0557942390441895, "learning_rate": 2.168342410117662e-05, "loss": 4.1183, "step": 960250 }, { "epoch": 5.790451152301588, "grad_norm": 2.217604160308838, "learning_rate": 2.167288664261337e-05, "loss": 4.5806, "step": 960300 }, { "epoch": 5.7907526440829225, "grad_norm": 4.078126430511475, "learning_rate": 2.166235154572745e-05, "loss": 4.283, "step": 960350 }, { "epoch": 5.791054135864257, "grad_norm": 5.0141706466674805, "learning_rate": 2.165181881071282e-05, "loss": 4.5644, "step": 960400 }, { "epoch": 5.79135562764559, "grad_norm": 4.97548770904541, "learning_rate": 2.1641288437763344e-05, "loss": 4.3763, "step": 960450 }, { "epoch": 5.791657119426924, "grad_norm": 5.002674102783203, "learning_rate": 2.1630760427072725e-05, "loss": 4.5116, "step": 960500 }, { "epoch": 5.791958611208258, "grad_norm": 5.161297798156738, "learning_rate": 2.1620234778834804e-05, "loss": 4.4859, "step": 960550 }, { "epoch": 5.792260102989593, "grad_norm": 3.7698473930358887, "learning_rate": 2.1609711493243265e-05, "loss": 4.3476, "step": 960600 }, { "epoch": 5.792561594770927, "grad_norm": 5.513879776000977, "learning_rate": 2.159919057049173e-05, "loss": 4.1056, "step": 960650 }, { "epoch": 5.792863086552261, "grad_norm": 5.70977258682251, "learning_rate": 2.158867201077385e-05, "loss": 4.5042, "step": 960700 }, { "epoch": 5.793164578333594, "grad_norm": 4.397405624389648, "learning_rate": 2.157815581428323e-05, "loss": 4.0999, "step": 960750 }, { "epoch": 5.7934660701149285, "grad_norm": 6.053720474243164, "learning_rate": 2.1567641981213356e-05, "loss": 4.4732, "step": 960800 }, { "epoch": 5.793767561896263, "grad_norm": 3.3044657707214355, "learning_rate": 2.155713051175777e-05, "loss": 4.4277, "step": 960850 }, { "epoch": 5.794069053677597, "grad_norm": 8.242375373840332, "learning_rate": 2.1546621406109915e-05, "loss": 4.5223, "step": 960900 }, { "epoch": 5.794370545458931, "grad_norm": 3.678574562072754, "learning_rate": 2.153611466446317e-05, "loss": 4.4487, "step": 960950 }, { "epoch": 5.7946720372402645, "grad_norm": 3.6005828380584717, "learning_rate": 2.1525610287010887e-05, "loss": 4.2708, "step": 961000 }, { "epoch": 5.794973529021599, "grad_norm": 4.446502208709717, "learning_rate": 2.1515108273946436e-05, "loss": 4.3799, "step": 961050 }, { "epoch": 5.795275020802933, "grad_norm": 4.2114481925964355, "learning_rate": 2.1504608625463032e-05, "loss": 4.5524, "step": 961100 }, { "epoch": 5.795576512584267, "grad_norm": 1.3192583322525024, "learning_rate": 2.1494111341753923e-05, "loss": 4.6224, "step": 961150 }, { "epoch": 5.795878004365601, "grad_norm": 4.996825218200684, "learning_rate": 2.148361642301235e-05, "loss": 4.6866, "step": 961200 }, { "epoch": 5.7961794961469355, "grad_norm": 5.87127685546875, "learning_rate": 2.1473123869431364e-05, "loss": 4.6124, "step": 961250 }, { "epoch": 5.796480987928269, "grad_norm": 4.486720085144043, "learning_rate": 2.1462633681204137e-05, "loss": 4.6536, "step": 961300 }, { "epoch": 5.796782479709603, "grad_norm": 3.332125663757324, "learning_rate": 2.1452145858523722e-05, "loss": 4.4708, "step": 961350 }, { "epoch": 5.797083971490937, "grad_norm": 3.208695888519287, "learning_rate": 2.1441660401583127e-05, "loss": 4.6423, "step": 961400 }, { "epoch": 5.7973854632722714, "grad_norm": 4.634199619293213, "learning_rate": 2.1431177310575265e-05, "loss": 4.4509, "step": 961450 }, { "epoch": 5.797686955053605, "grad_norm": 4.1108012199401855, "learning_rate": 2.1420696585693148e-05, "loss": 4.7241, "step": 961500 }, { "epoch": 5.797988446834939, "grad_norm": 1.8728656768798828, "learning_rate": 2.141021822712957e-05, "loss": 4.6586, "step": 961550 }, { "epoch": 5.798289938616273, "grad_norm": 4.267939567565918, "learning_rate": 2.1399742235077433e-05, "loss": 4.7295, "step": 961600 }, { "epoch": 5.798591430397607, "grad_norm": 6.239266872406006, "learning_rate": 2.1389268609729543e-05, "loss": 4.3623, "step": 961650 }, { "epoch": 5.798892922178942, "grad_norm": 5.797143936157227, "learning_rate": 2.1378797351278598e-05, "loss": 4.4085, "step": 961700 }, { "epoch": 5.799194413960276, "grad_norm": 7.462108612060547, "learning_rate": 2.136832845991738e-05, "loss": 4.2791, "step": 961750 }, { "epoch": 5.799495905741609, "grad_norm": 1.3323783874511719, "learning_rate": 2.1357861935838495e-05, "loss": 4.3785, "step": 961800 }, { "epoch": 5.799797397522943, "grad_norm": 4.40871000289917, "learning_rate": 2.1347397779234543e-05, "loss": 4.6636, "step": 961850 }, { "epoch": 5.8000988893042775, "grad_norm": 2.492300033569336, "learning_rate": 2.1336935990298153e-05, "loss": 4.1407, "step": 961900 }, { "epoch": 5.800400381085612, "grad_norm": 3.4911558628082275, "learning_rate": 2.1326476569221856e-05, "loss": 4.657, "step": 961950 }, { "epoch": 5.800701872866946, "grad_norm": 3.77150297164917, "learning_rate": 2.1316019516198113e-05, "loss": 4.4216, "step": 962000 }, { "epoch": 5.801003364648279, "grad_norm": 3.419347047805786, "learning_rate": 2.1305564831419385e-05, "loss": 4.2037, "step": 962050 }, { "epoch": 5.8013048564296135, "grad_norm": 4.282702922821045, "learning_rate": 2.129511251507812e-05, "loss": 4.6134, "step": 962100 }, { "epoch": 5.801606348210948, "grad_norm": 3.4073188304901123, "learning_rate": 2.1284662567366635e-05, "loss": 4.5092, "step": 962150 }, { "epoch": 5.801907839992282, "grad_norm": 2.3904666900634766, "learning_rate": 2.12742149884772e-05, "loss": 4.2529, "step": 962200 }, { "epoch": 5.802209331773616, "grad_norm": 5.329861640930176, "learning_rate": 2.1263769778602185e-05, "loss": 4.4654, "step": 962250 }, { "epoch": 5.80251082355495, "grad_norm": 3.3452701568603516, "learning_rate": 2.1253326937933734e-05, "loss": 4.4612, "step": 962300 }, { "epoch": 5.802812315336284, "grad_norm": 3.850947380065918, "learning_rate": 2.1242886466664062e-05, "loss": 4.3256, "step": 962350 }, { "epoch": 5.803113807117618, "grad_norm": 3.2977349758148193, "learning_rate": 2.1232448364985326e-05, "loss": 4.3502, "step": 962400 }, { "epoch": 5.803415298898952, "grad_norm": 3.2544105052948, "learning_rate": 2.1222012633089647e-05, "loss": 4.4762, "step": 962450 }, { "epoch": 5.803716790680286, "grad_norm": 3.621178388595581, "learning_rate": 2.1211579271169017e-05, "loss": 4.1279, "step": 962500 }, { "epoch": 5.80401828246162, "grad_norm": 6.000004291534424, "learning_rate": 2.1201148279415513e-05, "loss": 4.8211, "step": 962550 }, { "epoch": 5.804319774242954, "grad_norm": 5.944278717041016, "learning_rate": 2.119071965802107e-05, "loss": 4.7271, "step": 962600 }, { "epoch": 5.804621266024288, "grad_norm": 4.428544521331787, "learning_rate": 2.1180293407177583e-05, "loss": 4.6783, "step": 962650 }, { "epoch": 5.804922757805622, "grad_norm": 3.2917802333831787, "learning_rate": 2.116986952707695e-05, "loss": 4.3645, "step": 962700 }, { "epoch": 5.805224249586956, "grad_norm": 4.2207441329956055, "learning_rate": 2.115944801791106e-05, "loss": 3.9946, "step": 962750 }, { "epoch": 5.805525741368291, "grad_norm": 3.3557660579681396, "learning_rate": 2.114902887987164e-05, "loss": 4.2763, "step": 962800 }, { "epoch": 5.805827233149625, "grad_norm": 1.7113628387451172, "learning_rate": 2.113861211315044e-05, "loss": 4.494, "step": 962850 }, { "epoch": 5.806128724930958, "grad_norm": 3.4612553119659424, "learning_rate": 2.1128197717939228e-05, "loss": 4.6188, "step": 962900 }, { "epoch": 5.806430216712292, "grad_norm": 4.226151943206787, "learning_rate": 2.1117785694429633e-05, "loss": 4.3274, "step": 962950 }, { "epoch": 5.8067317084936265, "grad_norm": 6.452605247497559, "learning_rate": 2.110737604281324e-05, "loss": 4.1998, "step": 963000 }, { "epoch": 5.8067317084936265, "eval_loss": 4.876678943634033, "eval_runtime": 38.9301, "eval_samples_per_second": 13.152, "eval_steps_per_second": 6.576, "eval_tts_loss": 7.973017873315209, "step": 963000 }, { "epoch": 5.807033200274961, "grad_norm": 5.097949981689453, "learning_rate": 2.1096968763281674e-05, "loss": 4.5459, "step": 963050 }, { "epoch": 5.807334692056295, "grad_norm": 3.4562795162200928, "learning_rate": 2.1086563856026416e-05, "loss": 4.1781, "step": 963100 }, { "epoch": 5.807636183837628, "grad_norm": 5.230889320373535, "learning_rate": 2.1076161321238966e-05, "loss": 4.4322, "step": 963150 }, { "epoch": 5.8079376756189625, "grad_norm": 1.7735942602157593, "learning_rate": 2.1065761159110834e-05, "loss": 3.9753, "step": 963200 }, { "epoch": 5.808239167400297, "grad_norm": 3.568840265274048, "learning_rate": 2.1055363369833335e-05, "loss": 4.0423, "step": 963250 }, { "epoch": 5.808540659181631, "grad_norm": 4.857228755950928, "learning_rate": 2.104496795359788e-05, "loss": 4.2096, "step": 963300 }, { "epoch": 5.808842150962965, "grad_norm": 4.261373043060303, "learning_rate": 2.1034574910595735e-05, "loss": 4.4174, "step": 963350 }, { "epoch": 5.809143642744299, "grad_norm": 2.5405309200286865, "learning_rate": 2.102418424101823e-05, "loss": 4.7473, "step": 963400 }, { "epoch": 5.809445134525633, "grad_norm": 5.243120193481445, "learning_rate": 2.1013795945056518e-05, "loss": 3.948, "step": 963450 }, { "epoch": 5.809746626306967, "grad_norm": 2.3232884407043457, "learning_rate": 2.1003410022901867e-05, "loss": 4.2894, "step": 963500 }, { "epoch": 5.810048118088301, "grad_norm": 3.811640501022339, "learning_rate": 2.0993026474745316e-05, "loss": 4.5534, "step": 963550 }, { "epoch": 5.810349609869635, "grad_norm": 5.986536502838135, "learning_rate": 2.0982645300778013e-05, "loss": 4.3905, "step": 963600 }, { "epoch": 5.8106511016509685, "grad_norm": 5.741425514221191, "learning_rate": 2.0972266501191048e-05, "loss": 4.438, "step": 963650 }, { "epoch": 5.810952593432303, "grad_norm": 4.337397575378418, "learning_rate": 2.0961890076175343e-05, "loss": 3.7777, "step": 963700 }, { "epoch": 5.811254085213637, "grad_norm": 3.551408290863037, "learning_rate": 2.0951516025921937e-05, "loss": 4.4254, "step": 963750 }, { "epoch": 5.811555576994971, "grad_norm": 4.785838603973389, "learning_rate": 2.0941144350621718e-05, "loss": 4.7948, "step": 963800 }, { "epoch": 5.811857068776305, "grad_norm": 3.7397804260253906, "learning_rate": 2.0930775050465533e-05, "loss": 4.3056, "step": 963850 }, { "epoch": 5.81215856055764, "grad_norm": 3.68184232711792, "learning_rate": 2.092040812564422e-05, "loss": 4.1469, "step": 963900 }, { "epoch": 5.812460052338973, "grad_norm": 4.3986496925354, "learning_rate": 2.091004357634864e-05, "loss": 4.5598, "step": 963950 }, { "epoch": 5.812761544120307, "grad_norm": 1.1889783143997192, "learning_rate": 2.0899681402769463e-05, "loss": 4.1212, "step": 964000 }, { "epoch": 5.813063035901641, "grad_norm": 6.325162887573242, "learning_rate": 2.088932160509739e-05, "loss": 4.644, "step": 964050 }, { "epoch": 5.8133645276829755, "grad_norm": 5.15403413772583, "learning_rate": 2.0878964183523155e-05, "loss": 4.3425, "step": 964100 }, { "epoch": 5.81366601946431, "grad_norm": 4.267897129058838, "learning_rate": 2.086860913823732e-05, "loss": 4.2373, "step": 964150 }, { "epoch": 5.813967511245643, "grad_norm": 3.0163702964782715, "learning_rate": 2.0858256469430395e-05, "loss": 4.2484, "step": 964200 }, { "epoch": 5.814269003026977, "grad_norm": 4.733093738555908, "learning_rate": 2.0847906177293016e-05, "loss": 4.6299, "step": 964250 }, { "epoch": 5.8145704948083115, "grad_norm": 3.007228136062622, "learning_rate": 2.0837558262015582e-05, "loss": 4.0982, "step": 964300 }, { "epoch": 5.814871986589646, "grad_norm": 3.3531265258789062, "learning_rate": 2.0827212723788568e-05, "loss": 4.3311, "step": 964350 }, { "epoch": 5.81517347837098, "grad_norm": 4.750550270080566, "learning_rate": 2.081686956280239e-05, "loss": 4.3578, "step": 964400 }, { "epoch": 5.815474970152314, "grad_norm": 4.661322116851807, "learning_rate": 2.0806528779247335e-05, "loss": 4.4368, "step": 964450 }, { "epoch": 5.815776461933647, "grad_norm": 3.441621780395508, "learning_rate": 2.0796190373313758e-05, "loss": 4.4372, "step": 964500 }, { "epoch": 5.816077953714982, "grad_norm": 4.728804588317871, "learning_rate": 2.0785854345191948e-05, "loss": 4.0756, "step": 964550 }, { "epoch": 5.816379445496316, "grad_norm": 3.579451322555542, "learning_rate": 2.0775520695072085e-05, "loss": 4.4568, "step": 964600 }, { "epoch": 5.81668093727765, "grad_norm": 4.325102806091309, "learning_rate": 2.0765189423144313e-05, "loss": 4.4709, "step": 964650 }, { "epoch": 5.816982429058984, "grad_norm": 6.22654914855957, "learning_rate": 2.0754860529598833e-05, "loss": 4.6333, "step": 964700 }, { "epoch": 5.8172839208403175, "grad_norm": 3.97688889503479, "learning_rate": 2.0744534014625675e-05, "loss": 4.4689, "step": 964750 }, { "epoch": 5.817585412621652, "grad_norm": 3.6184093952178955, "learning_rate": 2.0734209878414922e-05, "loss": 4.4196, "step": 964800 }, { "epoch": 5.817886904402986, "grad_norm": 3.8073363304138184, "learning_rate": 2.0723888121156585e-05, "loss": 4.0846, "step": 964850 }, { "epoch": 5.81818839618432, "grad_norm": 3.780594825744629, "learning_rate": 2.0713568743040564e-05, "loss": 4.3363, "step": 964900 }, { "epoch": 5.818489887965654, "grad_norm": 3.972905397415161, "learning_rate": 2.0703251744256844e-05, "loss": 4.2602, "step": 964950 }, { "epoch": 5.818791379746989, "grad_norm": 4.898406505584717, "learning_rate": 2.0692937124995267e-05, "loss": 4.5399, "step": 965000 }, { "epoch": 5.819092871528322, "grad_norm": 4.309586048126221, "learning_rate": 2.0682624885445603e-05, "loss": 4.3503, "step": 965050 }, { "epoch": 5.819394363309656, "grad_norm": 3.6636321544647217, "learning_rate": 2.06723150257977e-05, "loss": 4.3823, "step": 965100 }, { "epoch": 5.81969585509099, "grad_norm": 1.094915747642517, "learning_rate": 2.06620075462413e-05, "loss": 4.2611, "step": 965150 }, { "epoch": 5.8199973468723245, "grad_norm": 6.171432018280029, "learning_rate": 2.0651702446966038e-05, "loss": 4.8184, "step": 965200 }, { "epoch": 5.820298838653658, "grad_norm": 4.283547878265381, "learning_rate": 2.0641399728161617e-05, "loss": 4.4781, "step": 965250 }, { "epoch": 5.820600330434992, "grad_norm": 3.4053828716278076, "learning_rate": 2.0631099390017663e-05, "loss": 4.2303, "step": 965300 }, { "epoch": 5.820901822216326, "grad_norm": 5.330726146697998, "learning_rate": 2.0620801432723695e-05, "loss": 4.3802, "step": 965350 }, { "epoch": 5.8212033139976604, "grad_norm": 4.828314781188965, "learning_rate": 2.0610505856469205e-05, "loss": 4.3872, "step": 965400 }, { "epoch": 5.821504805778995, "grad_norm": 3.5170836448669434, "learning_rate": 2.060021266144375e-05, "loss": 4.2218, "step": 965450 }, { "epoch": 5.821806297560329, "grad_norm": 4.3143744468688965, "learning_rate": 2.0589921847836688e-05, "loss": 4.4323, "step": 965500 }, { "epoch": 5.822107789341662, "grad_norm": 4.412014484405518, "learning_rate": 2.057963341583742e-05, "loss": 4.2978, "step": 965550 }, { "epoch": 5.822409281122996, "grad_norm": 4.050715923309326, "learning_rate": 2.0569347365635363e-05, "loss": 4.3963, "step": 965600 }, { "epoch": 5.822710772904331, "grad_norm": 3.6566715240478516, "learning_rate": 2.0559063697419715e-05, "loss": 4.2901, "step": 965650 }, { "epoch": 5.823012264685665, "grad_norm": 2.334287643432617, "learning_rate": 2.0548782411379772e-05, "loss": 4.2382, "step": 965700 }, { "epoch": 5.823313756466999, "grad_norm": 4.736079692840576, "learning_rate": 2.0538503507704784e-05, "loss": 4.5948, "step": 965750 }, { "epoch": 5.823615248248332, "grad_norm": 4.18605899810791, "learning_rate": 2.05282269865839e-05, "loss": 4.3918, "step": 965800 }, { "epoch": 5.8239167400296665, "grad_norm": 7.748193740844727, "learning_rate": 2.0517952848206187e-05, "loss": 4.4097, "step": 965850 }, { "epoch": 5.824218231811001, "grad_norm": 2.7470149993896484, "learning_rate": 2.0507681092760775e-05, "loss": 4.2186, "step": 965900 }, { "epoch": 5.824519723592335, "grad_norm": 2.1886894702911377, "learning_rate": 2.049741172043673e-05, "loss": 4.5966, "step": 965950 }, { "epoch": 5.824821215373669, "grad_norm": 5.896630764007568, "learning_rate": 2.048714473142298e-05, "loss": 4.3933, "step": 966000 }, { "epoch": 5.824821215373669, "eval_loss": 4.878879547119141, "eval_runtime": 39.1512, "eval_samples_per_second": 13.078, "eval_steps_per_second": 6.539, "eval_tts_loss": 7.925484678881063, "step": 966000 }, { "epoch": 5.825122707155003, "grad_norm": 5.3063225746154785, "learning_rate": 2.0476880125908502e-05, "loss": 4.5268, "step": 966050 }, { "epoch": 5.825424198936337, "grad_norm": 5.849755764007568, "learning_rate": 2.046661790408225e-05, "loss": 4.3971, "step": 966100 }, { "epoch": 5.825725690717671, "grad_norm": 4.942874908447266, "learning_rate": 2.0456358066133027e-05, "loss": 4.1473, "step": 966150 }, { "epoch": 5.826027182499005, "grad_norm": 3.8850481510162354, "learning_rate": 2.044610061224965e-05, "loss": 4.194, "step": 966200 }, { "epoch": 5.826328674280339, "grad_norm": 4.882343769073486, "learning_rate": 2.0435845542620915e-05, "loss": 4.2683, "step": 966250 }, { "epoch": 5.8266301660616735, "grad_norm": 7.64434814453125, "learning_rate": 2.042559285743554e-05, "loss": 4.5362, "step": 966300 }, { "epoch": 5.826931657843007, "grad_norm": 5.707789897918701, "learning_rate": 2.0415342556882193e-05, "loss": 4.4713, "step": 966350 }, { "epoch": 5.827233149624341, "grad_norm": 3.591036558151245, "learning_rate": 2.0405094641149584e-05, "loss": 4.2085, "step": 966400 }, { "epoch": 5.827534641405675, "grad_norm": 4.886037826538086, "learning_rate": 2.0394849110426232e-05, "loss": 4.7856, "step": 966450 }, { "epoch": 5.827836133187009, "grad_norm": 4.523858070373535, "learning_rate": 2.038460596490077e-05, "loss": 4.5684, "step": 966500 }, { "epoch": 5.828137624968344, "grad_norm": 5.178028106689453, "learning_rate": 2.0374365204761643e-05, "loss": 4.4625, "step": 966550 }, { "epoch": 5.828439116749678, "grad_norm": 1.5405473709106445, "learning_rate": 2.036412683019731e-05, "loss": 4.2157, "step": 966600 }, { "epoch": 5.828740608531011, "grad_norm": 5.135117530822754, "learning_rate": 2.0353890841396225e-05, "loss": 4.7401, "step": 966650 }, { "epoch": 5.829042100312345, "grad_norm": 3.744114398956299, "learning_rate": 2.034365723854681e-05, "loss": 4.4529, "step": 966700 }, { "epoch": 5.82934359209368, "grad_norm": 6.33007287979126, "learning_rate": 2.03334260218373e-05, "loss": 4.6595, "step": 966750 }, { "epoch": 5.829645083875014, "grad_norm": 3.909332275390625, "learning_rate": 2.0323197191456057e-05, "loss": 4.6021, "step": 966800 }, { "epoch": 5.829946575656348, "grad_norm": 3.9600484371185303, "learning_rate": 2.0312970747591335e-05, "loss": 4.0061, "step": 966850 }, { "epoch": 5.830248067437681, "grad_norm": 3.7077274322509766, "learning_rate": 2.0302746690431277e-05, "loss": 4.5784, "step": 966900 }, { "epoch": 5.8305495592190155, "grad_norm": 6.939572811126709, "learning_rate": 2.029252502016412e-05, "loss": 4.8163, "step": 966950 }, { "epoch": 5.83085105100035, "grad_norm": 4.371758937835693, "learning_rate": 2.0282305736977944e-05, "loss": 4.4605, "step": 967000 }, { "epoch": 5.831152542781684, "grad_norm": 5.622526168823242, "learning_rate": 2.0272088841060768e-05, "loss": 4.1609, "step": 967050 }, { "epoch": 5.831454034563018, "grad_norm": 3.543942451477051, "learning_rate": 2.026187433260066e-05, "loss": 4.6304, "step": 967100 }, { "epoch": 5.831755526344352, "grad_norm": 3.9332685470581055, "learning_rate": 2.0251662211785647e-05, "loss": 4.4234, "step": 967150 }, { "epoch": 5.832057018125686, "grad_norm": 6.563753604888916, "learning_rate": 2.02414524788036e-05, "loss": 4.4402, "step": 967200 }, { "epoch": 5.83235850990702, "grad_norm": 4.4388813972473145, "learning_rate": 2.023124513384245e-05, "loss": 4.5448, "step": 967250 }, { "epoch": 5.832660001688354, "grad_norm": 2.2425310611724854, "learning_rate": 2.022104017709008e-05, "loss": 4.4247, "step": 967300 }, { "epoch": 5.832961493469688, "grad_norm": 3.580935478210449, "learning_rate": 2.0210837608734242e-05, "loss": 4.5642, "step": 967350 }, { "epoch": 5.833262985251022, "grad_norm": 4.535197734832764, "learning_rate": 2.02006374289627e-05, "loss": 4.2018, "step": 967400 }, { "epoch": 5.833564477032356, "grad_norm": 4.9104390144348145, "learning_rate": 2.0190439637963223e-05, "loss": 4.3049, "step": 967450 }, { "epoch": 5.83386596881369, "grad_norm": 3.7242181301116943, "learning_rate": 2.0180244235923427e-05, "loss": 4.1072, "step": 967500 }, { "epoch": 5.834167460595024, "grad_norm": 4.066199779510498, "learning_rate": 2.017005122303096e-05, "loss": 4.5847, "step": 967550 }, { "epoch": 5.834468952376358, "grad_norm": 1.2007708549499512, "learning_rate": 2.0159860599473454e-05, "loss": 4.6382, "step": 967600 }, { "epoch": 5.834770444157693, "grad_norm": 3.360651731491089, "learning_rate": 2.0149672365438414e-05, "loss": 4.7121, "step": 967650 }, { "epoch": 5.835071935939026, "grad_norm": 5.095125198364258, "learning_rate": 2.0139486521113352e-05, "loss": 4.4568, "step": 967700 }, { "epoch": 5.83537342772036, "grad_norm": 5.450149059295654, "learning_rate": 2.0129303066685703e-05, "loss": 3.9939, "step": 967750 }, { "epoch": 5.835674919501694, "grad_norm": 4.052934646606445, "learning_rate": 2.0119122002342936e-05, "loss": 4.0952, "step": 967800 }, { "epoch": 5.835976411283029, "grad_norm": 5.410974979400635, "learning_rate": 2.0108943328272332e-05, "loss": 4.2675, "step": 967850 }, { "epoch": 5.836277903064363, "grad_norm": 3.8507235050201416, "learning_rate": 2.009876704466129e-05, "loss": 4.5834, "step": 967900 }, { "epoch": 5.836579394845696, "grad_norm": 1.3802505731582642, "learning_rate": 2.0088593151697045e-05, "loss": 4.2033, "step": 967950 }, { "epoch": 5.83688088662703, "grad_norm": 5.058741092681885, "learning_rate": 2.0078421649566834e-05, "loss": 4.3729, "step": 968000 }, { "epoch": 5.8371823784083645, "grad_norm": 3.57487416267395, "learning_rate": 2.00682525384579e-05, "loss": 4.5509, "step": 968050 }, { "epoch": 5.837483870189699, "grad_norm": 4.803381443023682, "learning_rate": 2.0058085818557335e-05, "loss": 4.331, "step": 968100 }, { "epoch": 5.837785361971033, "grad_norm": 3.4146978855133057, "learning_rate": 2.0047921490052283e-05, "loss": 4.4961, "step": 968150 }, { "epoch": 5.838086853752367, "grad_norm": 4.076310634613037, "learning_rate": 2.0037759553129795e-05, "loss": 4.2116, "step": 968200 }, { "epoch": 5.8383883455337005, "grad_norm": 4.171250820159912, "learning_rate": 2.002760000797684e-05, "loss": 4.4912, "step": 968250 }, { "epoch": 5.838689837315035, "grad_norm": 3.0525333881378174, "learning_rate": 2.001744285478042e-05, "loss": 4.397, "step": 968300 }, { "epoch": 5.838991329096369, "grad_norm": 3.6600894927978516, "learning_rate": 2.0007288093727497e-05, "loss": 4.2661, "step": 968350 }, { "epoch": 5.839292820877703, "grad_norm": 3.6583213806152344, "learning_rate": 1.999713572500491e-05, "loss": 4.748, "step": 968400 }, { "epoch": 5.839594312659037, "grad_norm": 1.4240198135375977, "learning_rate": 1.9986985748799507e-05, "loss": 4.4384, "step": 968450 }, { "epoch": 5.839895804440371, "grad_norm": 4.4452738761901855, "learning_rate": 1.9976838165298136e-05, "loss": 4.6179, "step": 968500 }, { "epoch": 5.840197296221705, "grad_norm": 4.16596794128418, "learning_rate": 1.9966692974687488e-05, "loss": 4.2911, "step": 968550 }, { "epoch": 5.840498788003039, "grad_norm": 4.559218406677246, "learning_rate": 1.9956550177154272e-05, "loss": 4.5236, "step": 968600 }, { "epoch": 5.840800279784373, "grad_norm": 1.151795744895935, "learning_rate": 1.9946409772885198e-05, "loss": 4.2865, "step": 968650 }, { "epoch": 5.841101771565707, "grad_norm": 4.601001262664795, "learning_rate": 1.9936271762066806e-05, "loss": 4.1388, "step": 968700 }, { "epoch": 5.841403263347042, "grad_norm": 3.0413882732391357, "learning_rate": 1.9926136144885742e-05, "loss": 4.1627, "step": 968750 }, { "epoch": 5.841704755128375, "grad_norm": 3.5970287322998047, "learning_rate": 1.9916002921528545e-05, "loss": 4.9459, "step": 968800 }, { "epoch": 5.842006246909709, "grad_norm": 4.575346946716309, "learning_rate": 1.9905872092181625e-05, "loss": 4.2877, "step": 968850 }, { "epoch": 5.842307738691043, "grad_norm": 4.76402473449707, "learning_rate": 1.9895743657031525e-05, "loss": 4.6911, "step": 968900 }, { "epoch": 5.842609230472378, "grad_norm": 6.606489181518555, "learning_rate": 1.9885617616264555e-05, "loss": 4.3942, "step": 968950 }, { "epoch": 5.842910722253711, "grad_norm": 3.9387013912200928, "learning_rate": 1.9875493970067157e-05, "loss": 4.3278, "step": 969000 }, { "epoch": 5.842910722253711, "eval_loss": 4.870733737945557, "eval_runtime": 39.062, "eval_samples_per_second": 13.107, "eval_steps_per_second": 6.554, "eval_tts_loss": 7.943303594803877, "step": 969000 }, { "epoch": 5.843212214035045, "grad_norm": 4.491587162017822, "learning_rate": 1.986537271862554e-05, "loss": 4.388, "step": 969050 }, { "epoch": 5.843513705816379, "grad_norm": 1.1932077407836914, "learning_rate": 1.9855253862126032e-05, "loss": 4.2572, "step": 969100 }, { "epoch": 5.8438151975977135, "grad_norm": 5.635510444641113, "learning_rate": 1.98451374007549e-05, "loss": 4.7123, "step": 969150 }, { "epoch": 5.844116689379048, "grad_norm": 4.254388332366943, "learning_rate": 1.983502333469822e-05, "loss": 4.6991, "step": 969200 }, { "epoch": 5.844418181160382, "grad_norm": 3.902071714401245, "learning_rate": 1.9824911664142183e-05, "loss": 4.7036, "step": 969250 }, { "epoch": 5.844719672941715, "grad_norm": 3.1270642280578613, "learning_rate": 1.9814802389272915e-05, "loss": 4.3815, "step": 969300 }, { "epoch": 5.8450211647230494, "grad_norm": 4.309293746948242, "learning_rate": 1.9804695510276432e-05, "loss": 4.1396, "step": 969350 }, { "epoch": 5.845322656504384, "grad_norm": 4.124812126159668, "learning_rate": 1.979459102733868e-05, "loss": 4.3646, "step": 969400 }, { "epoch": 5.845624148285718, "grad_norm": 6.490222454071045, "learning_rate": 1.9784488940645715e-05, "loss": 4.4231, "step": 969450 }, { "epoch": 5.845925640067052, "grad_norm": 5.864051342010498, "learning_rate": 1.9774389250383354e-05, "loss": 4.4498, "step": 969500 }, { "epoch": 5.846227131848385, "grad_norm": 3.9920456409454346, "learning_rate": 1.9764291956737533e-05, "loss": 4.0319, "step": 969550 }, { "epoch": 5.84652862362972, "grad_norm": 4.804906368255615, "learning_rate": 1.9754197059894082e-05, "loss": 4.5815, "step": 969600 }, { "epoch": 5.846830115411054, "grad_norm": 5.00455904006958, "learning_rate": 1.9744104560038744e-05, "loss": 4.6193, "step": 969650 }, { "epoch": 5.847131607192388, "grad_norm": 3.141686201095581, "learning_rate": 1.973401445735731e-05, "loss": 4.7148, "step": 969700 }, { "epoch": 5.847433098973722, "grad_norm": 3.478799343109131, "learning_rate": 1.9723926752035424e-05, "loss": 4.6585, "step": 969750 }, { "epoch": 5.847734590755056, "grad_norm": 1.392079472541809, "learning_rate": 1.971384144425873e-05, "loss": 4.26, "step": 969800 }, { "epoch": 5.84803608253639, "grad_norm": 4.690187454223633, "learning_rate": 1.9703758534212856e-05, "loss": 4.1946, "step": 969850 }, { "epoch": 5.848337574317724, "grad_norm": 4.572935581207275, "learning_rate": 1.9693678022083393e-05, "loss": 4.459, "step": 969900 }, { "epoch": 5.848639066099058, "grad_norm": 4.282377243041992, "learning_rate": 1.9683599908055796e-05, "loss": 4.7938, "step": 969950 }, { "epoch": 5.848940557880392, "grad_norm": 2.533109426498413, "learning_rate": 1.967352419231557e-05, "loss": 4.1308, "step": 970000 }, { "epoch": 5.849242049661727, "grad_norm": 2.9010696411132812, "learning_rate": 1.966345087504818e-05, "loss": 4.0983, "step": 970050 }, { "epoch": 5.84954354144306, "grad_norm": 2.711892604827881, "learning_rate": 1.9653379956438948e-05, "loss": 4.2148, "step": 970100 }, { "epoch": 5.849845033224394, "grad_norm": 4.6784210205078125, "learning_rate": 1.9643311436673286e-05, "loss": 4.6525, "step": 970150 }, { "epoch": 5.850146525005728, "grad_norm": 3.2557594776153564, "learning_rate": 1.9633245315936435e-05, "loss": 4.3505, "step": 970200 }, { "epoch": 5.8504480167870625, "grad_norm": 4.879347801208496, "learning_rate": 1.9623181594413628e-05, "loss": 4.2889, "step": 970250 }, { "epoch": 5.850749508568397, "grad_norm": 5.018052101135254, "learning_rate": 1.96131202722901e-05, "loss": 4.4744, "step": 970300 }, { "epoch": 5.851051000349731, "grad_norm": 4.436483860015869, "learning_rate": 1.9603061349751053e-05, "loss": 4.554, "step": 970350 }, { "epoch": 5.851352492131064, "grad_norm": 3.0133540630340576, "learning_rate": 1.9593004826981558e-05, "loss": 4.5778, "step": 970400 }, { "epoch": 5.851653983912398, "grad_norm": 2.741614818572998, "learning_rate": 1.9582950704166694e-05, "loss": 4.3705, "step": 970450 }, { "epoch": 5.851955475693733, "grad_norm": 4.100342750549316, "learning_rate": 1.957289898149155e-05, "loss": 4.6665, "step": 970500 }, { "epoch": 5.852256967475067, "grad_norm": 4.501344203948975, "learning_rate": 1.9562849659141066e-05, "loss": 4.6037, "step": 970550 }, { "epoch": 5.852558459256401, "grad_norm": 5.880162239074707, "learning_rate": 1.9552802737300157e-05, "loss": 4.5002, "step": 970600 }, { "epoch": 5.852859951037734, "grad_norm": 3.7728304862976074, "learning_rate": 1.9542758216153786e-05, "loss": 4.5613, "step": 970650 }, { "epoch": 5.853161442819069, "grad_norm": 3.3363213539123535, "learning_rate": 1.9532716095886746e-05, "loss": 4.3052, "step": 970700 }, { "epoch": 5.853462934600403, "grad_norm": 5.942025184631348, "learning_rate": 1.952267637668389e-05, "loss": 4.3588, "step": 970750 }, { "epoch": 5.853764426381737, "grad_norm": 4.339900016784668, "learning_rate": 1.9512639058729995e-05, "loss": 4.4101, "step": 970800 }, { "epoch": 5.854065918163071, "grad_norm": 6.086250305175781, "learning_rate": 1.9502604142209737e-05, "loss": 4.617, "step": 970850 }, { "epoch": 5.854367409944405, "grad_norm": 3.557842254638672, "learning_rate": 1.949257162730785e-05, "loss": 4.2562, "step": 970900 }, { "epoch": 5.854668901725739, "grad_norm": 4.35530948638916, "learning_rate": 1.948254151420894e-05, "loss": 4.1722, "step": 970950 }, { "epoch": 5.854970393507073, "grad_norm": 5.081782817840576, "learning_rate": 1.947251380309755e-05, "loss": 4.6885, "step": 971000 }, { "epoch": 5.855271885288407, "grad_norm": 6.689834117889404, "learning_rate": 1.946248849415829e-05, "loss": 4.2483, "step": 971050 }, { "epoch": 5.855573377069741, "grad_norm": 5.590266704559326, "learning_rate": 1.945246558757566e-05, "loss": 4.0954, "step": 971100 }, { "epoch": 5.855874868851075, "grad_norm": 3.5268301963806152, "learning_rate": 1.9442445083534062e-05, "loss": 4.4365, "step": 971150 }, { "epoch": 5.856176360632409, "grad_norm": 3.2017178535461426, "learning_rate": 1.9432426982217967e-05, "loss": 4.5565, "step": 971200 }, { "epoch": 5.856477852413743, "grad_norm": 4.9254937171936035, "learning_rate": 1.9422411283811744e-05, "loss": 4.58, "step": 971250 }, { "epoch": 5.856779344195077, "grad_norm": 3.6050615310668945, "learning_rate": 1.9412397988499663e-05, "loss": 4.391, "step": 971300 }, { "epoch": 5.8570808359764115, "grad_norm": 3.489804744720459, "learning_rate": 1.9402387096466072e-05, "loss": 4.1844, "step": 971350 }, { "epoch": 5.857382327757746, "grad_norm": 3.194823741912842, "learning_rate": 1.9392378607895177e-05, "loss": 4.3667, "step": 971400 }, { "epoch": 5.857683819539079, "grad_norm": 3.7654213905334473, "learning_rate": 1.9382372522971145e-05, "loss": 4.3683, "step": 971450 }, { "epoch": 5.857985311320413, "grad_norm": 5.16851282119751, "learning_rate": 1.9372368841878132e-05, "loss": 4.687, "step": 971500 }, { "epoch": 5.858286803101747, "grad_norm": 3.012112855911255, "learning_rate": 1.936236756480029e-05, "loss": 4.6436, "step": 971550 }, { "epoch": 5.858588294883082, "grad_norm": 3.171234130859375, "learning_rate": 1.9352368691921604e-05, "loss": 4.4511, "step": 971600 }, { "epoch": 5.858889786664416, "grad_norm": 4.404847621917725, "learning_rate": 1.934237222342613e-05, "loss": 3.9373, "step": 971650 }, { "epoch": 5.859191278445749, "grad_norm": 6.843914985656738, "learning_rate": 1.933237815949788e-05, "loss": 4.4973, "step": 971700 }, { "epoch": 5.859492770227083, "grad_norm": 2.979463815689087, "learning_rate": 1.932238650032073e-05, "loss": 4.2231, "step": 971750 }, { "epoch": 5.859794262008418, "grad_norm": 4.785796165466309, "learning_rate": 1.931239724607853e-05, "loss": 4.3978, "step": 971800 }, { "epoch": 5.860095753789752, "grad_norm": 3.915130138397217, "learning_rate": 1.9302410396955205e-05, "loss": 4.5765, "step": 971850 }, { "epoch": 5.860397245571086, "grad_norm": 5.746490478515625, "learning_rate": 1.929242595313445e-05, "loss": 4.1541, "step": 971900 }, { "epoch": 5.86069873735242, "grad_norm": 2.041956901550293, "learning_rate": 1.9282443914800087e-05, "loss": 4.2712, "step": 971950 }, { "epoch": 5.8610002291337535, "grad_norm": 4.752374649047852, "learning_rate": 1.927246428213581e-05, "loss": 4.6325, "step": 972000 }, { "epoch": 5.8610002291337535, "eval_loss": 4.8693952560424805, "eval_runtime": 39.202, "eval_samples_per_second": 13.061, "eval_steps_per_second": 6.53, "eval_tts_loss": 7.937887298445664, "step": 972000 }, { "epoch": 5.861301720915088, "grad_norm": 1.9041974544525146, "learning_rate": 1.926248705532526e-05, "loss": 4.5592, "step": 972050 }, { "epoch": 5.861603212696422, "grad_norm": 3.528597593307495, "learning_rate": 1.9252512234552082e-05, "loss": 4.7809, "step": 972100 }, { "epoch": 5.861904704477756, "grad_norm": 3.7688305377960205, "learning_rate": 1.9242539819999824e-05, "loss": 4.5407, "step": 972150 }, { "epoch": 5.86220619625909, "grad_norm": 5.660110950469971, "learning_rate": 1.9232569811851994e-05, "loss": 4.458, "step": 972200 }, { "epoch": 5.862507688040424, "grad_norm": 4.28179931640625, "learning_rate": 1.922260221029211e-05, "loss": 4.0975, "step": 972250 }, { "epoch": 5.862809179821758, "grad_norm": 5.090517520904541, "learning_rate": 1.9212637015503624e-05, "loss": 4.6941, "step": 972300 }, { "epoch": 5.863110671603092, "grad_norm": 3.5549821853637695, "learning_rate": 1.920267422766989e-05, "loss": 4.4535, "step": 972350 }, { "epoch": 5.863412163384426, "grad_norm": 4.9224066734313965, "learning_rate": 1.919271384697426e-05, "loss": 4.5897, "step": 972400 }, { "epoch": 5.8637136551657605, "grad_norm": 5.06522798538208, "learning_rate": 1.9182755873600074e-05, "loss": 4.6359, "step": 972450 }, { "epoch": 5.864015146947095, "grad_norm": 4.490753650665283, "learning_rate": 1.9172800307730618e-05, "loss": 4.5913, "step": 972500 }, { "epoch": 5.864316638728428, "grad_norm": 5.262799263000488, "learning_rate": 1.916284714954906e-05, "loss": 4.2865, "step": 972550 }, { "epoch": 5.864618130509762, "grad_norm": 2.9671714305877686, "learning_rate": 1.9152896399238554e-05, "loss": 4.1851, "step": 972600 }, { "epoch": 5.864919622291096, "grad_norm": 6.328374862670898, "learning_rate": 1.914294805698229e-05, "loss": 4.5355, "step": 972650 }, { "epoch": 5.865221114072431, "grad_norm": 4.66180419921875, "learning_rate": 1.91330021229633e-05, "loss": 4.1977, "step": 972700 }, { "epoch": 5.865522605853764, "grad_norm": 4.113300323486328, "learning_rate": 1.9123058597364655e-05, "loss": 4.3318, "step": 972750 }, { "epoch": 5.865824097635098, "grad_norm": 4.029775142669678, "learning_rate": 1.911311748036936e-05, "loss": 4.8027, "step": 972800 }, { "epoch": 5.866125589416432, "grad_norm": 4.67848014831543, "learning_rate": 1.910317877216032e-05, "loss": 4.879, "step": 972850 }, { "epoch": 5.866427081197767, "grad_norm": 4.608735084533691, "learning_rate": 1.909324247292052e-05, "loss": 4.3898, "step": 972900 }, { "epoch": 5.866728572979101, "grad_norm": 3.4059154987335205, "learning_rate": 1.908330858283277e-05, "loss": 4.3988, "step": 972950 }, { "epoch": 5.867030064760435, "grad_norm": 5.142661094665527, "learning_rate": 1.9073377102079877e-05, "loss": 4.4123, "step": 973000 }, { "epoch": 5.867331556541768, "grad_norm": 4.830638408660889, "learning_rate": 1.9063448030844626e-05, "loss": 4.2283, "step": 973050 }, { "epoch": 5.8676330483231025, "grad_norm": 4.709137916564941, "learning_rate": 1.9053521369309794e-05, "loss": 4.0774, "step": 973100 }, { "epoch": 5.867934540104437, "grad_norm": 3.5062553882598877, "learning_rate": 1.904359711765799e-05, "loss": 4.6974, "step": 973150 }, { "epoch": 5.868236031885771, "grad_norm": 4.9335551261901855, "learning_rate": 1.9033675276071895e-05, "loss": 4.3129, "step": 973200 }, { "epoch": 5.868537523667105, "grad_norm": 2.555645704269409, "learning_rate": 1.9023755844734157e-05, "loss": 4.5219, "step": 973250 }, { "epoch": 5.8688390154484384, "grad_norm": 4.991402626037598, "learning_rate": 1.901383882382727e-05, "loss": 4.258, "step": 973300 }, { "epoch": 5.869140507229773, "grad_norm": 3.480776786804199, "learning_rate": 1.900392421353371e-05, "loss": 4.4557, "step": 973350 }, { "epoch": 5.869441999011107, "grad_norm": 6.671649932861328, "learning_rate": 1.899401201403603e-05, "loss": 4.5311, "step": 973400 }, { "epoch": 5.869743490792441, "grad_norm": 4.0903730392456055, "learning_rate": 1.8984102225516568e-05, "loss": 4.4618, "step": 973450 }, { "epoch": 5.870044982573775, "grad_norm": 4.481087684631348, "learning_rate": 1.8974194848157725e-05, "loss": 4.1193, "step": 973500 }, { "epoch": 5.8703464743551095, "grad_norm": 3.5358035564422607, "learning_rate": 1.896428988214189e-05, "loss": 4.1766, "step": 973550 }, { "epoch": 5.870647966136443, "grad_norm": 5.1236348152160645, "learning_rate": 1.8954387327651254e-05, "loss": 4.6807, "step": 973600 }, { "epoch": 5.870949457917777, "grad_norm": 5.781940937042236, "learning_rate": 1.8944487184868118e-05, "loss": 4.3008, "step": 973650 }, { "epoch": 5.871250949699111, "grad_norm": 3.950582265853882, "learning_rate": 1.89345894539747e-05, "loss": 4.4447, "step": 973700 }, { "epoch": 5.871552441480445, "grad_norm": 4.980359077453613, "learning_rate": 1.8924694135153128e-05, "loss": 4.5139, "step": 973750 }, { "epoch": 5.87185393326178, "grad_norm": 3.3907341957092285, "learning_rate": 1.891480122858545e-05, "loss": 4.566, "step": 973800 }, { "epoch": 5.872155425043113, "grad_norm": 4.111329555511475, "learning_rate": 1.8904910734453842e-05, "loss": 4.4812, "step": 973850 }, { "epoch": 5.872456916824447, "grad_norm": 4.307956695556641, "learning_rate": 1.8895022652940238e-05, "loss": 4.4643, "step": 973900 }, { "epoch": 5.872758408605781, "grad_norm": 6.082640647888184, "learning_rate": 1.8885136984226628e-05, "loss": 4.5123, "step": 973950 }, { "epoch": 5.873059900387116, "grad_norm": 4.856224060058594, "learning_rate": 1.8875253728495016e-05, "loss": 4.2509, "step": 974000 }, { "epoch": 5.87336139216845, "grad_norm": 3.4834232330322266, "learning_rate": 1.8865372885927186e-05, "loss": 4.2893, "step": 974050 }, { "epoch": 5.873662883949784, "grad_norm": 3.8670814037323, "learning_rate": 1.885549445670505e-05, "loss": 4.2136, "step": 974100 }, { "epoch": 5.873964375731117, "grad_norm": 3.6713614463806152, "learning_rate": 1.8845618441010408e-05, "loss": 4.3392, "step": 974150 }, { "epoch": 5.8742658675124515, "grad_norm": 3.536773443222046, "learning_rate": 1.8835744839024946e-05, "loss": 4.6482, "step": 974200 }, { "epoch": 5.874567359293786, "grad_norm": 4.569503307342529, "learning_rate": 1.8825873650930422e-05, "loss": 4.3134, "step": 974250 }, { "epoch": 5.87486885107512, "grad_norm": 3.4697933197021484, "learning_rate": 1.8816004876908543e-05, "loss": 4.5799, "step": 974300 }, { "epoch": 5.875170342856454, "grad_norm": 1.9044487476348877, "learning_rate": 1.8806138517140824e-05, "loss": 4.5076, "step": 974350 }, { "epoch": 5.875471834637787, "grad_norm": 4.811226844787598, "learning_rate": 1.879627457180892e-05, "loss": 4.1833, "step": 974400 }, { "epoch": 5.875773326419122, "grad_norm": 3.6180458068847656, "learning_rate": 1.878641304109438e-05, "loss": 4.3633, "step": 974450 }, { "epoch": 5.876074818200456, "grad_norm": 3.6283962726593018, "learning_rate": 1.8776553925178612e-05, "loss": 4.2226, "step": 974500 }, { "epoch": 5.87637630998179, "grad_norm": 3.4586431980133057, "learning_rate": 1.8766697224243143e-05, "loss": 4.2941, "step": 974550 }, { "epoch": 5.876677801763124, "grad_norm": 3.2307093143463135, "learning_rate": 1.875684293846933e-05, "loss": 4.4485, "step": 974600 }, { "epoch": 5.8769792935444585, "grad_norm": 3.909358024597168, "learning_rate": 1.8746991068038508e-05, "loss": 4.5946, "step": 974650 }, { "epoch": 5.877280785325792, "grad_norm": 4.778355121612549, "learning_rate": 1.8737141613131984e-05, "loss": 4.4802, "step": 974700 }, { "epoch": 5.877582277107126, "grad_norm": 4.835002422332764, "learning_rate": 1.8727294573931097e-05, "loss": 3.9501, "step": 974750 }, { "epoch": 5.87788376888846, "grad_norm": 4.527201175689697, "learning_rate": 1.871744995061698e-05, "loss": 4.4387, "step": 974800 }, { "epoch": 5.878185260669794, "grad_norm": 4.308828830718994, "learning_rate": 1.8707607743370845e-05, "loss": 4.3088, "step": 974850 }, { "epoch": 5.878486752451128, "grad_norm": 4.864686012268066, "learning_rate": 1.869776795237386e-05, "loss": 4.2466, "step": 974900 }, { "epoch": 5.878788244232462, "grad_norm": 3.4718616008758545, "learning_rate": 1.8687930577807082e-05, "loss": 4.1391, "step": 974950 }, { "epoch": 5.879089736013796, "grad_norm": 5.86060094833374, "learning_rate": 1.8678095619851497e-05, "loss": 4.439, "step": 975000 }, { "epoch": 5.879089736013796, "eval_loss": 4.866596698760986, "eval_runtime": 38.9763, "eval_samples_per_second": 13.136, "eval_steps_per_second": 6.568, "eval_tts_loss": 7.9240611975231126, "step": 975000 }, { "epoch": 5.87939122779513, "grad_norm": 3.722712516784668, "learning_rate": 1.8668263078688213e-05, "loss": 4.1095, "step": 975050 }, { "epoch": 5.8796927195764646, "grad_norm": 2.1173453330993652, "learning_rate": 1.8658432954498082e-05, "loss": 4.0867, "step": 975100 }, { "epoch": 5.879994211357799, "grad_norm": 2.7367076873779297, "learning_rate": 1.864860524746206e-05, "loss": 4.2901, "step": 975150 }, { "epoch": 5.880295703139132, "grad_norm": 4.959259033203125, "learning_rate": 1.8638779957761042e-05, "loss": 4.3826, "step": 975200 }, { "epoch": 5.880597194920466, "grad_norm": 4.984342575073242, "learning_rate": 1.8628957085575776e-05, "loss": 4.4917, "step": 975250 }, { "epoch": 5.8808986867018005, "grad_norm": 3.7315142154693604, "learning_rate": 1.8619136631087116e-05, "loss": 4.2512, "step": 975300 }, { "epoch": 5.881200178483135, "grad_norm": 3.61983323097229, "learning_rate": 1.860931859447576e-05, "loss": 4.6758, "step": 975350 }, { "epoch": 5.881501670264469, "grad_norm": 3.5755271911621094, "learning_rate": 1.8599502975922352e-05, "loss": 4.4858, "step": 975400 }, { "epoch": 5.881803162045802, "grad_norm": 2.701289176940918, "learning_rate": 1.8589689775607574e-05, "loss": 4.361, "step": 975450 }, { "epoch": 5.882104653827136, "grad_norm": 3.407623767852783, "learning_rate": 1.8579878993712062e-05, "loss": 4.3772, "step": 975500 }, { "epoch": 5.882406145608471, "grad_norm": 3.799684762954712, "learning_rate": 1.85700706304163e-05, "loss": 4.4344, "step": 975550 }, { "epoch": 5.882707637389805, "grad_norm": 6.748619079589844, "learning_rate": 1.8560264685900817e-05, "loss": 4.3307, "step": 975600 }, { "epoch": 5.883009129171139, "grad_norm": 2.087207555770874, "learning_rate": 1.855046116034613e-05, "loss": 4.5196, "step": 975650 }, { "epoch": 5.883310620952473, "grad_norm": 3.9788382053375244, "learning_rate": 1.854066005393258e-05, "loss": 4.0177, "step": 975700 }, { "epoch": 5.883612112733807, "grad_norm": 4.219176769256592, "learning_rate": 1.853086136684061e-05, "loss": 4.324, "step": 975750 }, { "epoch": 5.883913604515141, "grad_norm": 4.531744003295898, "learning_rate": 1.8521065099250488e-05, "loss": 4.471, "step": 975800 }, { "epoch": 5.884215096296475, "grad_norm": 3.9980087280273438, "learning_rate": 1.8511271251342567e-05, "loss": 4.3346, "step": 975850 }, { "epoch": 5.884516588077809, "grad_norm": 4.095396041870117, "learning_rate": 1.850147982329703e-05, "loss": 4.6787, "step": 975900 }, { "epoch": 5.884818079859143, "grad_norm": 5.6036834716796875, "learning_rate": 1.8491690815294084e-05, "loss": 4.2506, "step": 975950 }, { "epoch": 5.885119571640477, "grad_norm": 3.2905776500701904, "learning_rate": 1.848190422751393e-05, "loss": 4.5763, "step": 976000 }, { "epoch": 5.885421063421811, "grad_norm": 5.297740936279297, "learning_rate": 1.847212006013661e-05, "loss": 4.607, "step": 976050 }, { "epoch": 5.885722555203145, "grad_norm": 5.279068946838379, "learning_rate": 1.846233831334225e-05, "loss": 3.9521, "step": 976100 }, { "epoch": 5.886024046984479, "grad_norm": 2.4473063945770264, "learning_rate": 1.845255898731085e-05, "loss": 4.3312, "step": 976150 }, { "epoch": 5.8863255387658135, "grad_norm": 4.057143688201904, "learning_rate": 1.8442782082222328e-05, "loss": 4.3377, "step": 976200 }, { "epoch": 5.886627030547148, "grad_norm": 7.562154293060303, "learning_rate": 1.8433007598256677e-05, "loss": 4.3938, "step": 976250 }, { "epoch": 5.886928522328481, "grad_norm": 3.602287769317627, "learning_rate": 1.8423235535593782e-05, "loss": 4.4992, "step": 976300 }, { "epoch": 5.887230014109815, "grad_norm": 4.560611724853516, "learning_rate": 1.8413465894413436e-05, "loss": 4.2388, "step": 976350 }, { "epoch": 5.8875315058911495, "grad_norm": 4.219459056854248, "learning_rate": 1.840369867489548e-05, "loss": 4.7578, "step": 976400 }, { "epoch": 5.887832997672484, "grad_norm": 3.871006727218628, "learning_rate": 1.839393387721968e-05, "loss": 4.6033, "step": 976450 }, { "epoch": 5.888134489453817, "grad_norm": 1.3763419389724731, "learning_rate": 1.8384171501565698e-05, "loss": 4.2914, "step": 976500 }, { "epoch": 5.888435981235151, "grad_norm": 1.997152328491211, "learning_rate": 1.8374411548113206e-05, "loss": 4.4957, "step": 976550 }, { "epoch": 5.888737473016485, "grad_norm": 3.7615814208984375, "learning_rate": 1.836465401704184e-05, "loss": 4.3173, "step": 976600 }, { "epoch": 5.88903896479782, "grad_norm": 5.015923976898193, "learning_rate": 1.8354898908531145e-05, "loss": 4.5591, "step": 976650 }, { "epoch": 5.889340456579154, "grad_norm": 3.192643165588379, "learning_rate": 1.834514622276066e-05, "loss": 4.4467, "step": 976700 }, { "epoch": 5.889641948360488, "grad_norm": 5.46851921081543, "learning_rate": 1.8335395959909903e-05, "loss": 4.6561, "step": 976750 }, { "epoch": 5.889943440141821, "grad_norm": 4.766038417816162, "learning_rate": 1.8325648120158272e-05, "loss": 4.1165, "step": 976800 }, { "epoch": 5.890244931923156, "grad_norm": 5.054922580718994, "learning_rate": 1.8315902703685165e-05, "loss": 4.4707, "step": 976850 }, { "epoch": 5.89054642370449, "grad_norm": 4.318323135375977, "learning_rate": 1.8306159710669993e-05, "loss": 4.3913, "step": 976900 }, { "epoch": 5.890847915485824, "grad_norm": 2.4675393104553223, "learning_rate": 1.8296419141291996e-05, "loss": 4.3596, "step": 976950 }, { "epoch": 5.891149407267158, "grad_norm": 5.090371608734131, "learning_rate": 1.828668099573043e-05, "loss": 4.291, "step": 977000 }, { "epoch": 5.8914508990484915, "grad_norm": 3.666515827178955, "learning_rate": 1.8276945274164566e-05, "loss": 4.4918, "step": 977050 }, { "epoch": 5.891752390829826, "grad_norm": 3.3370859622955322, "learning_rate": 1.8267211976773516e-05, "loss": 4.4885, "step": 977100 }, { "epoch": 5.89205388261116, "grad_norm": 5.431835174560547, "learning_rate": 1.825748110373643e-05, "loss": 4.3565, "step": 977150 }, { "epoch": 5.892355374392494, "grad_norm": 3.525399923324585, "learning_rate": 1.824775265523244e-05, "loss": 4.4602, "step": 977200 }, { "epoch": 5.892656866173828, "grad_norm": 3.337379217147827, "learning_rate": 1.8238026631440512e-05, "loss": 3.9871, "step": 977250 }, { "epoch": 5.8929583579551625, "grad_norm": 3.735752582550049, "learning_rate": 1.8228303032539687e-05, "loss": 4.2947, "step": 977300 }, { "epoch": 5.893259849736496, "grad_norm": 3.8763885498046875, "learning_rate": 1.8218581858708907e-05, "loss": 4.1583, "step": 977350 }, { "epoch": 5.89356134151783, "grad_norm": 3.255751609802246, "learning_rate": 1.8208863110127033e-05, "loss": 4.3622, "step": 977400 }, { "epoch": 5.893862833299164, "grad_norm": 4.632932662963867, "learning_rate": 1.8199146786972947e-05, "loss": 4.1446, "step": 977450 }, { "epoch": 5.8941643250804985, "grad_norm": 4.564955234527588, "learning_rate": 1.8189432889425527e-05, "loss": 4.1317, "step": 977500 }, { "epoch": 5.894465816861833, "grad_norm": 3.0796258449554443, "learning_rate": 1.8179721417663444e-05, "loss": 4.7584, "step": 977550 }, { "epoch": 5.894767308643166, "grad_norm": 3.723079204559326, "learning_rate": 1.817001237186546e-05, "loss": 4.7156, "step": 977600 }, { "epoch": 5.8950688004245, "grad_norm": 3.6721041202545166, "learning_rate": 1.8160305752210312e-05, "loss": 4.267, "step": 977650 }, { "epoch": 5.895370292205834, "grad_norm": 4.691669464111328, "learning_rate": 1.8150601558876592e-05, "loss": 4.4018, "step": 977700 }, { "epoch": 5.895671783987169, "grad_norm": 3.748300790786743, "learning_rate": 1.8140899792042855e-05, "loss": 4.492, "step": 977750 }, { "epoch": 5.895973275768503, "grad_norm": 4.003288745880127, "learning_rate": 1.8131200451887727e-05, "loss": 4.2864, "step": 977800 }, { "epoch": 5.896274767549837, "grad_norm": 3.43825364112854, "learning_rate": 1.8121503538589634e-05, "loss": 4.3487, "step": 977850 }, { "epoch": 5.89657625933117, "grad_norm": 4.935304164886475, "learning_rate": 1.811180905232706e-05, "loss": 4.7396, "step": 977900 }, { "epoch": 5.896877751112505, "grad_norm": 4.846500396728516, "learning_rate": 1.8102116993278465e-05, "loss": 4.5545, "step": 977950 }, { "epoch": 5.897179242893839, "grad_norm": 5.491771697998047, "learning_rate": 1.809242736162214e-05, "loss": 4.3903, "step": 978000 }, { "epoch": 5.897179242893839, "eval_loss": 4.858097076416016, "eval_runtime": 39.0389, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.558, "eval_tts_loss": 7.987699900306004, "step": 978000 }, { "epoch": 5.897480734675173, "grad_norm": 3.864499807357788, "learning_rate": 1.8082740157536448e-05, "loss": 4.4698, "step": 978050 }, { "epoch": 5.897782226456507, "grad_norm": 2.9591517448425293, "learning_rate": 1.8073055381199685e-05, "loss": 4.8648, "step": 978100 }, { "epoch": 5.8980837182378405, "grad_norm": 4.769310474395752, "learning_rate": 1.8063373032790085e-05, "loss": 4.215, "step": 978150 }, { "epoch": 5.898385210019175, "grad_norm": 4.699121952056885, "learning_rate": 1.8053693112485783e-05, "loss": 4.3551, "step": 978200 }, { "epoch": 5.898686701800509, "grad_norm": 7.259535789489746, "learning_rate": 1.8044015620465007e-05, "loss": 4.3982, "step": 978250 }, { "epoch": 5.898988193581843, "grad_norm": 3.5635783672332764, "learning_rate": 1.8034340556905762e-05, "loss": 4.4963, "step": 978300 }, { "epoch": 5.899289685363177, "grad_norm": 4.092696189880371, "learning_rate": 1.8024667921986175e-05, "loss": 4.509, "step": 978350 }, { "epoch": 5.8995911771445115, "grad_norm": 4.988131523132324, "learning_rate": 1.801499771588425e-05, "loss": 4.302, "step": 978400 }, { "epoch": 5.899892668925845, "grad_norm": 4.455691814422607, "learning_rate": 1.8005329938777913e-05, "loss": 4.2463, "step": 978450 }, { "epoch": 5.900194160707179, "grad_norm": 4.980223655700684, "learning_rate": 1.7995664590845138e-05, "loss": 4.3777, "step": 978500 }, { "epoch": 5.900495652488513, "grad_norm": 4.656064510345459, "learning_rate": 1.7986001672263785e-05, "loss": 4.1415, "step": 978550 }, { "epoch": 5.9007971442698475, "grad_norm": 6.925421237945557, "learning_rate": 1.797634118321164e-05, "loss": 4.3201, "step": 978600 }, { "epoch": 5.901098636051181, "grad_norm": 4.973115921020508, "learning_rate": 1.7966683123866533e-05, "loss": 4.3081, "step": 978650 }, { "epoch": 5.901400127832515, "grad_norm": 0.9122458100318909, "learning_rate": 1.7957027494406235e-05, "loss": 4.0763, "step": 978700 }, { "epoch": 5.901701619613849, "grad_norm": 3.826215982437134, "learning_rate": 1.794737429500837e-05, "loss": 4.3163, "step": 978750 }, { "epoch": 5.902003111395183, "grad_norm": 3.8378381729125977, "learning_rate": 1.7937723525850633e-05, "loss": 4.2062, "step": 978800 }, { "epoch": 5.902304603176518, "grad_norm": 5.5557756423950195, "learning_rate": 1.7928075187110663e-05, "loss": 4.1302, "step": 978850 }, { "epoch": 5.902606094957852, "grad_norm": 3.975740671157837, "learning_rate": 1.7918429278966002e-05, "loss": 4.6979, "step": 978900 }, { "epoch": 5.902907586739185, "grad_norm": 3.7983345985412598, "learning_rate": 1.7908785801594105e-05, "loss": 4.2644, "step": 978950 }, { "epoch": 5.903209078520519, "grad_norm": 5.295588493347168, "learning_rate": 1.789914475517255e-05, "loss": 4.2637, "step": 979000 }, { "epoch": 5.9035105703018536, "grad_norm": 4.750382900238037, "learning_rate": 1.7889506139878694e-05, "loss": 4.44, "step": 979050 }, { "epoch": 5.903812062083188, "grad_norm": 3.079785108566284, "learning_rate": 1.7879869955889932e-05, "loss": 4.0455, "step": 979100 }, { "epoch": 5.904113553864522, "grad_norm": 3.6168670654296875, "learning_rate": 1.7870236203383648e-05, "loss": 4.4504, "step": 979150 }, { "epoch": 5.904415045645855, "grad_norm": 1.590810775756836, "learning_rate": 1.786060488253709e-05, "loss": 4.0146, "step": 979200 }, { "epoch": 5.9047165374271895, "grad_norm": 2.0574896335601807, "learning_rate": 1.785097599352751e-05, "loss": 4.2011, "step": 979250 }, { "epoch": 5.905018029208524, "grad_norm": 5.292647838592529, "learning_rate": 1.7841349536532172e-05, "loss": 4.3144, "step": 979300 }, { "epoch": 5.905319520989858, "grad_norm": 3.147834539413452, "learning_rate": 1.78317255117282e-05, "loss": 4.2231, "step": 979350 }, { "epoch": 5.905621012771192, "grad_norm": 6.518436908721924, "learning_rate": 1.7822103919292652e-05, "loss": 4.644, "step": 979400 }, { "epoch": 5.905922504552526, "grad_norm": 2.400001049041748, "learning_rate": 1.7812484759402668e-05, "loss": 4.6856, "step": 979450 }, { "epoch": 5.90622399633386, "grad_norm": 3.528200626373291, "learning_rate": 1.7802868032235307e-05, "loss": 4.5108, "step": 979500 }, { "epoch": 5.906525488115194, "grad_norm": 4.6877121925354, "learning_rate": 1.7793253737967446e-05, "loss": 4.0086, "step": 979550 }, { "epoch": 5.906826979896528, "grad_norm": 3.6062095165252686, "learning_rate": 1.7783641876776088e-05, "loss": 4.0749, "step": 979600 }, { "epoch": 5.907128471677862, "grad_norm": 2.800950050354004, "learning_rate": 1.7774032448838165e-05, "loss": 4.3839, "step": 979650 }, { "epoch": 5.9074299634591965, "grad_norm": 5.849330425262451, "learning_rate": 1.7764425454330466e-05, "loss": 4.3429, "step": 979700 }, { "epoch": 5.90773145524053, "grad_norm": 3.9336960315704346, "learning_rate": 1.7754820893429783e-05, "loss": 4.4285, "step": 979750 }, { "epoch": 5.908032947021864, "grad_norm": 4.605312824249268, "learning_rate": 1.7745218766312906e-05, "loss": 4.2775, "step": 979800 }, { "epoch": 5.908334438803198, "grad_norm": 4.02051305770874, "learning_rate": 1.773561907315653e-05, "loss": 4.2303, "step": 979850 }, { "epoch": 5.908635930584532, "grad_norm": 4.985191345214844, "learning_rate": 1.7726021814137332e-05, "loss": 4.3478, "step": 979900 }, { "epoch": 5.908937422365867, "grad_norm": 2.648287057876587, "learning_rate": 1.7716426989431965e-05, "loss": 4.5785, "step": 979950 }, { "epoch": 5.909238914147201, "grad_norm": 4.187495231628418, "learning_rate": 1.770683459921694e-05, "loss": 4.1067, "step": 980000 }, { "epoch": 5.909540405928534, "grad_norm": 2.84623384475708, "learning_rate": 1.7697244643668886e-05, "loss": 4.3996, "step": 980050 }, { "epoch": 5.909841897709868, "grad_norm": 3.3426737785339355, "learning_rate": 1.768765712296419e-05, "loss": 4.5872, "step": 980100 }, { "epoch": 5.9101433894912025, "grad_norm": 5.031069755554199, "learning_rate": 1.767807203727938e-05, "loss": 4.4173, "step": 980150 }, { "epoch": 5.910444881272537, "grad_norm": 2.5485692024230957, "learning_rate": 1.766848938679078e-05, "loss": 4.2816, "step": 980200 }, { "epoch": 5.91074637305387, "grad_norm": 4.204954624176025, "learning_rate": 1.7658909171674834e-05, "loss": 4.678, "step": 980250 }, { "epoch": 5.911047864835204, "grad_norm": 4.584493637084961, "learning_rate": 1.764933139210777e-05, "loss": 4.0809, "step": 980300 }, { "epoch": 5.9113493566165385, "grad_norm": 5.96432638168335, "learning_rate": 1.7639756048265875e-05, "loss": 4.6919, "step": 980350 }, { "epoch": 5.911650848397873, "grad_norm": 4.070333957672119, "learning_rate": 1.763018314032541e-05, "loss": 4.132, "step": 980400 }, { "epoch": 5.911952340179207, "grad_norm": 4.650291442871094, "learning_rate": 1.7620612668462503e-05, "loss": 3.9793, "step": 980450 }, { "epoch": 5.912253831960541, "grad_norm": 4.837878704071045, "learning_rate": 1.7611044632853326e-05, "loss": 4.5177, "step": 980500 }, { "epoch": 5.912555323741874, "grad_norm": 5.612051010131836, "learning_rate": 1.760147903367396e-05, "loss": 4.3763, "step": 980550 }, { "epoch": 5.912856815523209, "grad_norm": 1.8375362157821655, "learning_rate": 1.7591915871100393e-05, "loss": 4.1304, "step": 980600 }, { "epoch": 5.913158307304543, "grad_norm": 3.715153455734253, "learning_rate": 1.758235514530865e-05, "loss": 4.4811, "step": 980650 }, { "epoch": 5.913459799085877, "grad_norm": 3.8523342609405518, "learning_rate": 1.757279685647473e-05, "loss": 4.2603, "step": 980700 }, { "epoch": 5.913761290867211, "grad_norm": 2.596477508544922, "learning_rate": 1.7563241004774454e-05, "loss": 4.4397, "step": 980750 }, { "epoch": 5.914062782648545, "grad_norm": 3.448751211166382, "learning_rate": 1.755368759038375e-05, "loss": 4.6314, "step": 980800 }, { "epoch": 5.914364274429879, "grad_norm": 4.679253101348877, "learning_rate": 1.754413661347842e-05, "loss": 4.5881, "step": 980850 }, { "epoch": 5.914665766211213, "grad_norm": 3.899998903274536, "learning_rate": 1.7534588074234252e-05, "loss": 4.3639, "step": 980900 }, { "epoch": 5.914967257992547, "grad_norm": 5.0473432540893555, "learning_rate": 1.7525041972826913e-05, "loss": 4.297, "step": 980950 }, { "epoch": 5.915268749773881, "grad_norm": 5.867341041564941, "learning_rate": 1.7515498309432167e-05, "loss": 4.497, "step": 981000 }, { "epoch": 5.915268749773881, "eval_loss": 4.857856750488281, "eval_runtime": 38.8943, "eval_samples_per_second": 13.164, "eval_steps_per_second": 6.582, "eval_tts_loss": 7.971245755933453, "step": 981000 }, { "epoch": 5.915570241555216, "grad_norm": 5.143268585205078, "learning_rate": 1.7505957084225575e-05, "loss": 4.6755, "step": 981050 }, { "epoch": 5.915871733336549, "grad_norm": 3.186105251312256, "learning_rate": 1.7496418297382758e-05, "loss": 4.403, "step": 981100 }, { "epoch": 5.916173225117883, "grad_norm": 5.159701824188232, "learning_rate": 1.7486881949079294e-05, "loss": 4.5582, "step": 981150 }, { "epoch": 5.916474716899217, "grad_norm": 5.1135663986206055, "learning_rate": 1.747734803949063e-05, "loss": 4.3102, "step": 981200 }, { "epoch": 5.9167762086805515, "grad_norm": 7.92324161529541, "learning_rate": 1.746781656879227e-05, "loss": 4.6272, "step": 981250 }, { "epoch": 5.917077700461886, "grad_norm": 5.142251014709473, "learning_rate": 1.7458287537159645e-05, "loss": 4.6401, "step": 981300 }, { "epoch": 5.917379192243219, "grad_norm": 4.088005065917969, "learning_rate": 1.7448760944768076e-05, "loss": 4.0863, "step": 981350 }, { "epoch": 5.917680684024553, "grad_norm": 3.5655901432037354, "learning_rate": 1.7439236791792876e-05, "loss": 4.0254, "step": 981400 }, { "epoch": 5.9179821758058875, "grad_norm": 4.318368911743164, "learning_rate": 1.7429715078409374e-05, "loss": 4.7363, "step": 981450 }, { "epoch": 5.918283667587222, "grad_norm": 3.561112403869629, "learning_rate": 1.742019580479276e-05, "loss": 4.5635, "step": 981500 }, { "epoch": 5.918585159368556, "grad_norm": 2.857628345489502, "learning_rate": 1.7410678971118225e-05, "loss": 4.3043, "step": 981550 }, { "epoch": 5.91888665114989, "grad_norm": 4.285181999206543, "learning_rate": 1.7401164577560965e-05, "loss": 3.8447, "step": 981600 }, { "epoch": 5.919188142931223, "grad_norm": 6.119715690612793, "learning_rate": 1.739165262429601e-05, "loss": 4.4066, "step": 981650 }, { "epoch": 5.919489634712558, "grad_norm": 6.480530738830566, "learning_rate": 1.7382143111498482e-05, "loss": 4.411, "step": 981700 }, { "epoch": 5.919791126493892, "grad_norm": 5.138076305389404, "learning_rate": 1.7372636039343358e-05, "loss": 4.467, "step": 981750 }, { "epoch": 5.920092618275226, "grad_norm": 5.479776382446289, "learning_rate": 1.7363131408005564e-05, "loss": 4.5322, "step": 981800 }, { "epoch": 5.92039411005656, "grad_norm": 2.6812644004821777, "learning_rate": 1.735362921766005e-05, "loss": 4.5968, "step": 981850 }, { "epoch": 5.920695601837894, "grad_norm": 7.683592319488525, "learning_rate": 1.734412946848173e-05, "loss": 4.4552, "step": 981900 }, { "epoch": 5.920997093619228, "grad_norm": 2.617466926574707, "learning_rate": 1.733463216064536e-05, "loss": 4.2521, "step": 981950 }, { "epoch": 5.921298585400562, "grad_norm": 4.545928001403809, "learning_rate": 1.7325137294325764e-05, "loss": 4.3013, "step": 982000 }, { "epoch": 5.921600077181896, "grad_norm": 4.507788181304932, "learning_rate": 1.7315644869697714e-05, "loss": 4.5055, "step": 982050 }, { "epoch": 5.92190156896323, "grad_norm": 5.205670356750488, "learning_rate": 1.7306154886935858e-05, "loss": 4.3246, "step": 982100 }, { "epoch": 5.922203060744565, "grad_norm": 3.432572364807129, "learning_rate": 1.7296667346214837e-05, "loss": 4.5164, "step": 982150 }, { "epoch": 5.922504552525898, "grad_norm": 3.220604419708252, "learning_rate": 1.7287182247709297e-05, "loss": 4.4863, "step": 982200 }, { "epoch": 5.922806044307232, "grad_norm": 1.8091800212860107, "learning_rate": 1.7277699591593746e-05, "loss": 4.3501, "step": 982250 }, { "epoch": 5.923107536088566, "grad_norm": 3.97472882270813, "learning_rate": 1.7268219378042748e-05, "loss": 4.7734, "step": 982300 }, { "epoch": 5.9234090278699005, "grad_norm": 3.088085889816284, "learning_rate": 1.7258741607230757e-05, "loss": 4.106, "step": 982350 }, { "epoch": 5.923710519651234, "grad_norm": 2.0167808532714844, "learning_rate": 1.7249266279332192e-05, "loss": 4.2896, "step": 982400 }, { "epoch": 5.924012011432568, "grad_norm": 4.1644744873046875, "learning_rate": 1.7239793394521423e-05, "loss": 4.5449, "step": 982450 }, { "epoch": 5.924313503213902, "grad_norm": 5.064780235290527, "learning_rate": 1.7230322952972815e-05, "loss": 4.5023, "step": 982500 }, { "epoch": 5.9246149949952365, "grad_norm": 2.3607900142669678, "learning_rate": 1.722085495486066e-05, "loss": 4.3721, "step": 982550 }, { "epoch": 5.924916486776571, "grad_norm": 4.251748561859131, "learning_rate": 1.7211389400359148e-05, "loss": 4.1203, "step": 982600 }, { "epoch": 5.925217978557905, "grad_norm": 5.345974445343018, "learning_rate": 1.7201926289642516e-05, "loss": 4.5387, "step": 982650 }, { "epoch": 5.925519470339238, "grad_norm": 3.8777382373809814, "learning_rate": 1.7192465622884932e-05, "loss": 4.3898, "step": 982700 }, { "epoch": 5.925820962120572, "grad_norm": 7.693628787994385, "learning_rate": 1.718300740026048e-05, "loss": 4.408, "step": 982750 }, { "epoch": 5.926122453901907, "grad_norm": 4.4818501472473145, "learning_rate": 1.7173551621943232e-05, "loss": 4.4388, "step": 982800 }, { "epoch": 5.926423945683241, "grad_norm": 4.363343238830566, "learning_rate": 1.7164098288107237e-05, "loss": 3.955, "step": 982850 }, { "epoch": 5.926725437464575, "grad_norm": 4.755974292755127, "learning_rate": 1.7154647398926452e-05, "loss": 4.6776, "step": 982900 }, { "epoch": 5.927026929245908, "grad_norm": 3.9981112480163574, "learning_rate": 1.7145198954574772e-05, "loss": 4.0979, "step": 982950 }, { "epoch": 5.9273284210272426, "grad_norm": 4.182511806488037, "learning_rate": 1.7135752955226144e-05, "loss": 4.2929, "step": 983000 }, { "epoch": 5.927629912808577, "grad_norm": 5.344959259033203, "learning_rate": 1.712630940105433e-05, "loss": 4.1937, "step": 983050 }, { "epoch": 5.927931404589911, "grad_norm": 5.125295162200928, "learning_rate": 1.711686829223318e-05, "loss": 3.9247, "step": 983100 }, { "epoch": 5.928232896371245, "grad_norm": 3.733581304550171, "learning_rate": 1.710742962893647e-05, "loss": 4.0532, "step": 983150 }, { "epoch": 5.928534388152579, "grad_norm": 3.448859453201294, "learning_rate": 1.7097993411337836e-05, "loss": 4.0802, "step": 983200 }, { "epoch": 5.928835879933913, "grad_norm": 6.201766014099121, "learning_rate": 1.7088559639611006e-05, "loss": 4.5719, "step": 983250 }, { "epoch": 5.929137371715247, "grad_norm": 3.2892098426818848, "learning_rate": 1.7079128313929557e-05, "loss": 4.0843, "step": 983300 }, { "epoch": 5.929438863496581, "grad_norm": 3.8852992057800293, "learning_rate": 1.706969943446702e-05, "loss": 4.2819, "step": 983350 }, { "epoch": 5.929740355277915, "grad_norm": 5.375982761383057, "learning_rate": 1.7060273001396974e-05, "loss": 4.1146, "step": 983400 }, { "epoch": 5.9300418470592495, "grad_norm": 3.166733980178833, "learning_rate": 1.7050849014892926e-05, "loss": 4.3817, "step": 983450 }, { "epoch": 5.930343338840583, "grad_norm": 3.2019219398498535, "learning_rate": 1.7041427475128237e-05, "loss": 4.556, "step": 983500 }, { "epoch": 5.930644830621917, "grad_norm": 4.642408847808838, "learning_rate": 1.7032008382276324e-05, "loss": 4.4375, "step": 983550 }, { "epoch": 5.930946322403251, "grad_norm": 2.4354469776153564, "learning_rate": 1.7022591736510595e-05, "loss": 4.2297, "step": 983600 }, { "epoch": 5.9312478141845855, "grad_norm": 3.4762370586395264, "learning_rate": 1.701317753800424e-05, "loss": 3.9697, "step": 983650 }, { "epoch": 5.93154930596592, "grad_norm": 4.416313648223877, "learning_rate": 1.7003765786930612e-05, "loss": 4.4302, "step": 983700 }, { "epoch": 5.931850797747254, "grad_norm": 4.07663631439209, "learning_rate": 1.6994356483462883e-05, "loss": 4.4809, "step": 983750 }, { "epoch": 5.932152289528587, "grad_norm": 2.3962314128875732, "learning_rate": 1.6984949627774184e-05, "loss": 4.4479, "step": 983800 }, { "epoch": 5.932453781309921, "grad_norm": 5.871977806091309, "learning_rate": 1.6975545220037662e-05, "loss": 4.3919, "step": 983850 }, { "epoch": 5.932755273091256, "grad_norm": 6.079034328460693, "learning_rate": 1.6966143260426406e-05, "loss": 4.5158, "step": 983900 }, { "epoch": 5.93305676487259, "grad_norm": 5.21881628036499, "learning_rate": 1.6956743749113432e-05, "loss": 4.2125, "step": 983950 }, { "epoch": 5.933358256653923, "grad_norm": 2.722011089324951, "learning_rate": 1.6947346686271717e-05, "loss": 4.6346, "step": 984000 }, { "epoch": 5.933358256653923, "eval_loss": 4.855856418609619, "eval_runtime": 39.0536, "eval_samples_per_second": 13.11, "eval_steps_per_second": 6.555, "eval_tts_loss": 7.987743214955928, "step": 984000 }, { "epoch": 5.933659748435257, "grad_norm": 5.823754787445068, "learning_rate": 1.6937952072074225e-05, "loss": 4.4732, "step": 984050 }, { "epoch": 5.9339612402165915, "grad_norm": 3.8477163314819336, "learning_rate": 1.6928559906693845e-05, "loss": 4.4317, "step": 984100 }, { "epoch": 5.934262731997926, "grad_norm": 4.374720096588135, "learning_rate": 1.6919170190303394e-05, "loss": 4.343, "step": 984150 }, { "epoch": 5.93456422377926, "grad_norm": 4.283677101135254, "learning_rate": 1.6909782923075733e-05, "loss": 4.6846, "step": 984200 }, { "epoch": 5.934865715560594, "grad_norm": 4.865111827850342, "learning_rate": 1.6900398105183538e-05, "loss": 4.3425, "step": 984250 }, { "epoch": 5.9351672073419275, "grad_norm": 6.035644054412842, "learning_rate": 1.689101573679959e-05, "loss": 4.6332, "step": 984300 }, { "epoch": 5.935468699123262, "grad_norm": 5.773984432220459, "learning_rate": 1.6881635818096567e-05, "loss": 4.7129, "step": 984350 }, { "epoch": 5.935770190904596, "grad_norm": 4.058032035827637, "learning_rate": 1.6872258349247032e-05, "loss": 4.4351, "step": 984400 }, { "epoch": 5.93607168268593, "grad_norm": 3.9491403102874756, "learning_rate": 1.6862883330423627e-05, "loss": 4.444, "step": 984450 }, { "epoch": 5.936373174467264, "grad_norm": 4.182639122009277, "learning_rate": 1.6853510761798866e-05, "loss": 4.5612, "step": 984500 }, { "epoch": 5.936674666248598, "grad_norm": 5.851808071136475, "learning_rate": 1.6844140643545177e-05, "loss": 4.2385, "step": 984550 }, { "epoch": 5.936976158029932, "grad_norm": 7.304453372955322, "learning_rate": 1.6834772975835075e-05, "loss": 4.548, "step": 984600 }, { "epoch": 5.937277649811266, "grad_norm": 3.331996202468872, "learning_rate": 1.682540775884097e-05, "loss": 4.5201, "step": 984650 }, { "epoch": 5.9375791415926, "grad_norm": 3.693697452545166, "learning_rate": 1.6816044992735157e-05, "loss": 3.9129, "step": 984700 }, { "epoch": 5.9378806333739345, "grad_norm": 6.3998003005981445, "learning_rate": 1.6806684677689962e-05, "loss": 4.5642, "step": 984750 }, { "epoch": 5.938182125155269, "grad_norm": 1.3210731744766235, "learning_rate": 1.6797326813877686e-05, "loss": 4.0838, "step": 984800 }, { "epoch": 5.938483616936602, "grad_norm": 5.2584452629089355, "learning_rate": 1.6787971401470506e-05, "loss": 4.5651, "step": 984850 }, { "epoch": 5.938785108717936, "grad_norm": 3.5999443531036377, "learning_rate": 1.6778618440640635e-05, "loss": 4.0006, "step": 984900 }, { "epoch": 5.93908660049927, "grad_norm": 3.313828468322754, "learning_rate": 1.6769267931560166e-05, "loss": 4.4076, "step": 984950 }, { "epoch": 5.939388092280605, "grad_norm": 5.141165733337402, "learning_rate": 1.6759919874401162e-05, "loss": 4.3958, "step": 985000 }, { "epoch": 5.939689584061939, "grad_norm": 3.620471477508545, "learning_rate": 1.675057426933569e-05, "loss": 4.3199, "step": 985050 }, { "epoch": 5.939991075843272, "grad_norm": 4.46602725982666, "learning_rate": 1.674123111653577e-05, "loss": 4.1472, "step": 985100 }, { "epoch": 5.940292567624606, "grad_norm": 4.13111686706543, "learning_rate": 1.673189041617329e-05, "loss": 4.2969, "step": 985150 }, { "epoch": 5.9405940594059405, "grad_norm": 7.955294132232666, "learning_rate": 1.672255216842019e-05, "loss": 4.1907, "step": 985200 }, { "epoch": 5.940895551187275, "grad_norm": 3.5574820041656494, "learning_rate": 1.6713216373448334e-05, "loss": 4.4031, "step": 985250 }, { "epoch": 5.941197042968609, "grad_norm": 5.530332088470459, "learning_rate": 1.6703883031429538e-05, "loss": 4.57, "step": 985300 }, { "epoch": 5.941498534749943, "grad_norm": 5.479222774505615, "learning_rate": 1.6694552142535512e-05, "loss": 4.1012, "step": 985350 }, { "epoch": 5.9418000265312765, "grad_norm": 0.8556482195854187, "learning_rate": 1.668522370693805e-05, "loss": 4.2739, "step": 985400 }, { "epoch": 5.942101518312611, "grad_norm": 3.5062711238861084, "learning_rate": 1.667589772480877e-05, "loss": 4.7907, "step": 985450 }, { "epoch": 5.942403010093945, "grad_norm": 7.096410751342773, "learning_rate": 1.6666574196319326e-05, "loss": 4.4038, "step": 985500 }, { "epoch": 5.942704501875279, "grad_norm": 2.763983726501465, "learning_rate": 1.6657253121641323e-05, "loss": 3.762, "step": 985550 }, { "epoch": 5.943005993656613, "grad_norm": 3.3564116954803467, "learning_rate": 1.664793450094627e-05, "loss": 4.1686, "step": 985600 }, { "epoch": 5.943307485437947, "grad_norm": 5.27370023727417, "learning_rate": 1.6638618334405714e-05, "loss": 4.4376, "step": 985650 }, { "epoch": 5.943608977219281, "grad_norm": 3.87634539604187, "learning_rate": 1.662930462219103e-05, "loss": 4.3969, "step": 985700 }, { "epoch": 5.943910469000615, "grad_norm": 3.5320935249328613, "learning_rate": 1.6619993364473704e-05, "loss": 4.2671, "step": 985750 }, { "epoch": 5.944211960781949, "grad_norm": 4.683406829833984, "learning_rate": 1.6610684561425014e-05, "loss": 4.4335, "step": 985800 }, { "epoch": 5.9445134525632835, "grad_norm": 4.912902355194092, "learning_rate": 1.6601378213216353e-05, "loss": 4.4489, "step": 985850 }, { "epoch": 5.944814944344618, "grad_norm": 4.49929141998291, "learning_rate": 1.6592074320018923e-05, "loss": 4.4944, "step": 985900 }, { "epoch": 5.945116436125951, "grad_norm": 4.372183799743652, "learning_rate": 1.658277288200398e-05, "loss": 4.258, "step": 985950 }, { "epoch": 5.945417927907285, "grad_norm": 3.678820848464966, "learning_rate": 1.6573473899342727e-05, "loss": 4.3335, "step": 986000 }, { "epoch": 5.945719419688619, "grad_norm": 3.970493793487549, "learning_rate": 1.6564177372206287e-05, "loss": 4.3121, "step": 986050 }, { "epoch": 5.946020911469954, "grad_norm": 5.301519393920898, "learning_rate": 1.6554883300765746e-05, "loss": 4.4253, "step": 986100 }, { "epoch": 5.946322403251287, "grad_norm": 4.33554744720459, "learning_rate": 1.65455916851921e-05, "loss": 3.9962, "step": 986150 }, { "epoch": 5.946623895032621, "grad_norm": 4.595118999481201, "learning_rate": 1.6536302525656424e-05, "loss": 4.5417, "step": 986200 }, { "epoch": 5.946925386813955, "grad_norm": 1.2562936544418335, "learning_rate": 1.652701582232962e-05, "loss": 3.8047, "step": 986250 }, { "epoch": 5.9472268785952895, "grad_norm": 4.313316822052002, "learning_rate": 1.6517731575382597e-05, "loss": 4.3016, "step": 986300 }, { "epoch": 5.947528370376624, "grad_norm": 3.8136582374572754, "learning_rate": 1.6508449784986273e-05, "loss": 4.2638, "step": 986350 }, { "epoch": 5.947829862157958, "grad_norm": 2.1962969303131104, "learning_rate": 1.6499170451311407e-05, "loss": 4.3837, "step": 986400 }, { "epoch": 5.948131353939291, "grad_norm": 3.840430498123169, "learning_rate": 1.648989357452881e-05, "loss": 4.4993, "step": 986450 }, { "epoch": 5.9484328457206255, "grad_norm": 4.555734634399414, "learning_rate": 1.6480619154809188e-05, "loss": 4.4313, "step": 986500 }, { "epoch": 5.94873433750196, "grad_norm": 4.005029678344727, "learning_rate": 1.6471347192323213e-05, "loss": 4.301, "step": 986550 }, { "epoch": 5.949035829283294, "grad_norm": 3.864119052886963, "learning_rate": 1.6462077687241537e-05, "loss": 4.1651, "step": 986600 }, { "epoch": 5.949337321064628, "grad_norm": 1.0110496282577515, "learning_rate": 1.645281063973477e-05, "loss": 4.236, "step": 986650 }, { "epoch": 5.949638812845961, "grad_norm": 3.6490836143493652, "learning_rate": 1.644354604997341e-05, "loss": 4.2603, "step": 986700 }, { "epoch": 5.949940304627296, "grad_norm": 3.917797565460205, "learning_rate": 1.6434283918128003e-05, "loss": 4.2428, "step": 986750 }, { "epoch": 5.95024179640863, "grad_norm": 2.260329484939575, "learning_rate": 1.6425024244369017e-05, "loss": 4.5523, "step": 986800 }, { "epoch": 5.950543288189964, "grad_norm": 3.5358176231384277, "learning_rate": 1.641576702886681e-05, "loss": 4.4912, "step": 986850 }, { "epoch": 5.950844779971298, "grad_norm": 3.845012903213501, "learning_rate": 1.64065122717918e-05, "loss": 4.5595, "step": 986900 }, { "epoch": 5.951146271752632, "grad_norm": 4.13739538192749, "learning_rate": 1.639725997331428e-05, "loss": 4.4537, "step": 986950 }, { "epoch": 5.951447763533966, "grad_norm": 1.5298658609390259, "learning_rate": 1.6388010133604516e-05, "loss": 4.2804, "step": 987000 }, { "epoch": 5.951447763533966, "eval_loss": 4.852077007293701, "eval_runtime": 39.1757, "eval_samples_per_second": 13.069, "eval_steps_per_second": 6.535, "eval_tts_loss": 7.965365776284553, "step": 987000 }, { "epoch": 5.9517492553153, "grad_norm": 5.700860977172852, "learning_rate": 1.6378762752832742e-05, "loss": 4.201, "step": 987050 }, { "epoch": 5.952050747096634, "grad_norm": 5.039409637451172, "learning_rate": 1.6369517831169165e-05, "loss": 4.492, "step": 987100 }, { "epoch": 5.952352238877968, "grad_norm": 8.774152755737305, "learning_rate": 1.6360275368783903e-05, "loss": 4.5169, "step": 987150 }, { "epoch": 5.952653730659303, "grad_norm": 3.7785239219665527, "learning_rate": 1.6351035365847054e-05, "loss": 4.4584, "step": 987200 }, { "epoch": 5.952955222440636, "grad_norm": 4.303296089172363, "learning_rate": 1.6341797822528696e-05, "loss": 4.2743, "step": 987250 }, { "epoch": 5.95325671422197, "grad_norm": 7.122345447540283, "learning_rate": 1.633256273899881e-05, "loss": 4.5928, "step": 987300 }, { "epoch": 5.953558206003304, "grad_norm": 1.378609299659729, "learning_rate": 1.632333011542733e-05, "loss": 3.8615, "step": 987350 }, { "epoch": 5.9538596977846385, "grad_norm": 2.693561315536499, "learning_rate": 1.6314099951984216e-05, "loss": 4.9344, "step": 987400 }, { "epoch": 5.954161189565973, "grad_norm": 3.0576956272125244, "learning_rate": 1.6304872248839283e-05, "loss": 4.5576, "step": 987450 }, { "epoch": 5.954462681347307, "grad_norm": 7.83306360244751, "learning_rate": 1.6295647006162384e-05, "loss": 4.4051, "step": 987500 }, { "epoch": 5.95476417312864, "grad_norm": 3.922938346862793, "learning_rate": 1.6286424224123325e-05, "loss": 4.5932, "step": 987550 }, { "epoch": 5.9550656649099745, "grad_norm": 5.745452880859375, "learning_rate": 1.6277203902891787e-05, "loss": 4.2891, "step": 987600 }, { "epoch": 5.955367156691309, "grad_norm": 3.026470899581909, "learning_rate": 1.6267986042637493e-05, "loss": 4.1504, "step": 987650 }, { "epoch": 5.955668648472643, "grad_norm": 4.864970684051514, "learning_rate": 1.6258770643530082e-05, "loss": 4.7483, "step": 987700 }, { "epoch": 5.955970140253976, "grad_norm": 4.343578815460205, "learning_rate": 1.6249557705739107e-05, "loss": 4.4912, "step": 987750 }, { "epoch": 5.95627163203531, "grad_norm": 4.930344581604004, "learning_rate": 1.6240347229434147e-05, "loss": 4.2561, "step": 987800 }, { "epoch": 5.956573123816645, "grad_norm": 3.3216965198516846, "learning_rate": 1.6231139214784734e-05, "loss": 4.3186, "step": 987850 }, { "epoch": 5.956874615597979, "grad_norm": 6.3943610191345215, "learning_rate": 1.622193366196028e-05, "loss": 4.7247, "step": 987900 }, { "epoch": 5.957176107379313, "grad_norm": 3.924572706222534, "learning_rate": 1.6212730571130234e-05, "loss": 4.3627, "step": 987950 }, { "epoch": 5.957477599160647, "grad_norm": 3.535759687423706, "learning_rate": 1.6203529942463976e-05, "loss": 4.586, "step": 988000 }, { "epoch": 5.9577790909419805, "grad_norm": 3.7735462188720703, "learning_rate": 1.6194331776130794e-05, "loss": 4.4591, "step": 988050 }, { "epoch": 5.958080582723315, "grad_norm": 4.209403991699219, "learning_rate": 1.618513607230001e-05, "loss": 4.3401, "step": 988100 }, { "epoch": 5.958382074504649, "grad_norm": 5.095142841339111, "learning_rate": 1.6175942831140827e-05, "loss": 4.3783, "step": 988150 }, { "epoch": 5.958683566285983, "grad_norm": 4.375344753265381, "learning_rate": 1.6166752052822406e-05, "loss": 4.6849, "step": 988200 }, { "epoch": 5.958985058067317, "grad_norm": 5.456417560577393, "learning_rate": 1.6157563737513933e-05, "loss": 4.4215, "step": 988250 }, { "epoch": 5.959286549848651, "grad_norm": 4.7994704246521, "learning_rate": 1.6148377885384522e-05, "loss": 3.9047, "step": 988300 }, { "epoch": 5.959588041629985, "grad_norm": 3.6346511840820312, "learning_rate": 1.613919449660317e-05, "loss": 4.6213, "step": 988350 }, { "epoch": 5.959889533411319, "grad_norm": 4.279477596282959, "learning_rate": 1.6130013571338924e-05, "loss": 4.2053, "step": 988400 }, { "epoch": 5.960191025192653, "grad_norm": 2.142956495285034, "learning_rate": 1.6120835109760747e-05, "loss": 4.4376, "step": 988450 }, { "epoch": 5.9604925169739875, "grad_norm": 1.099271297454834, "learning_rate": 1.6111659112037544e-05, "loss": 4.4551, "step": 988500 }, { "epoch": 5.960794008755322, "grad_norm": 3.549899101257324, "learning_rate": 1.6102485578338175e-05, "loss": 4.52, "step": 988550 }, { "epoch": 5.961095500536655, "grad_norm": 5.4975152015686035, "learning_rate": 1.609331450883149e-05, "loss": 4.4217, "step": 988600 }, { "epoch": 5.961396992317989, "grad_norm": 3.493776798248291, "learning_rate": 1.6084145903686235e-05, "loss": 4.4285, "step": 988650 }, { "epoch": 5.9616984840993235, "grad_norm": 4.070940017700195, "learning_rate": 1.6074979763071162e-05, "loss": 4.0859, "step": 988700 }, { "epoch": 5.961999975880658, "grad_norm": 5.126143932342529, "learning_rate": 1.6065816087155e-05, "loss": 4.6025, "step": 988750 }, { "epoch": 5.962301467661992, "grad_norm": 4.848209857940674, "learning_rate": 1.605665487610633e-05, "loss": 4.6115, "step": 988800 }, { "epoch": 5.962602959443325, "grad_norm": 5.484229564666748, "learning_rate": 1.6047496130093802e-05, "loss": 4.4974, "step": 988850 }, { "epoch": 5.962904451224659, "grad_norm": 6.276095390319824, "learning_rate": 1.603833984928595e-05, "loss": 4.0435, "step": 988900 }, { "epoch": 5.963205943005994, "grad_norm": 4.137619972229004, "learning_rate": 1.602918603385125e-05, "loss": 4.4588, "step": 988950 }, { "epoch": 5.963507434787328, "grad_norm": 3.935830593109131, "learning_rate": 1.602003468395819e-05, "loss": 4.3982, "step": 989000 }, { "epoch": 5.963808926568662, "grad_norm": 4.451808929443359, "learning_rate": 1.601088579977523e-05, "loss": 4.5893, "step": 989050 }, { "epoch": 5.964110418349996, "grad_norm": 3.901951313018799, "learning_rate": 1.6001739381470673e-05, "loss": 4.1509, "step": 989100 }, { "epoch": 5.9644119101313295, "grad_norm": 6.705533504486084, "learning_rate": 1.5992595429212868e-05, "loss": 4.6315, "step": 989150 }, { "epoch": 5.964713401912664, "grad_norm": 4.1472063064575195, "learning_rate": 1.5983453943170145e-05, "loss": 4.5208, "step": 989200 }, { "epoch": 5.965014893693998, "grad_norm": 6.522915363311768, "learning_rate": 1.5974314923510667e-05, "loss": 4.4579, "step": 989250 }, { "epoch": 5.965316385475332, "grad_norm": 3.625149726867676, "learning_rate": 1.5965178370402687e-05, "loss": 4.198, "step": 989300 }, { "epoch": 5.965617877256666, "grad_norm": 6.149247646331787, "learning_rate": 1.5956044284014286e-05, "loss": 4.4097, "step": 989350 }, { "epoch": 5.965919369038, "grad_norm": 4.973310947418213, "learning_rate": 1.5946912664513645e-05, "loss": 4.434, "step": 989400 }, { "epoch": 5.966220860819334, "grad_norm": 3.6079025268554688, "learning_rate": 1.593778351206873e-05, "loss": 4.4053, "step": 989450 }, { "epoch": 5.966522352600668, "grad_norm": 3.9501283168792725, "learning_rate": 1.592865682684759e-05, "loss": 4.4658, "step": 989500 }, { "epoch": 5.966823844382002, "grad_norm": 3.9704201221466064, "learning_rate": 1.5919532609018227e-05, "loss": 4.0608, "step": 989550 }, { "epoch": 5.9671253361633365, "grad_norm": 6.084621429443359, "learning_rate": 1.59104108587485e-05, "loss": 4.4965, "step": 989600 }, { "epoch": 5.967426827944671, "grad_norm": 3.6511309146881104, "learning_rate": 1.5901291576206314e-05, "loss": 4.1209, "step": 989650 }, { "epoch": 5.967728319726004, "grad_norm": 3.7535548210144043, "learning_rate": 1.5892174761559502e-05, "loss": 4.6403, "step": 989700 }, { "epoch": 5.968029811507338, "grad_norm": 3.91623592376709, "learning_rate": 1.5883060414975808e-05, "loss": 4.0133, "step": 989750 }, { "epoch": 5.9683313032886725, "grad_norm": 5.76961088180542, "learning_rate": 1.5873948536622984e-05, "loss": 4.1616, "step": 989800 }, { "epoch": 5.968632795070007, "grad_norm": 6.561680793762207, "learning_rate": 1.586483912666876e-05, "loss": 4.6122, "step": 989850 }, { "epoch": 5.96893428685134, "grad_norm": 2.5355467796325684, "learning_rate": 1.5855732185280707e-05, "loss": 4.5249, "step": 989900 }, { "epoch": 5.969235778632674, "grad_norm": 1.5499279499053955, "learning_rate": 1.5846627712626487e-05, "loss": 4.2111, "step": 989950 }, { "epoch": 5.969537270414008, "grad_norm": 3.7170636653900146, "learning_rate": 1.5837525708873665e-05, "loss": 4.4478, "step": 990000 }, { "epoch": 5.969537270414008, "eval_loss": 4.850494861602783, "eval_runtime": 39.288, "eval_samples_per_second": 13.032, "eval_steps_per_second": 6.516, "eval_tts_loss": 7.960781312021543, "step": 990000 }, { "epoch": 5.969838762195343, "grad_norm": 3.0269386768341064, "learning_rate": 1.5828426174189714e-05, "loss": 4.5811, "step": 990050 }, { "epoch": 5.970140253976677, "grad_norm": 3.7035162448883057, "learning_rate": 1.5819329108742092e-05, "loss": 4.5569, "step": 990100 }, { "epoch": 5.970441745758011, "grad_norm": 3.5687313079833984, "learning_rate": 1.5810234512698234e-05, "loss": 4.3068, "step": 990150 }, { "epoch": 5.970743237539344, "grad_norm": 1.7919840812683105, "learning_rate": 1.5801142386225508e-05, "loss": 4.257, "step": 990200 }, { "epoch": 5.9710447293206785, "grad_norm": 3.3785464763641357, "learning_rate": 1.5792052729491228e-05, "loss": 4.7684, "step": 990250 }, { "epoch": 5.971346221102013, "grad_norm": 3.8558712005615234, "learning_rate": 1.578296554266273e-05, "loss": 4.5341, "step": 990300 }, { "epoch": 5.971647712883347, "grad_norm": 4.192984580993652, "learning_rate": 1.5773880825907177e-05, "loss": 3.979, "step": 990350 }, { "epoch": 5.971949204664681, "grad_norm": 3.8637051582336426, "learning_rate": 1.5764798579391803e-05, "loss": 4.2934, "step": 990400 }, { "epoch": 5.9722506964460145, "grad_norm": 4.553633213043213, "learning_rate": 1.5755718803283775e-05, "loss": 4.4553, "step": 990450 }, { "epoch": 5.972552188227349, "grad_norm": 8.713153839111328, "learning_rate": 1.574664149775017e-05, "loss": 4.2434, "step": 990500 }, { "epoch": 5.972853680008683, "grad_norm": 3.9884674549102783, "learning_rate": 1.5737566662958e-05, "loss": 4.2641, "step": 990550 }, { "epoch": 5.973155171790017, "grad_norm": 4.883981227874756, "learning_rate": 1.572849429907434e-05, "loss": 4.5184, "step": 990600 }, { "epoch": 5.973456663571351, "grad_norm": 5.061344146728516, "learning_rate": 1.5719424406266105e-05, "loss": 3.9951, "step": 990650 }, { "epoch": 5.9737581553526855, "grad_norm": 5.212470531463623, "learning_rate": 1.5710356984700234e-05, "loss": 4.5526, "step": 990700 }, { "epoch": 5.974059647134019, "grad_norm": 6.275105953216553, "learning_rate": 1.570129203454364e-05, "loss": 4.6688, "step": 990750 }, { "epoch": 5.974361138915353, "grad_norm": 4.091090679168701, "learning_rate": 1.569222955596307e-05, "loss": 4.5968, "step": 990800 }, { "epoch": 5.974662630696687, "grad_norm": 2.2158865928649902, "learning_rate": 1.568316954912538e-05, "loss": 4.647, "step": 990850 }, { "epoch": 5.9749641224780214, "grad_norm": 3.5995993614196777, "learning_rate": 1.5674112014197264e-05, "loss": 4.0982, "step": 990900 }, { "epoch": 5.975265614259356, "grad_norm": 3.7306716442108154, "learning_rate": 1.566505695134541e-05, "loss": 4.3589, "step": 990950 }, { "epoch": 5.975567106040689, "grad_norm": 4.256673336029053, "learning_rate": 1.5656004360736467e-05, "loss": 4.1091, "step": 991000 }, { "epoch": 5.975868597822023, "grad_norm": 2.236137628555298, "learning_rate": 1.5646954242537068e-05, "loss": 4.3811, "step": 991050 }, { "epoch": 5.976170089603357, "grad_norm": 2.2953875064849854, "learning_rate": 1.5637906596913728e-05, "loss": 4.4529, "step": 991100 }, { "epoch": 5.976471581384692, "grad_norm": 3.5711984634399414, "learning_rate": 1.562886142403297e-05, "loss": 4.2022, "step": 991150 }, { "epoch": 5.976773073166026, "grad_norm": 2.9642186164855957, "learning_rate": 1.5619818724061285e-05, "loss": 4.0199, "step": 991200 }, { "epoch": 5.97707456494736, "grad_norm": 5.4433698654174805, "learning_rate": 1.561077849716506e-05, "loss": 4.5773, "step": 991250 }, { "epoch": 5.977376056728693, "grad_norm": 2.6853678226470947, "learning_rate": 1.5601740743510667e-05, "loss": 4.5804, "step": 991300 }, { "epoch": 5.9776775485100275, "grad_norm": 4.278421878814697, "learning_rate": 1.5592705463264448e-05, "loss": 4.276, "step": 991350 }, { "epoch": 5.977979040291362, "grad_norm": 2.5117762088775635, "learning_rate": 1.5583672656592644e-05, "loss": 4.7697, "step": 991400 }, { "epoch": 5.978280532072696, "grad_norm": 5.007069110870361, "learning_rate": 1.5574642323661535e-05, "loss": 4.2545, "step": 991450 }, { "epoch": 5.978582023854029, "grad_norm": 4.494686603546143, "learning_rate": 1.556561446463732e-05, "loss": 4.4222, "step": 991500 }, { "epoch": 5.9788835156353635, "grad_norm": 2.8914852142333984, "learning_rate": 1.5556589079686088e-05, "loss": 4.3614, "step": 991550 }, { "epoch": 5.979185007416698, "grad_norm": 3.5104262828826904, "learning_rate": 1.5547566168973986e-05, "loss": 4.0526, "step": 991600 }, { "epoch": 5.979486499198032, "grad_norm": 6.872646808624268, "learning_rate": 1.553854573266708e-05, "loss": 4.2075, "step": 991650 }, { "epoch": 5.979787990979366, "grad_norm": 5.2922563552856445, "learning_rate": 1.5529527770931338e-05, "loss": 4.3523, "step": 991700 }, { "epoch": 5.9800894827607, "grad_norm": 6.209910869598389, "learning_rate": 1.552051228393273e-05, "loss": 4.2409, "step": 991750 }, { "epoch": 5.980390974542034, "grad_norm": 5.45849609375, "learning_rate": 1.5511499271837202e-05, "loss": 4.5669, "step": 991800 }, { "epoch": 5.980692466323368, "grad_norm": 6.278318881988525, "learning_rate": 1.5502488734810575e-05, "loss": 4.1436, "step": 991850 }, { "epoch": 5.980993958104702, "grad_norm": 4.838593482971191, "learning_rate": 1.54934806730187e-05, "loss": 4.0099, "step": 991900 }, { "epoch": 5.981295449886036, "grad_norm": 6.5713276863098145, "learning_rate": 1.548447508662739e-05, "loss": 4.4775, "step": 991950 }, { "epoch": 5.98159694166737, "grad_norm": 4.890938758850098, "learning_rate": 1.5475471975802333e-05, "loss": 4.2115, "step": 992000 }, { "epoch": 5.981898433448704, "grad_norm": 2.0598971843719482, "learning_rate": 1.5466471340709265e-05, "loss": 4.2649, "step": 992050 }, { "epoch": 5.982199925230038, "grad_norm": 5.156960964202881, "learning_rate": 1.5457473181513796e-05, "loss": 4.4842, "step": 992100 }, { "epoch": 5.982501417011372, "grad_norm": 3.221560001373291, "learning_rate": 1.54484774983815e-05, "loss": 4.3175, "step": 992150 }, { "epoch": 5.982802908792706, "grad_norm": 5.130057334899902, "learning_rate": 1.5439484291477962e-05, "loss": 4.5291, "step": 992200 }, { "epoch": 5.983104400574041, "grad_norm": 3.5411083698272705, "learning_rate": 1.5430493560968726e-05, "loss": 4.3512, "step": 992250 }, { "epoch": 5.983405892355375, "grad_norm": 5.287300109863281, "learning_rate": 1.5421505307019182e-05, "loss": 4.2367, "step": 992300 }, { "epoch": 5.983707384136708, "grad_norm": 4.298800468444824, "learning_rate": 1.5412519529794765e-05, "loss": 4.7376, "step": 992350 }, { "epoch": 5.984008875918042, "grad_norm": 4.356636047363281, "learning_rate": 1.5403536229460905e-05, "loss": 4.56, "step": 992400 }, { "epoch": 5.9843103676993765, "grad_norm": 7.668010711669922, "learning_rate": 1.5394555406182856e-05, "loss": 3.8775, "step": 992450 }, { "epoch": 5.984611859480711, "grad_norm": 3.304849863052368, "learning_rate": 1.5385577060125937e-05, "loss": 4.272, "step": 992500 }, { "epoch": 5.984913351262045, "grad_norm": 4.285327911376953, "learning_rate": 1.537660119145538e-05, "loss": 4.2755, "step": 992550 }, { "epoch": 5.985214843043378, "grad_norm": 4.553359031677246, "learning_rate": 1.5367627800336318e-05, "loss": 3.926, "step": 992600 }, { "epoch": 5.9855163348247125, "grad_norm": 4.946171760559082, "learning_rate": 1.535865688693394e-05, "loss": 4.426, "step": 992650 }, { "epoch": 5.985817826606047, "grad_norm": 3.724978446960449, "learning_rate": 1.5349688451413345e-05, "loss": 4.8241, "step": 992700 }, { "epoch": 5.986119318387381, "grad_norm": 1.976184606552124, "learning_rate": 1.53407224939396e-05, "loss": 4.0527, "step": 992750 }, { "epoch": 5.986420810168715, "grad_norm": 5.92244291305542, "learning_rate": 1.5331759014677657e-05, "loss": 4.2547, "step": 992800 }, { "epoch": 5.986722301950049, "grad_norm": 5.044709205627441, "learning_rate": 1.5322798013792535e-05, "loss": 4.0616, "step": 992850 }, { "epoch": 5.987023793731383, "grad_norm": 3.531721830368042, "learning_rate": 1.531383949144914e-05, "loss": 4.2991, "step": 992900 }, { "epoch": 5.987325285512717, "grad_norm": 3.9354400634765625, "learning_rate": 1.530488344781228e-05, "loss": 4.5233, "step": 992950 }, { "epoch": 5.987626777294051, "grad_norm": 1.3322128057479858, "learning_rate": 1.529592988304681e-05, "loss": 4.0293, "step": 993000 }, { "epoch": 5.987626777294051, "eval_loss": 4.848410606384277, "eval_runtime": 39.2626, "eval_samples_per_second": 13.04, "eval_steps_per_second": 6.52, "eval_tts_loss": 7.950866635684636, "step": 993000 }, { "epoch": 5.987928269075385, "grad_norm": 5.846860885620117, "learning_rate": 1.5286978797317552e-05, "loss": 4.4359, "step": 993050 }, { "epoch": 5.988229760856719, "grad_norm": 4.394814491271973, "learning_rate": 1.5278030190789188e-05, "loss": 4.2475, "step": 993100 }, { "epoch": 5.988531252638053, "grad_norm": 7.044231414794922, "learning_rate": 1.5269084063626403e-05, "loss": 4.1797, "step": 993150 }, { "epoch": 5.988832744419387, "grad_norm": 3.95660400390625, "learning_rate": 1.5260140415993882e-05, "loss": 4.1182, "step": 993200 }, { "epoch": 5.989134236200721, "grad_norm": 5.153783798217773, "learning_rate": 1.5251199248056212e-05, "loss": 4.5469, "step": 993250 }, { "epoch": 5.989435727982055, "grad_norm": 4.566433906555176, "learning_rate": 1.5242260559977876e-05, "loss": 4.0991, "step": 993300 }, { "epoch": 5.98973721976339, "grad_norm": 3.587484121322632, "learning_rate": 1.523332435192346e-05, "loss": 4.3731, "step": 993350 }, { "epoch": 5.990038711544724, "grad_norm": 5.355910301208496, "learning_rate": 1.5224390624057364e-05, "loss": 4.5595, "step": 993400 }, { "epoch": 5.990340203326057, "grad_norm": 3.7578065395355225, "learning_rate": 1.5215459376544009e-05, "loss": 4.7805, "step": 993450 }, { "epoch": 5.990641695107391, "grad_norm": 0.7172632217407227, "learning_rate": 1.5206530609547812e-05, "loss": 4.6283, "step": 993500 }, { "epoch": 5.9909431868887255, "grad_norm": 3.8780157566070557, "learning_rate": 1.5197604323233042e-05, "loss": 4.5799, "step": 993550 }, { "epoch": 5.99124467867006, "grad_norm": 8.330458641052246, "learning_rate": 1.5188680517763967e-05, "loss": 4.3805, "step": 993600 }, { "epoch": 5.991546170451393, "grad_norm": 3.5177619457244873, "learning_rate": 1.5179759193304892e-05, "loss": 4.0529, "step": 993650 }, { "epoch": 5.991847662232727, "grad_norm": 4.07930850982666, "learning_rate": 1.5170840350019946e-05, "loss": 4.5176, "step": 993700 }, { "epoch": 5.9921491540140615, "grad_norm": 5.679322719573975, "learning_rate": 1.5161923988073238e-05, "loss": 4.3226, "step": 993750 }, { "epoch": 5.992450645795396, "grad_norm": 5.254493236541748, "learning_rate": 1.5153010107628915e-05, "loss": 4.2275, "step": 993800 }, { "epoch": 5.99275213757673, "grad_norm": 5.982423305511475, "learning_rate": 1.5144098708850998e-05, "loss": 4.1805, "step": 993850 }, { "epoch": 5.993053629358064, "grad_norm": 2.6255502700805664, "learning_rate": 1.5135189791903474e-05, "loss": 4.3444, "step": 993900 }, { "epoch": 5.993355121139397, "grad_norm": 4.010504245758057, "learning_rate": 1.5126283356950358e-05, "loss": 4.3922, "step": 993950 }, { "epoch": 5.993656612920732, "grad_norm": 4.3216166496276855, "learning_rate": 1.5117379404155489e-05, "loss": 4.1691, "step": 994000 }, { "epoch": 5.993958104702066, "grad_norm": 4.703359603881836, "learning_rate": 1.51084779336828e-05, "loss": 4.2713, "step": 994050 }, { "epoch": 5.9942595964834, "grad_norm": 3.957547187805176, "learning_rate": 1.5099578945696062e-05, "loss": 4.28, "step": 994100 }, { "epoch": 5.994561088264734, "grad_norm": 3.858375072479248, "learning_rate": 1.5090682440359043e-05, "loss": 4.273, "step": 994150 }, { "epoch": 5.9948625800460675, "grad_norm": 2.3026959896087646, "learning_rate": 1.5081788417835477e-05, "loss": 4.5508, "step": 994200 }, { "epoch": 5.995164071827402, "grad_norm": 6.415427207946777, "learning_rate": 1.5072896878289086e-05, "loss": 4.3945, "step": 994250 }, { "epoch": 5.995465563608736, "grad_norm": 3.696098804473877, "learning_rate": 1.5064007821883456e-05, "loss": 4.3863, "step": 994300 }, { "epoch": 5.99576705539007, "grad_norm": 3.6854686737060547, "learning_rate": 1.5055121248782205e-05, "loss": 4.2391, "step": 994350 }, { "epoch": 5.996068547171404, "grad_norm": 3.354118824005127, "learning_rate": 1.5046237159148883e-05, "loss": 4.3265, "step": 994400 }, { "epoch": 5.996370038952739, "grad_norm": 3.721531391143799, "learning_rate": 1.503735555314698e-05, "loss": 4.3759, "step": 994450 }, { "epoch": 5.996671530734072, "grad_norm": 4.971909999847412, "learning_rate": 1.5028476430939912e-05, "loss": 4.0901, "step": 994500 }, { "epoch": 5.996973022515406, "grad_norm": 5.457627773284912, "learning_rate": 1.5019599792691167e-05, "loss": 4.4208, "step": 994550 }, { "epoch": 5.99727451429674, "grad_norm": 4.881081581115723, "learning_rate": 1.5010725638564014e-05, "loss": 4.6901, "step": 994600 }, { "epoch": 5.9975760060780745, "grad_norm": 3.5768303871154785, "learning_rate": 1.5001853968721838e-05, "loss": 4.4551, "step": 994650 }, { "epoch": 5.997877497859409, "grad_norm": 4.429909706115723, "learning_rate": 1.4992984783327894e-05, "loss": 4.3063, "step": 994700 }, { "epoch": 5.998178989640742, "grad_norm": 4.767649173736572, "learning_rate": 1.4984118082545382e-05, "loss": 4.5401, "step": 994750 }, { "epoch": 5.998480481422076, "grad_norm": 5.8221635818481445, "learning_rate": 1.497525386653749e-05, "loss": 4.3871, "step": 994800 }, { "epoch": 5.9987819732034104, "grad_norm": 1.7775181531906128, "learning_rate": 1.4966392135467403e-05, "loss": 4.3356, "step": 994850 }, { "epoch": 5.999083464984745, "grad_norm": 6.08511209487915, "learning_rate": 1.4957532889498142e-05, "loss": 4.5971, "step": 994900 }, { "epoch": 5.999384956766079, "grad_norm": 3.1599152088165283, "learning_rate": 1.4948676128792758e-05, "loss": 4.7602, "step": 994950 }, { "epoch": 5.999686448547413, "grad_norm": 4.62293004989624, "learning_rate": 1.4939821853514306e-05, "loss": 4.5181, "step": 995000 }, { "epoch": 5.999987940328746, "grad_norm": 3.4863317012786865, "learning_rate": 1.4930970063825636e-05, "loss": 4.3827, "step": 995050 }, { "epoch": 6.000289432110081, "grad_norm": 3.3916709423065186, "learning_rate": 1.4922120759889722e-05, "loss": 3.9417, "step": 995100 }, { "epoch": 6.000590923891415, "grad_norm": 6.098944664001465, "learning_rate": 1.491327394186943e-05, "loss": 4.2564, "step": 995150 }, { "epoch": 6.000892415672749, "grad_norm": 4.759034156799316, "learning_rate": 1.4904429609927533e-05, "loss": 4.303, "step": 995200 }, { "epoch": 6.001193907454083, "grad_norm": 3.185276746749878, "learning_rate": 1.4895587764226847e-05, "loss": 4.2028, "step": 995250 }, { "epoch": 6.0014953992354165, "grad_norm": 4.056685447692871, "learning_rate": 1.4886748404930044e-05, "loss": 4.2543, "step": 995300 }, { "epoch": 6.001796891016751, "grad_norm": 1.5146543979644775, "learning_rate": 1.487791153219981e-05, "loss": 3.8943, "step": 995350 }, { "epoch": 6.002098382798085, "grad_norm": 6.995954513549805, "learning_rate": 1.4869077146198783e-05, "loss": 3.6054, "step": 995400 }, { "epoch": 6.002399874579419, "grad_norm": 4.614500999450684, "learning_rate": 1.4860245247089564e-05, "loss": 4.1683, "step": 995450 }, { "epoch": 6.002701366360753, "grad_norm": 3.829916477203369, "learning_rate": 1.4851415835034659e-05, "loss": 4.239, "step": 995500 }, { "epoch": 6.003002858142087, "grad_norm": 4.195547103881836, "learning_rate": 1.4842588910196585e-05, "loss": 4.4174, "step": 995550 }, { "epoch": 6.003304349923421, "grad_norm": 4.205779552459717, "learning_rate": 1.4833764472737797e-05, "loss": 4.4083, "step": 995600 }, { "epoch": 6.003605841704755, "grad_norm": 5.923312664031982, "learning_rate": 1.4824942522820683e-05, "loss": 4.3902, "step": 995650 }, { "epoch": 6.003907333486089, "grad_norm": 6.2292375564575195, "learning_rate": 1.4816123060607577e-05, "loss": 4.0567, "step": 995700 }, { "epoch": 6.0042088252674235, "grad_norm": 4.972060680389404, "learning_rate": 1.4807306086260834e-05, "loss": 4.2453, "step": 995750 }, { "epoch": 6.004510317048757, "grad_norm": 3.315784454345703, "learning_rate": 1.4798491599942658e-05, "loss": 4.2067, "step": 995800 }, { "epoch": 6.004811808830091, "grad_norm": 5.39577054977417, "learning_rate": 1.4789679601815302e-05, "loss": 4.2667, "step": 995850 }, { "epoch": 6.005113300611425, "grad_norm": 3.5954842567443848, "learning_rate": 1.4780870092040969e-05, "loss": 4.0047, "step": 995900 }, { "epoch": 6.005414792392759, "grad_norm": 6.051171779632568, "learning_rate": 1.4772063070781731e-05, "loss": 4.4809, "step": 995950 }, { "epoch": 6.005716284174094, "grad_norm": 3.5335590839385986, "learning_rate": 1.476325853819969e-05, "loss": 4.2137, "step": 996000 }, { "epoch": 6.005716284174094, "eval_loss": 4.8567657470703125, "eval_runtime": 38.974, "eval_samples_per_second": 13.137, "eval_steps_per_second": 6.568, "eval_tts_loss": 8.073208599846335, "step": 996000 }, { "epoch": 6.006017775955428, "grad_norm": 3.8306350708007812, "learning_rate": 1.4754456494456901e-05, "loss": 4.1848, "step": 996050 }, { "epoch": 6.006319267736761, "grad_norm": 5.239724636077881, "learning_rate": 1.4745656939715333e-05, "loss": 4.1615, "step": 996100 }, { "epoch": 6.006620759518095, "grad_norm": 5.833479404449463, "learning_rate": 1.4736859874136908e-05, "loss": 3.851, "step": 996150 }, { "epoch": 6.00692225129943, "grad_norm": 4.269453525543213, "learning_rate": 1.4728065297883562e-05, "loss": 4.1288, "step": 996200 }, { "epoch": 6.007223743080764, "grad_norm": 5.512800693511963, "learning_rate": 1.4719273211117134e-05, "loss": 4.2625, "step": 996250 }, { "epoch": 6.007525234862098, "grad_norm": 4.967913627624512, "learning_rate": 1.4710483613999407e-05, "loss": 4.3032, "step": 996300 }, { "epoch": 6.007826726643431, "grad_norm": 7.522356033325195, "learning_rate": 1.4701696506692172e-05, "loss": 4.0449, "step": 996350 }, { "epoch": 6.0081282184247655, "grad_norm": 2.597677707672119, "learning_rate": 1.469291188935715e-05, "loss": 4.1578, "step": 996400 }, { "epoch": 6.0084297102061, "grad_norm": 4.440650463104248, "learning_rate": 1.4684129762155994e-05, "loss": 4.4688, "step": 996450 }, { "epoch": 6.008731201987434, "grad_norm": 4.815304756164551, "learning_rate": 1.4675350125250291e-05, "loss": 4.5322, "step": 996500 }, { "epoch": 6.009032693768768, "grad_norm": 5.926978588104248, "learning_rate": 1.4666572978801678e-05, "loss": 4.2299, "step": 996550 }, { "epoch": 6.009334185550102, "grad_norm": 5.196664333343506, "learning_rate": 1.4657798322971625e-05, "loss": 4.245, "step": 996600 }, { "epoch": 6.009635677331436, "grad_norm": 1.567948579788208, "learning_rate": 1.4649026157921655e-05, "loss": 4.4716, "step": 996650 }, { "epoch": 6.00993716911277, "grad_norm": 5.655208587646484, "learning_rate": 1.464025648381324e-05, "loss": 4.2848, "step": 996700 }, { "epoch": 6.010238660894104, "grad_norm": 4.571066379547119, "learning_rate": 1.4631489300807697e-05, "loss": 4.1661, "step": 996750 }, { "epoch": 6.010540152675438, "grad_norm": 3.937694549560547, "learning_rate": 1.462272460906645e-05, "loss": 4.4525, "step": 996800 }, { "epoch": 6.0108416444567725, "grad_norm": 4.854142189025879, "learning_rate": 1.4613962408750735e-05, "loss": 4.176, "step": 996850 }, { "epoch": 6.011143136238106, "grad_norm": 5.708829879760742, "learning_rate": 1.4605202700021856e-05, "loss": 4.1642, "step": 996900 }, { "epoch": 6.01144462801944, "grad_norm": 6.311600208282471, "learning_rate": 1.459644548304097e-05, "loss": 4.4024, "step": 996950 }, { "epoch": 6.011746119800774, "grad_norm": 3.8883402347564697, "learning_rate": 1.4587690757969312e-05, "loss": 4.1175, "step": 997000 }, { "epoch": 6.012047611582108, "grad_norm": 5.4482035636901855, "learning_rate": 1.457893852496792e-05, "loss": 4.2574, "step": 997050 }, { "epoch": 6.012349103363443, "grad_norm": 3.9459166526794434, "learning_rate": 1.4570188784197918e-05, "loss": 4.5141, "step": 997100 }, { "epoch": 6.012650595144776, "grad_norm": 3.9877400398254395, "learning_rate": 1.4561441535820357e-05, "loss": 4.1217, "step": 997150 }, { "epoch": 6.01295208692611, "grad_norm": 4.009861469268799, "learning_rate": 1.4552696779996142e-05, "loss": 4.2316, "step": 997200 }, { "epoch": 6.013253578707444, "grad_norm": 8.039348602294922, "learning_rate": 1.454395451688628e-05, "loss": 3.89, "step": 997250 }, { "epoch": 6.013555070488779, "grad_norm": 6.893671035766602, "learning_rate": 1.4535214746651641e-05, "loss": 3.9978, "step": 997300 }, { "epoch": 6.013856562270113, "grad_norm": 5.3263044357299805, "learning_rate": 1.4526477469453013e-05, "loss": 4.2828, "step": 997350 }, { "epoch": 6.014158054051447, "grad_norm": 2.1681525707244873, "learning_rate": 1.451774268545125e-05, "loss": 4.4522, "step": 997400 }, { "epoch": 6.01445954583278, "grad_norm": 3.9497008323669434, "learning_rate": 1.4509010394807109e-05, "loss": 4.2893, "step": 997450 }, { "epoch": 6.0147610376141145, "grad_norm": 5.115708827972412, "learning_rate": 1.4500280597681241e-05, "loss": 4.3783, "step": 997500 }, { "epoch": 6.015062529395449, "grad_norm": 6.857222080230713, "learning_rate": 1.4491553294234353e-05, "loss": 3.9295, "step": 997550 }, { "epoch": 6.015364021176783, "grad_norm": 4.40608549118042, "learning_rate": 1.4482828484627068e-05, "loss": 4.2057, "step": 997600 }, { "epoch": 6.015665512958117, "grad_norm": 7.281431674957275, "learning_rate": 1.4474106169019938e-05, "loss": 4.7668, "step": 997650 }, { "epoch": 6.0159670047394505, "grad_norm": 1.947826623916626, "learning_rate": 1.4465386347573438e-05, "loss": 4.0052, "step": 997700 }, { "epoch": 6.016268496520785, "grad_norm": 4.685832977294922, "learning_rate": 1.445666902044812e-05, "loss": 4.3415, "step": 997750 }, { "epoch": 6.016569988302119, "grad_norm": 10.821586608886719, "learning_rate": 1.4447954187804339e-05, "loss": 4.3341, "step": 997800 }, { "epoch": 6.016871480083453, "grad_norm": 2.3386800289154053, "learning_rate": 1.4439241849802536e-05, "loss": 4.3894, "step": 997850 }, { "epoch": 6.017172971864787, "grad_norm": 4.491759300231934, "learning_rate": 1.4430532006603046e-05, "loss": 4.0611, "step": 997900 }, { "epoch": 6.017474463646121, "grad_norm": 1.801466703414917, "learning_rate": 1.4421824658366127e-05, "loss": 4.2955, "step": 997950 }, { "epoch": 6.017775955427455, "grad_norm": 5.5827317237854, "learning_rate": 1.4413119805252032e-05, "loss": 4.1324, "step": 998000 }, { "epoch": 6.018077447208789, "grad_norm": 4.501517295837402, "learning_rate": 1.4404417447421018e-05, "loss": 4.155, "step": 998050 }, { "epoch": 6.018378938990123, "grad_norm": 5.6656174659729, "learning_rate": 1.439571758503319e-05, "loss": 4.2181, "step": 998100 }, { "epoch": 6.018680430771457, "grad_norm": 3.8122403621673584, "learning_rate": 1.4387020218248635e-05, "loss": 4.4952, "step": 998150 }, { "epoch": 6.018981922552792, "grad_norm": 5.7066569328308105, "learning_rate": 1.437832534722746e-05, "loss": 4.3631, "step": 998200 }, { "epoch": 6.019283414334125, "grad_norm": 3.903757333755493, "learning_rate": 1.4369632972129652e-05, "loss": 4.4594, "step": 998250 }, { "epoch": 6.019584906115459, "grad_norm": 3.906238317489624, "learning_rate": 1.4360943093115185e-05, "loss": 4.2056, "step": 998300 }, { "epoch": 6.019886397896793, "grad_norm": 3.739908456802368, "learning_rate": 1.4352255710344013e-05, "loss": 4.6277, "step": 998350 }, { "epoch": 6.020187889678128, "grad_norm": 3.1414477825164795, "learning_rate": 1.4343570823975976e-05, "loss": 4.6691, "step": 998400 }, { "epoch": 6.020489381459462, "grad_norm": 2.2579643726348877, "learning_rate": 1.433488843417096e-05, "loss": 3.9081, "step": 998450 }, { "epoch": 6.020790873240795, "grad_norm": 4.600595474243164, "learning_rate": 1.4326208541088708e-05, "loss": 4.2933, "step": 998500 }, { "epoch": 6.021092365022129, "grad_norm": 2.9859302043914795, "learning_rate": 1.4317531144888954e-05, "loss": 4.5479, "step": 998550 }, { "epoch": 6.0213938568034635, "grad_norm": 3.602787733078003, "learning_rate": 1.4308856245731393e-05, "loss": 4.3681, "step": 998600 }, { "epoch": 6.021695348584798, "grad_norm": 2.3702402114868164, "learning_rate": 1.4300183843775742e-05, "loss": 4.0416, "step": 998650 }, { "epoch": 6.021996840366132, "grad_norm": 2.699815511703491, "learning_rate": 1.429151393918151e-05, "loss": 4.3311, "step": 998700 }, { "epoch": 6.022298332147465, "grad_norm": 5.7149176597595215, "learning_rate": 1.42828465321083e-05, "loss": 4.0753, "step": 998750 }, { "epoch": 6.0225998239287994, "grad_norm": 6.748716831207275, "learning_rate": 1.4274181622715652e-05, "loss": 3.9507, "step": 998800 }, { "epoch": 6.022901315710134, "grad_norm": 6.260883331298828, "learning_rate": 1.426551921116299e-05, "loss": 4.218, "step": 998850 }, { "epoch": 6.023202807491468, "grad_norm": 4.428374767303467, "learning_rate": 1.4256859297609719e-05, "loss": 4.138, "step": 998900 }, { "epoch": 6.023504299272802, "grad_norm": 4.403228282928467, "learning_rate": 1.4248201882215277e-05, "loss": 4.6547, "step": 998950 }, { "epoch": 6.023805791054136, "grad_norm": 3.66404390335083, "learning_rate": 1.4239546965138904e-05, "loss": 4.3325, "step": 999000 }, { "epoch": 6.023805791054136, "eval_loss": 4.8548712730407715, "eval_runtime": 39.1705, "eval_samples_per_second": 13.071, "eval_steps_per_second": 6.536, "eval_tts_loss": 8.088326790517842, "step": 999000 }, { "epoch": 6.02410728283547, "grad_norm": 6.447748184204102, "learning_rate": 1.4230894546539939e-05, "loss": 4.0322, "step": 999050 }, { "epoch": 6.024408774616804, "grad_norm": 4.441551208496094, "learning_rate": 1.422224462657764e-05, "loss": 4.4061, "step": 999100 }, { "epoch": 6.024710266398138, "grad_norm": 3.2480485439300537, "learning_rate": 1.4213597205411127e-05, "loss": 4.162, "step": 999150 }, { "epoch": 6.025011758179472, "grad_norm": 1.1611220836639404, "learning_rate": 1.4204952283199588e-05, "loss": 4.3268, "step": 999200 }, { "epoch": 6.025313249960806, "grad_norm": 2.7975332736968994, "learning_rate": 1.4196309860102135e-05, "loss": 4.3316, "step": 999250 }, { "epoch": 6.02561474174214, "grad_norm": 2.8260128498077393, "learning_rate": 1.41876699362778e-05, "loss": 4.349, "step": 999300 }, { "epoch": 6.025916233523474, "grad_norm": 5.017808437347412, "learning_rate": 1.4179032511885546e-05, "loss": 4.2784, "step": 999350 }, { "epoch": 6.026217725304808, "grad_norm": 2.753918170928955, "learning_rate": 1.4170397587084408e-05, "loss": 4.0848, "step": 999400 }, { "epoch": 6.026519217086142, "grad_norm": 4.789759635925293, "learning_rate": 1.4161765162033245e-05, "loss": 4.3773, "step": 999450 }, { "epoch": 6.026820708867477, "grad_norm": 4.322310924530029, "learning_rate": 1.4153135236890928e-05, "loss": 4.2034, "step": 999500 }, { "epoch": 6.02712220064881, "grad_norm": 2.622042655944824, "learning_rate": 1.4144507811816296e-05, "loss": 4.4711, "step": 999550 }, { "epoch": 6.027423692430144, "grad_norm": 4.0262041091918945, "learning_rate": 1.4135882886968159e-05, "loss": 3.9732, "step": 999600 }, { "epoch": 6.027725184211478, "grad_norm": 4.7853851318359375, "learning_rate": 1.4127260462505202e-05, "loss": 4.338, "step": 999650 }, { "epoch": 6.0280266759928125, "grad_norm": 4.019760608673096, "learning_rate": 1.4118640538586084e-05, "loss": 4.3702, "step": 999700 }, { "epoch": 6.028328167774147, "grad_norm": 3.3650460243225098, "learning_rate": 1.4110023115369496e-05, "loss": 4.1141, "step": 999750 }, { "epoch": 6.028629659555481, "grad_norm": 5.314129829406738, "learning_rate": 1.4101408193013991e-05, "loss": 4.4477, "step": 999800 }, { "epoch": 6.028931151336814, "grad_norm": 2.7372241020202637, "learning_rate": 1.4092795771678128e-05, "loss": 4.1921, "step": 999850 }, { "epoch": 6.029232643118148, "grad_norm": 3.918640375137329, "learning_rate": 1.408418585152043e-05, "loss": 4.1191, "step": 999900 }, { "epoch": 6.029534134899483, "grad_norm": 3.7500994205474854, "learning_rate": 1.407557843269932e-05, "loss": 4.0323, "step": 999950 }, { "epoch": 6.029835626680817, "grad_norm": 4.640491485595703, "learning_rate": 1.406697351537322e-05, "loss": 4.2637, "step": 1000000 }, { "epoch": 6.030137118462151, "grad_norm": 4.347264289855957, "learning_rate": 1.4058371099700506e-05, "loss": 4.3018, "step": 1000050 }, { "epoch": 6.030438610243484, "grad_norm": 4.277031421661377, "learning_rate": 1.4049771185839431e-05, "loss": 3.8333, "step": 1000100 }, { "epoch": 6.030740102024819, "grad_norm": 4.268004894256592, "learning_rate": 1.4041173773948305e-05, "loss": 4.6667, "step": 1000150 }, { "epoch": 6.031041593806153, "grad_norm": 4.745704650878906, "learning_rate": 1.4032578864185385e-05, "loss": 4.5054, "step": 1000200 }, { "epoch": 6.031343085587487, "grad_norm": 5.899508476257324, "learning_rate": 1.4023986456708791e-05, "loss": 4.4, "step": 1000250 }, { "epoch": 6.031644577368821, "grad_norm": 9.631304740905762, "learning_rate": 1.4015396551676667e-05, "loss": 4.1777, "step": 1000300 }, { "epoch": 6.031946069150155, "grad_norm": 6.006819248199463, "learning_rate": 1.400680914924715e-05, "loss": 4.6485, "step": 1000350 }, { "epoch": 6.032247560931489, "grad_norm": 6.0637288093566895, "learning_rate": 1.3998224249578199e-05, "loss": 3.909, "step": 1000400 }, { "epoch": 6.032549052712823, "grad_norm": 2.061910390853882, "learning_rate": 1.398964185282787e-05, "loss": 4.169, "step": 1000450 }, { "epoch": 6.032850544494157, "grad_norm": 4.445291042327881, "learning_rate": 1.3981061959154105e-05, "loss": 4.3065, "step": 1000500 }, { "epoch": 6.033152036275491, "grad_norm": 5.237217426300049, "learning_rate": 1.3972484568714743e-05, "loss": 4.0351, "step": 1000550 }, { "epoch": 6.0334535280568256, "grad_norm": 4.73788595199585, "learning_rate": 1.3963909681667674e-05, "loss": 4.0944, "step": 1000600 }, { "epoch": 6.033755019838159, "grad_norm": 3.912609577178955, "learning_rate": 1.3955337298170754e-05, "loss": 4.0392, "step": 1000650 }, { "epoch": 6.034056511619493, "grad_norm": 4.242745399475098, "learning_rate": 1.3946767418381677e-05, "loss": 4.4556, "step": 1000700 }, { "epoch": 6.034358003400827, "grad_norm": 4.071492671966553, "learning_rate": 1.393820004245818e-05, "loss": 3.8929, "step": 1000750 }, { "epoch": 6.0346594951821615, "grad_norm": 4.188648223876953, "learning_rate": 1.3929635170557974e-05, "loss": 4.2144, "step": 1000800 }, { "epoch": 6.034960986963496, "grad_norm": 4.668923377990723, "learning_rate": 1.3921072802838645e-05, "loss": 4.0139, "step": 1000850 }, { "epoch": 6.035262478744829, "grad_norm": 7.3968729972839355, "learning_rate": 1.3912512939457754e-05, "loss": 4.4071, "step": 1000900 }, { "epoch": 6.035563970526163, "grad_norm": 4.202498435974121, "learning_rate": 1.3903955580572873e-05, "loss": 4.4883, "step": 1000950 }, { "epoch": 6.035865462307497, "grad_norm": 4.356410980224609, "learning_rate": 1.389540072634146e-05, "loss": 4.1634, "step": 1001000 }, { "epoch": 6.036166954088832, "grad_norm": 4.284942626953125, "learning_rate": 1.3886848376920956e-05, "loss": 4.5943, "step": 1001050 }, { "epoch": 6.036468445870166, "grad_norm": 3.6428496837615967, "learning_rate": 1.3878298532468785e-05, "loss": 4.4726, "step": 1001100 }, { "epoch": 6.0367699376515, "grad_norm": 3.4780750274658203, "learning_rate": 1.3869751193142253e-05, "loss": 4.0157, "step": 1001150 }, { "epoch": 6.037071429432833, "grad_norm": 5.624900817871094, "learning_rate": 1.3861206359098703e-05, "loss": 4.6094, "step": 1001200 }, { "epoch": 6.037372921214168, "grad_norm": 3.0532290935516357, "learning_rate": 1.3852664030495375e-05, "loss": 4.0763, "step": 1001250 }, { "epoch": 6.037674412995502, "grad_norm": 3.9345390796661377, "learning_rate": 1.3844124207489444e-05, "loss": 4.2869, "step": 1001300 }, { "epoch": 6.037975904776836, "grad_norm": 5.8692545890808105, "learning_rate": 1.3835586890238099e-05, "loss": 4.4454, "step": 1001350 }, { "epoch": 6.03827739655817, "grad_norm": 5.477871894836426, "learning_rate": 1.3827052078898499e-05, "loss": 4.2458, "step": 1001400 }, { "epoch": 6.0385788883395035, "grad_norm": 3.2276368141174316, "learning_rate": 1.3818519773627634e-05, "loss": 4.1187, "step": 1001450 }, { "epoch": 6.038880380120838, "grad_norm": 2.8657608032226562, "learning_rate": 1.380998997458258e-05, "loss": 4.41, "step": 1001500 }, { "epoch": 6.039181871902172, "grad_norm": 9.894784927368164, "learning_rate": 1.3801462681920328e-05, "loss": 4.3143, "step": 1001550 }, { "epoch": 6.039483363683506, "grad_norm": 4.271504878997803, "learning_rate": 1.3792937895797768e-05, "loss": 4.2708, "step": 1001600 }, { "epoch": 6.03978485546484, "grad_norm": 3.8491744995117188, "learning_rate": 1.3784415616371841e-05, "loss": 4.5991, "step": 1001650 }, { "epoch": 6.040086347246174, "grad_norm": 3.331562042236328, "learning_rate": 1.3775895843799339e-05, "loss": 4.1896, "step": 1001700 }, { "epoch": 6.040387839027508, "grad_norm": 4.2898430824279785, "learning_rate": 1.3767378578237055e-05, "loss": 4.26, "step": 1001750 }, { "epoch": 6.040689330808842, "grad_norm": 3.392843008041382, "learning_rate": 1.3758863819841758e-05, "loss": 4.1999, "step": 1001800 }, { "epoch": 6.040990822590176, "grad_norm": 4.51327657699585, "learning_rate": 1.3750351568770163e-05, "loss": 4.3667, "step": 1001850 }, { "epoch": 6.0412923143715105, "grad_norm": 3.3924341201782227, "learning_rate": 1.3741841825178889e-05, "loss": 4.3325, "step": 1001900 }, { "epoch": 6.041593806152845, "grad_norm": 4.734469413757324, "learning_rate": 1.3733334589224549e-05, "loss": 4.351, "step": 1001950 }, { "epoch": 6.041895297934178, "grad_norm": 3.872028112411499, "learning_rate": 1.3724829861063764e-05, "loss": 4.233, "step": 1002000 }, { "epoch": 6.041895297934178, "eval_loss": 4.852492809295654, "eval_runtime": 38.9769, "eval_samples_per_second": 13.136, "eval_steps_per_second": 6.568, "eval_tts_loss": 8.08977957289046, "step": 1002000 }, { "epoch": 6.042196789715512, "grad_norm": 4.7878031730651855, "learning_rate": 1.3716327640853009e-05, "loss": 3.9988, "step": 1002050 }, { "epoch": 6.042498281496846, "grad_norm": 4.331811428070068, "learning_rate": 1.3707827928748743e-05, "loss": 4.2027, "step": 1002100 }, { "epoch": 6.042799773278181, "grad_norm": 6.349465370178223, "learning_rate": 1.369933072490741e-05, "loss": 4.5399, "step": 1002150 }, { "epoch": 6.043101265059515, "grad_norm": 6.072861671447754, "learning_rate": 1.3690836029485364e-05, "loss": 4.2497, "step": 1002200 }, { "epoch": 6.043402756840848, "grad_norm": 4.278685569763184, "learning_rate": 1.3682343842638966e-05, "loss": 4.334, "step": 1002250 }, { "epoch": 6.043704248622182, "grad_norm": 4.448363780975342, "learning_rate": 1.3673854164524505e-05, "loss": 3.9557, "step": 1002300 }, { "epoch": 6.044005740403517, "grad_norm": 4.948603630065918, "learning_rate": 1.3665366995298194e-05, "loss": 4.3054, "step": 1002350 }, { "epoch": 6.044307232184851, "grad_norm": 7.0069050788879395, "learning_rate": 1.365688233511627e-05, "loss": 4.3354, "step": 1002400 }, { "epoch": 6.044608723966185, "grad_norm": 5.001265525817871, "learning_rate": 1.3648400184134828e-05, "loss": 4.3267, "step": 1002450 }, { "epoch": 6.044910215747518, "grad_norm": 3.8411059379577637, "learning_rate": 1.3639920542510008e-05, "loss": 4.1953, "step": 1002500 }, { "epoch": 6.0452117075288525, "grad_norm": 2.2485737800598145, "learning_rate": 1.3631443410397852e-05, "loss": 4.2329, "step": 1002550 }, { "epoch": 6.045513199310187, "grad_norm": 4.212012767791748, "learning_rate": 1.3622968787954385e-05, "loss": 3.797, "step": 1002600 }, { "epoch": 6.045814691091521, "grad_norm": 2.537306785583496, "learning_rate": 1.3614496675335517e-05, "loss": 4.5409, "step": 1002650 }, { "epoch": 6.046116182872855, "grad_norm": 5.040619850158691, "learning_rate": 1.3606027072697223e-05, "loss": 3.8128, "step": 1002700 }, { "epoch": 6.046417674654189, "grad_norm": 5.209597587585449, "learning_rate": 1.3597559980195378e-05, "loss": 4.1202, "step": 1002750 }, { "epoch": 6.046719166435523, "grad_norm": 5.627045154571533, "learning_rate": 1.3589095397985755e-05, "loss": 3.9508, "step": 1002800 }, { "epoch": 6.047020658216857, "grad_norm": 5.771944522857666, "learning_rate": 1.35806333262242e-05, "loss": 4.2491, "step": 1002850 }, { "epoch": 6.047322149998191, "grad_norm": 9.336252212524414, "learning_rate": 1.3572173765066368e-05, "loss": 4.563, "step": 1002900 }, { "epoch": 6.047623641779525, "grad_norm": 5.2875871658325195, "learning_rate": 1.356371671466802e-05, "loss": 4.1839, "step": 1002950 }, { "epoch": 6.0479251335608595, "grad_norm": 4.536770820617676, "learning_rate": 1.3555262175184733e-05, "loss": 4.1306, "step": 1003000 }, { "epoch": 6.048226625342193, "grad_norm": 5.8907999992370605, "learning_rate": 1.354681014677213e-05, "loss": 3.7411, "step": 1003050 }, { "epoch": 6.048528117123527, "grad_norm": 3.765608072280884, "learning_rate": 1.3538360629585788e-05, "loss": 4.2923, "step": 1003100 }, { "epoch": 6.048829608904861, "grad_norm": 3.9806129932403564, "learning_rate": 1.3529913623781147e-05, "loss": 4.2574, "step": 1003150 }, { "epoch": 6.049131100686195, "grad_norm": 4.581576824188232, "learning_rate": 1.3521469129513717e-05, "loss": 4.4704, "step": 1003200 }, { "epoch": 6.04943259246753, "grad_norm": 6.650818347930908, "learning_rate": 1.3513027146938876e-05, "loss": 4.386, "step": 1003250 }, { "epoch": 6.049734084248863, "grad_norm": 2.207125186920166, "learning_rate": 1.3504587676211964e-05, "loss": 4.0407, "step": 1003300 }, { "epoch": 6.050035576030197, "grad_norm": 1.5935115814208984, "learning_rate": 1.3496150717488341e-05, "loss": 3.6697, "step": 1003350 }, { "epoch": 6.050337067811531, "grad_norm": 4.1293134689331055, "learning_rate": 1.3487716270923282e-05, "loss": 4.3638, "step": 1003400 }, { "epoch": 6.050638559592866, "grad_norm": 1.8252493143081665, "learning_rate": 1.3479284336671964e-05, "loss": 4.072, "step": 1003450 }, { "epoch": 6.0509400513742, "grad_norm": 1.007738709449768, "learning_rate": 1.3470854914889594e-05, "loss": 4.318, "step": 1003500 }, { "epoch": 6.051241543155534, "grad_norm": 4.1732177734375, "learning_rate": 1.3462428005731335e-05, "loss": 4.0306, "step": 1003550 }, { "epoch": 6.051543034936867, "grad_norm": 4.769659996032715, "learning_rate": 1.3454003609352209e-05, "loss": 4.3813, "step": 1003600 }, { "epoch": 6.0518445267182015, "grad_norm": 4.564196586608887, "learning_rate": 1.3445581725907295e-05, "loss": 4.2552, "step": 1003650 }, { "epoch": 6.052146018499536, "grad_norm": 5.63437032699585, "learning_rate": 1.34371623555516e-05, "loss": 4.1038, "step": 1003700 }, { "epoch": 6.05244751028087, "grad_norm": 4.076622486114502, "learning_rate": 1.3428745498440003e-05, "loss": 4.3208, "step": 1003750 }, { "epoch": 6.052749002062204, "grad_norm": 3.709883213043213, "learning_rate": 1.342033115472746e-05, "loss": 4.3639, "step": 1003800 }, { "epoch": 6.053050493843537, "grad_norm": 5.280208110809326, "learning_rate": 1.3411919324568832e-05, "loss": 4.2917, "step": 1003850 }, { "epoch": 6.053351985624872, "grad_norm": 4.15906286239624, "learning_rate": 1.3403510008118879e-05, "loss": 4.4472, "step": 1003900 }, { "epoch": 6.053653477406206, "grad_norm": 4.738115310668945, "learning_rate": 1.3395103205532392e-05, "loss": 4.2515, "step": 1003950 }, { "epoch": 6.05395496918754, "grad_norm": 3.611372470855713, "learning_rate": 1.33866989169641e-05, "loss": 4.2615, "step": 1004000 }, { "epoch": 6.054256460968874, "grad_norm": 5.497255325317383, "learning_rate": 1.3378297142568678e-05, "loss": 4.4633, "step": 1004050 }, { "epoch": 6.0545579527502085, "grad_norm": 5.977441310882568, "learning_rate": 1.3369897882500686e-05, "loss": 4.4035, "step": 1004100 }, { "epoch": 6.054859444531542, "grad_norm": 4.0448832511901855, "learning_rate": 1.3361501136914782e-05, "loss": 4.1556, "step": 1004150 }, { "epoch": 6.055160936312876, "grad_norm": 4.359955310821533, "learning_rate": 1.3353106905965411e-05, "loss": 4.4011, "step": 1004200 }, { "epoch": 6.05546242809421, "grad_norm": 6.219391345977783, "learning_rate": 1.3344715189807114e-05, "loss": 4.2786, "step": 1004250 }, { "epoch": 6.055763919875544, "grad_norm": 6.2583417892456055, "learning_rate": 1.3336325988594338e-05, "loss": 3.9575, "step": 1004300 }, { "epoch": 6.056065411656879, "grad_norm": 1.905458688735962, "learning_rate": 1.3327939302481439e-05, "loss": 4.0486, "step": 1004350 }, { "epoch": 6.056366903438212, "grad_norm": 4.7057204246521, "learning_rate": 1.3319555131622795e-05, "loss": 4.0635, "step": 1004400 }, { "epoch": 6.056668395219546, "grad_norm": 5.709936618804932, "learning_rate": 1.3311173476172682e-05, "loss": 4.4327, "step": 1004450 }, { "epoch": 6.05696988700088, "grad_norm": 6.04281759262085, "learning_rate": 1.3302794336285327e-05, "loss": 3.9056, "step": 1004500 }, { "epoch": 6.0572713787822146, "grad_norm": 2.2970407009124756, "learning_rate": 1.3294417712114957e-05, "loss": 4.3447, "step": 1004550 }, { "epoch": 6.057572870563549, "grad_norm": 0.9370604157447815, "learning_rate": 1.3286043603815782e-05, "loss": 4.0992, "step": 1004600 }, { "epoch": 6.057874362344882, "grad_norm": 5.361557483673096, "learning_rate": 1.3277672011541845e-05, "loss": 4.5433, "step": 1004650 }, { "epoch": 6.058175854126216, "grad_norm": 3.568678617477417, "learning_rate": 1.326930293544724e-05, "loss": 4.009, "step": 1004700 }, { "epoch": 6.0584773459075505, "grad_norm": 4.606599807739258, "learning_rate": 1.326093637568601e-05, "loss": 4.3031, "step": 1004750 }, { "epoch": 6.058778837688885, "grad_norm": 5.323522567749023, "learning_rate": 1.325257233241208e-05, "loss": 4.1607, "step": 1004800 }, { "epoch": 6.059080329470219, "grad_norm": 3.441434621810913, "learning_rate": 1.3244210805779448e-05, "loss": 4.489, "step": 1004850 }, { "epoch": 6.059381821251553, "grad_norm": 5.304021835327148, "learning_rate": 1.3235851795941938e-05, "loss": 4.3224, "step": 1004900 }, { "epoch": 6.059683313032886, "grad_norm": 6.593399524688721, "learning_rate": 1.3227495303053392e-05, "loss": 4.5966, "step": 1004950 }, { "epoch": 6.059984804814221, "grad_norm": 4.5611066818237305, "learning_rate": 1.3219141327267607e-05, "loss": 4.2202, "step": 1005000 }, { "epoch": 6.059984804814221, "eval_loss": 4.8541669845581055, "eval_runtime": 39.2464, "eval_samples_per_second": 13.046, "eval_steps_per_second": 6.523, "eval_tts_loss": 8.116321777124519, "step": 1005000 }, { "epoch": 6.060286296595555, "grad_norm": 4.366276741027832, "learning_rate": 1.3210789868738342e-05, "loss": 4.4117, "step": 1005050 }, { "epoch": 6.060587788376889, "grad_norm": 4.98624849319458, "learning_rate": 1.3202440927619272e-05, "loss": 4.0026, "step": 1005100 }, { "epoch": 6.060889280158223, "grad_norm": 6.180444717407227, "learning_rate": 1.3194094504064045e-05, "loss": 4.1894, "step": 1005150 }, { "epoch": 6.061190771939557, "grad_norm": 5.4869465827941895, "learning_rate": 1.3185750598226302e-05, "loss": 4.4431, "step": 1005200 }, { "epoch": 6.061492263720891, "grad_norm": 3.0408456325531006, "learning_rate": 1.3177409210259588e-05, "loss": 3.9669, "step": 1005250 }, { "epoch": 6.061793755502225, "grad_norm": 4.052515029907227, "learning_rate": 1.316907034031736e-05, "loss": 4.3586, "step": 1005300 }, { "epoch": 6.062095247283559, "grad_norm": 4.829543113708496, "learning_rate": 1.3160733988553152e-05, "loss": 4.0738, "step": 1005350 }, { "epoch": 6.062396739064893, "grad_norm": 4.647202491760254, "learning_rate": 1.3152400155120335e-05, "loss": 4.3245, "step": 1005400 }, { "epoch": 6.062698230846227, "grad_norm": 4.497167587280273, "learning_rate": 1.3144068840172307e-05, "loss": 4.158, "step": 1005450 }, { "epoch": 6.062999722627561, "grad_norm": 4.753464698791504, "learning_rate": 1.3135740043862408e-05, "loss": 4.3082, "step": 1005500 }, { "epoch": 6.063301214408895, "grad_norm": 6.370828628540039, "learning_rate": 1.312741376634387e-05, "loss": 4.1685, "step": 1005550 }, { "epoch": 6.063602706190229, "grad_norm": 4.415899753570557, "learning_rate": 1.3119090007769968e-05, "loss": 4.1022, "step": 1005600 }, { "epoch": 6.0639041979715635, "grad_norm": 4.139736175537109, "learning_rate": 1.3110768768293895e-05, "loss": 4.3513, "step": 1005650 }, { "epoch": 6.064205689752898, "grad_norm": 3.5927727222442627, "learning_rate": 1.310245004806873e-05, "loss": 3.9367, "step": 1005700 }, { "epoch": 6.064507181534231, "grad_norm": 6.232890605926514, "learning_rate": 1.3094133847247618e-05, "loss": 4.27, "step": 1005750 }, { "epoch": 6.064808673315565, "grad_norm": 6.3769636154174805, "learning_rate": 1.3085820165983618e-05, "loss": 4.3892, "step": 1005800 }, { "epoch": 6.0651101650968995, "grad_norm": 6.139308929443359, "learning_rate": 1.3077509004429692e-05, "loss": 4.1897, "step": 1005850 }, { "epoch": 6.065411656878234, "grad_norm": 4.615022659301758, "learning_rate": 1.3069200362738802e-05, "loss": 3.8974, "step": 1005900 }, { "epoch": 6.065713148659568, "grad_norm": 6.784293174743652, "learning_rate": 1.3060894241063907e-05, "loss": 3.7584, "step": 1005950 }, { "epoch": 6.066014640440901, "grad_norm": 6.684248924255371, "learning_rate": 1.3052590639557802e-05, "loss": 4.0857, "step": 1006000 }, { "epoch": 6.066316132222235, "grad_norm": 3.590587615966797, "learning_rate": 1.304428955837335e-05, "loss": 4.0925, "step": 1006050 }, { "epoch": 6.06661762400357, "grad_norm": 3.9486513137817383, "learning_rate": 1.303599099766331e-05, "loss": 4.3718, "step": 1006100 }, { "epoch": 6.066919115784904, "grad_norm": 3.4941043853759766, "learning_rate": 1.3027694957580375e-05, "loss": 4.0729, "step": 1006150 }, { "epoch": 6.067220607566238, "grad_norm": 3.459268569946289, "learning_rate": 1.3019401438277228e-05, "loss": 4.5149, "step": 1006200 }, { "epoch": 6.067522099347571, "grad_norm": 6.036304473876953, "learning_rate": 1.3011110439906524e-05, "loss": 4.4988, "step": 1006250 }, { "epoch": 6.067823591128906, "grad_norm": 3.9980921745300293, "learning_rate": 1.300282196262088e-05, "loss": 3.81, "step": 1006300 }, { "epoch": 6.06812508291024, "grad_norm": 3.8553078174591064, "learning_rate": 1.2994536006572753e-05, "loss": 4.2156, "step": 1006350 }, { "epoch": 6.068426574691574, "grad_norm": 4.794662952423096, "learning_rate": 1.2986252571914691e-05, "loss": 4.0515, "step": 1006400 }, { "epoch": 6.068728066472908, "grad_norm": 5.72066593170166, "learning_rate": 1.2977971658799135e-05, "loss": 4.2865, "step": 1006450 }, { "epoch": 6.069029558254242, "grad_norm": 3.7919442653656006, "learning_rate": 1.2969693267378433e-05, "loss": 4.202, "step": 1006500 }, { "epoch": 6.069331050035576, "grad_norm": 5.086033344268799, "learning_rate": 1.2961417397804996e-05, "loss": 4.1881, "step": 1006550 }, { "epoch": 6.06963254181691, "grad_norm": 1.639626145362854, "learning_rate": 1.2953144050231117e-05, "loss": 4.0548, "step": 1006600 }, { "epoch": 6.069934033598244, "grad_norm": 4.474252700805664, "learning_rate": 1.294487322480901e-05, "loss": 3.9757, "step": 1006650 }, { "epoch": 6.070235525379578, "grad_norm": 8.287466049194336, "learning_rate": 1.2936604921690934e-05, "loss": 4.3543, "step": 1006700 }, { "epoch": 6.0705370171609125, "grad_norm": 2.3170087337493896, "learning_rate": 1.2928339141029071e-05, "loss": 4.4382, "step": 1006750 }, { "epoch": 6.070838508942246, "grad_norm": 3.0900089740753174, "learning_rate": 1.2920075882975511e-05, "loss": 3.9799, "step": 1006800 }, { "epoch": 6.07114000072358, "grad_norm": 6.137530326843262, "learning_rate": 1.2911815147682303e-05, "loss": 4.1909, "step": 1006850 }, { "epoch": 6.071441492504914, "grad_norm": 4.847071647644043, "learning_rate": 1.290355693530154e-05, "loss": 4.4732, "step": 1006900 }, { "epoch": 6.0717429842862485, "grad_norm": 2.1808197498321533, "learning_rate": 1.2895301245985118e-05, "loss": 4.2022, "step": 1006950 }, { "epoch": 6.072044476067583, "grad_norm": 5.3897199630737305, "learning_rate": 1.2887048079885031e-05, "loss": 4.2365, "step": 1007000 }, { "epoch": 6.072345967848916, "grad_norm": 5.575745582580566, "learning_rate": 1.2878797437153177e-05, "loss": 4.1518, "step": 1007050 }, { "epoch": 6.07264745963025, "grad_norm": 6.022787094116211, "learning_rate": 1.2870549317941347e-05, "loss": 4.5785, "step": 1007100 }, { "epoch": 6.072948951411584, "grad_norm": 3.8885343074798584, "learning_rate": 1.2862303722401374e-05, "loss": 4.1996, "step": 1007150 }, { "epoch": 6.073250443192919, "grad_norm": 2.774681568145752, "learning_rate": 1.2854060650685e-05, "loss": 4.281, "step": 1007200 }, { "epoch": 6.073551934974253, "grad_norm": 6.544476509094238, "learning_rate": 1.2845820102943937e-05, "loss": 4.333, "step": 1007250 }, { "epoch": 6.073853426755587, "grad_norm": 5.259943008422852, "learning_rate": 1.28375820793298e-05, "loss": 4.3195, "step": 1007300 }, { "epoch": 6.07415491853692, "grad_norm": 3.4399163722991943, "learning_rate": 1.2829346579994249e-05, "loss": 4.0585, "step": 1007350 }, { "epoch": 6.074456410318255, "grad_norm": 4.196422100067139, "learning_rate": 1.282111360508878e-05, "loss": 3.8922, "step": 1007400 }, { "epoch": 6.074757902099589, "grad_norm": 6.59808874130249, "learning_rate": 1.2812883154764968e-05, "loss": 4.1661, "step": 1007450 }, { "epoch": 6.075059393880923, "grad_norm": 6.669552803039551, "learning_rate": 1.2804655229174283e-05, "loss": 3.9381, "step": 1007500 }, { "epoch": 6.075360885662257, "grad_norm": 7.013058662414551, "learning_rate": 1.2796429828468114e-05, "loss": 4.2736, "step": 1007550 }, { "epoch": 6.0756623774435905, "grad_norm": 5.512246131896973, "learning_rate": 1.2788206952797859e-05, "loss": 4.3815, "step": 1007600 }, { "epoch": 6.075963869224925, "grad_norm": 4.040696144104004, "learning_rate": 1.2779986602314862e-05, "loss": 4.2038, "step": 1007650 }, { "epoch": 6.076265361006259, "grad_norm": 3.6032731533050537, "learning_rate": 1.2771768777170355e-05, "loss": 4.44, "step": 1007700 }, { "epoch": 6.076566852787593, "grad_norm": 4.278432369232178, "learning_rate": 1.2763553477515614e-05, "loss": 4.3137, "step": 1007750 }, { "epoch": 6.076868344568927, "grad_norm": 4.104910850524902, "learning_rate": 1.2755340703501838e-05, "loss": 4.4012, "step": 1007800 }, { "epoch": 6.0771698363502615, "grad_norm": 3.486199378967285, "learning_rate": 1.2747130455280135e-05, "loss": 4.3755, "step": 1007850 }, { "epoch": 6.077471328131595, "grad_norm": 8.913064002990723, "learning_rate": 1.273892273300164e-05, "loss": 3.806, "step": 1007900 }, { "epoch": 6.077772819912929, "grad_norm": 2.2127912044525146, "learning_rate": 1.2730717536817408e-05, "loss": 4.2325, "step": 1007950 }, { "epoch": 6.078074311694263, "grad_norm": 2.329796314239502, "learning_rate": 1.2722514866878425e-05, "loss": 4.3139, "step": 1008000 }, { "epoch": 6.078074311694263, "eval_loss": 4.8467793464660645, "eval_runtime": 39.0381, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.558, "eval_tts_loss": 8.07369188410002, "step": 1008000 }, { "epoch": 6.0783758034755975, "grad_norm": 3.9715588092803955, "learning_rate": 1.2714314723335616e-05, "loss": 4.3702, "step": 1008050 }, { "epoch": 6.078677295256932, "grad_norm": 5.0800862312316895, "learning_rate": 1.2706117106339964e-05, "loss": 4.3391, "step": 1008100 }, { "epoch": 6.078978787038265, "grad_norm": 4.104764938354492, "learning_rate": 1.2697922016042278e-05, "loss": 4.4101, "step": 1008150 }, { "epoch": 6.079280278819599, "grad_norm": 6.962893009185791, "learning_rate": 1.2689729452593372e-05, "loss": 4.4559, "step": 1008200 }, { "epoch": 6.079581770600933, "grad_norm": 2.745828628540039, "learning_rate": 1.2681539416144092e-05, "loss": 4.0983, "step": 1008250 }, { "epoch": 6.079883262382268, "grad_norm": 3.5329296588897705, "learning_rate": 1.2673351906845087e-05, "loss": 3.9969, "step": 1008300 }, { "epoch": 6.080184754163602, "grad_norm": 5.055573463439941, "learning_rate": 1.2665166924847048e-05, "loss": 4.0672, "step": 1008350 }, { "epoch": 6.080486245944935, "grad_norm": 4.448140621185303, "learning_rate": 1.2656984470300674e-05, "loss": 4.179, "step": 1008400 }, { "epoch": 6.080787737726269, "grad_norm": 3.5271737575531006, "learning_rate": 1.2648804543356478e-05, "loss": 4.4735, "step": 1008450 }, { "epoch": 6.0810892295076036, "grad_norm": 5.161962032318115, "learning_rate": 1.2640627144165005e-05, "loss": 4.005, "step": 1008500 }, { "epoch": 6.081390721288938, "grad_norm": 5.144651889801025, "learning_rate": 1.2632452272876802e-05, "loss": 4.257, "step": 1008550 }, { "epoch": 6.081692213070272, "grad_norm": 4.614532947540283, "learning_rate": 1.262427992964225e-05, "loss": 4.4432, "step": 1008600 }, { "epoch": 6.081993704851606, "grad_norm": 2.4626994132995605, "learning_rate": 1.2616110114611777e-05, "loss": 4.1313, "step": 1008650 }, { "epoch": 6.0822951966329395, "grad_norm": 3.1375815868377686, "learning_rate": 1.260794282793578e-05, "loss": 4.136, "step": 1008700 }, { "epoch": 6.082596688414274, "grad_norm": 4.338687419891357, "learning_rate": 1.2599778069764487e-05, "loss": 4.4443, "step": 1008750 }, { "epoch": 6.082898180195608, "grad_norm": 6.22871208190918, "learning_rate": 1.2591615840248231e-05, "loss": 4.7116, "step": 1008800 }, { "epoch": 6.083199671976942, "grad_norm": 4.5030012130737305, "learning_rate": 1.2583456139537191e-05, "loss": 4.2663, "step": 1008850 }, { "epoch": 6.083501163758276, "grad_norm": 7.743555068969727, "learning_rate": 1.2575298967781494e-05, "loss": 4.8592, "step": 1008900 }, { "epoch": 6.08380265553961, "grad_norm": 4.020897388458252, "learning_rate": 1.2567144325131322e-05, "loss": 4.152, "step": 1008950 }, { "epoch": 6.084104147320944, "grad_norm": 4.2191548347473145, "learning_rate": 1.2558992211736757e-05, "loss": 3.6665, "step": 1009000 }, { "epoch": 6.084405639102278, "grad_norm": 4.195882320404053, "learning_rate": 1.2550842627747776e-05, "loss": 4.4621, "step": 1009050 }, { "epoch": 6.084707130883612, "grad_norm": 6.488932132720947, "learning_rate": 1.2542695573314376e-05, "loss": 4.1016, "step": 1009100 }, { "epoch": 6.0850086226649465, "grad_norm": 3.8431522846221924, "learning_rate": 1.2534551048586538e-05, "loss": 4.7041, "step": 1009150 }, { "epoch": 6.08531011444628, "grad_norm": 4.477029323577881, "learning_rate": 1.2526409053714076e-05, "loss": 3.9156, "step": 1009200 }, { "epoch": 6.085611606227614, "grad_norm": 4.138172149658203, "learning_rate": 1.2518269588846919e-05, "loss": 4.0795, "step": 1009250 }, { "epoch": 6.085913098008948, "grad_norm": 4.854744911193848, "learning_rate": 1.2510132654134797e-05, "loss": 4.257, "step": 1009300 }, { "epoch": 6.086214589790282, "grad_norm": 2.617283821105957, "learning_rate": 1.250199824972744e-05, "loss": 4.7733, "step": 1009350 }, { "epoch": 6.086516081571617, "grad_norm": 3.878873348236084, "learning_rate": 1.2493866375774598e-05, "loss": 4.208, "step": 1009400 }, { "epoch": 6.086817573352951, "grad_norm": 2.30999755859375, "learning_rate": 1.2485737032425947e-05, "loss": 4.1329, "step": 1009450 }, { "epoch": 6.087119065134284, "grad_norm": 5.536127090454102, "learning_rate": 1.247761021983102e-05, "loss": 3.8683, "step": 1009500 }, { "epoch": 6.087420556915618, "grad_norm": 4.731263160705566, "learning_rate": 1.2469485938139429e-05, "loss": 4.0611, "step": 1009550 }, { "epoch": 6.0877220486969525, "grad_norm": 3.039741277694702, "learning_rate": 1.2461364187500704e-05, "loss": 4.1755, "step": 1009600 }, { "epoch": 6.088023540478287, "grad_norm": 0.9331954717636108, "learning_rate": 1.2453244968064296e-05, "loss": 4.2036, "step": 1009650 }, { "epoch": 6.088325032259621, "grad_norm": 4.3721699714660645, "learning_rate": 1.2445128279979599e-05, "loss": 4.0115, "step": 1009700 }, { "epoch": 6.088626524040954, "grad_norm": 7.647670745849609, "learning_rate": 1.2437014123396027e-05, "loss": 4.2326, "step": 1009750 }, { "epoch": 6.0889280158222885, "grad_norm": 3.472224235534668, "learning_rate": 1.2428902498462912e-05, "loss": 4.4281, "step": 1009800 }, { "epoch": 6.089229507603623, "grad_norm": 3.342453718185425, "learning_rate": 1.2420793405329483e-05, "loss": 3.9988, "step": 1009850 }, { "epoch": 6.089530999384957, "grad_norm": 4.138881206512451, "learning_rate": 1.2412686844145036e-05, "loss": 4.2054, "step": 1009900 }, { "epoch": 6.089832491166291, "grad_norm": 4.572276592254639, "learning_rate": 1.2404582815058755e-05, "loss": 4.1397, "step": 1009950 }, { "epoch": 6.090133982947624, "grad_norm": 3.089932680130005, "learning_rate": 1.239648131821977e-05, "loss": 4.4291, "step": 1010000 }, { "epoch": 6.090435474728959, "grad_norm": 3.5174806118011475, "learning_rate": 1.238838235377716e-05, "loss": 4.0883, "step": 1010050 }, { "epoch": 6.090736966510293, "grad_norm": 5.744992256164551, "learning_rate": 1.2380285921880007e-05, "loss": 4.2236, "step": 1010100 }, { "epoch": 6.091038458291627, "grad_norm": 2.290144681930542, "learning_rate": 1.2372192022677257e-05, "loss": 4.1845, "step": 1010150 }, { "epoch": 6.091339950072961, "grad_norm": 6.492090702056885, "learning_rate": 1.2364100656317927e-05, "loss": 4.2273, "step": 1010200 }, { "epoch": 6.0916414418542955, "grad_norm": 5.569454193115234, "learning_rate": 1.2356011822950928e-05, "loss": 4.5506, "step": 1010250 }, { "epoch": 6.091942933635629, "grad_norm": 4.051942825317383, "learning_rate": 1.2347925522725078e-05, "loss": 4.0417, "step": 1010300 }, { "epoch": 6.092244425416963, "grad_norm": 6.1129469871521, "learning_rate": 1.233984175578922e-05, "loss": 3.8599, "step": 1010350 }, { "epoch": 6.092545917198297, "grad_norm": 4.314840793609619, "learning_rate": 1.2331760522292155e-05, "loss": 4.1037, "step": 1010400 }, { "epoch": 6.092847408979631, "grad_norm": 1.5568351745605469, "learning_rate": 1.2323681822382564e-05, "loss": 4.2831, "step": 1010450 }, { "epoch": 6.093148900760966, "grad_norm": 3.956097364425659, "learning_rate": 1.231560565620911e-05, "loss": 4.3684, "step": 1010500 }, { "epoch": 6.093450392542299, "grad_norm": 4.674776077270508, "learning_rate": 1.2307532023920474e-05, "loss": 4.2518, "step": 1010550 }, { "epoch": 6.093751884323633, "grad_norm": 4.70447301864624, "learning_rate": 1.229946092566519e-05, "loss": 3.6923, "step": 1010600 }, { "epoch": 6.094053376104967, "grad_norm": 4.284597396850586, "learning_rate": 1.2291392361591818e-05, "loss": 4.2446, "step": 1010650 }, { "epoch": 6.0943548678863015, "grad_norm": 3.569708824157715, "learning_rate": 1.2283326331848875e-05, "loss": 4.1487, "step": 1010700 }, { "epoch": 6.094656359667636, "grad_norm": 4.318901538848877, "learning_rate": 1.2275262836584744e-05, "loss": 4.5933, "step": 1010750 }, { "epoch": 6.094957851448969, "grad_norm": 5.595326900482178, "learning_rate": 1.2267201875947885e-05, "loss": 3.9377, "step": 1010800 }, { "epoch": 6.095259343230303, "grad_norm": 5.205841541290283, "learning_rate": 1.2259143450086634e-05, "loss": 4.3236, "step": 1010850 }, { "epoch": 6.0955608350116375, "grad_norm": 5.666046619415283, "learning_rate": 1.2251087559149254e-05, "loss": 4.2424, "step": 1010900 }, { "epoch": 6.095862326792972, "grad_norm": 7.130800724029541, "learning_rate": 1.2243034203284008e-05, "loss": 4.5274, "step": 1010950 }, { "epoch": 6.096163818574306, "grad_norm": 10.130331039428711, "learning_rate": 1.223498338263918e-05, "loss": 4.281, "step": 1011000 }, { "epoch": 6.096163818574306, "eval_loss": 4.847867488861084, "eval_runtime": 39.1373, "eval_samples_per_second": 13.082, "eval_steps_per_second": 6.541, "eval_tts_loss": 8.124205967487464, "step": 1011000 }, { "epoch": 6.09646531035564, "grad_norm": 2.993549108505249, "learning_rate": 1.2226935097362833e-05, "loss": 4.1899, "step": 1011050 }, { "epoch": 6.096766802136973, "grad_norm": 4.188024044036865, "learning_rate": 1.221888934760315e-05, "loss": 3.9076, "step": 1011100 }, { "epoch": 6.097068293918308, "grad_norm": 6.414023399353027, "learning_rate": 1.2210846133508196e-05, "loss": 4.5154, "step": 1011150 }, { "epoch": 6.097369785699642, "grad_norm": 5.677143573760986, "learning_rate": 1.2202805455225984e-05, "loss": 4.1448, "step": 1011200 }, { "epoch": 6.097671277480976, "grad_norm": 5.107029438018799, "learning_rate": 1.219476731290448e-05, "loss": 4.2687, "step": 1011250 }, { "epoch": 6.09797276926231, "grad_norm": 4.939112663269043, "learning_rate": 1.2186731706691632e-05, "loss": 4.3894, "step": 1011300 }, { "epoch": 6.098274261043644, "grad_norm": 3.5273077487945557, "learning_rate": 1.2178698636735307e-05, "loss": 4.0072, "step": 1011350 }, { "epoch": 6.098575752824978, "grad_norm": 3.9631004333496094, "learning_rate": 1.217066810318335e-05, "loss": 4.6356, "step": 1011400 }, { "epoch": 6.098877244606312, "grad_norm": 4.627447128295898, "learning_rate": 1.2162640106183563e-05, "loss": 4.3945, "step": 1011450 }, { "epoch": 6.099178736387646, "grad_norm": 4.195650577545166, "learning_rate": 1.2154614645883675e-05, "loss": 4.1782, "step": 1011500 }, { "epoch": 6.09948022816898, "grad_norm": 3.6804261207580566, "learning_rate": 1.2146591722431387e-05, "loss": 3.9491, "step": 1011550 }, { "epoch": 6.099781719950315, "grad_norm": 11.640316009521484, "learning_rate": 1.2138571335974361e-05, "loss": 4.1553, "step": 1011600 }, { "epoch": 6.100083211731648, "grad_norm": 3.7401578426361084, "learning_rate": 1.21305534866602e-05, "loss": 4.269, "step": 1011650 }, { "epoch": 6.100384703512982, "grad_norm": 4.194705963134766, "learning_rate": 1.2122538174636432e-05, "loss": 4.0546, "step": 1011700 }, { "epoch": 6.100686195294316, "grad_norm": 1.6852794885635376, "learning_rate": 1.2114525400050607e-05, "loss": 4.2151, "step": 1011750 }, { "epoch": 6.1009876870756505, "grad_norm": 3.891789436340332, "learning_rate": 1.210651516305014e-05, "loss": 4.1512, "step": 1011800 }, { "epoch": 6.101289178856985, "grad_norm": 3.773864507675171, "learning_rate": 1.209850746378248e-05, "loss": 4.4717, "step": 1011850 }, { "epoch": 6.101590670638318, "grad_norm": 4.303004264831543, "learning_rate": 1.209050230239501e-05, "loss": 4.3994, "step": 1011900 }, { "epoch": 6.101892162419652, "grad_norm": 4.110435962677002, "learning_rate": 1.208249967903503e-05, "loss": 4.4462, "step": 1011950 }, { "epoch": 6.1021936542009865, "grad_norm": 5.842707633972168, "learning_rate": 1.2074499593849834e-05, "loss": 4.3046, "step": 1012000 }, { "epoch": 6.102495145982321, "grad_norm": 5.094550609588623, "learning_rate": 1.2066502046986658e-05, "loss": 4.045, "step": 1012050 }, { "epoch": 6.102796637763655, "grad_norm": 1.914636254310608, "learning_rate": 1.2058507038592634e-05, "loss": 4.0935, "step": 1012100 }, { "epoch": 6.103098129544988, "grad_norm": 3.9048922061920166, "learning_rate": 1.2050514568814929e-05, "loss": 4.4965, "step": 1012150 }, { "epoch": 6.103399621326322, "grad_norm": 3.973245143890381, "learning_rate": 1.2042524637800687e-05, "loss": 4.6266, "step": 1012200 }, { "epoch": 6.103701113107657, "grad_norm": 5.639232158660889, "learning_rate": 1.2034537245696863e-05, "loss": 4.4852, "step": 1012250 }, { "epoch": 6.104002604888991, "grad_norm": 3.6096231937408447, "learning_rate": 1.2026552392650484e-05, "loss": 4.2093, "step": 1012300 }, { "epoch": 6.104304096670325, "grad_norm": 7.054574012756348, "learning_rate": 1.2018570078808554e-05, "loss": 4.344, "step": 1012350 }, { "epoch": 6.104605588451659, "grad_norm": 4.481039524078369, "learning_rate": 1.2010590304317919e-05, "loss": 4.3264, "step": 1012400 }, { "epoch": 6.1049070802329926, "grad_norm": 4.746547698974609, "learning_rate": 1.2002613069325428e-05, "loss": 4.2785, "step": 1012450 }, { "epoch": 6.105208572014327, "grad_norm": 4.761368274688721, "learning_rate": 1.1994638373977932e-05, "loss": 4.4472, "step": 1012500 }, { "epoch": 6.105510063795661, "grad_norm": 6.784380912780762, "learning_rate": 1.1986666218422147e-05, "loss": 3.7333, "step": 1012550 }, { "epoch": 6.105811555576995, "grad_norm": 5.060460567474365, "learning_rate": 1.1978696602804822e-05, "loss": 3.9896, "step": 1012600 }, { "epoch": 6.106113047358329, "grad_norm": 5.885990142822266, "learning_rate": 1.1970729527272638e-05, "loss": 4.117, "step": 1012650 }, { "epoch": 6.106414539139663, "grad_norm": 5.1651835441589355, "learning_rate": 1.1962764991972178e-05, "loss": 4.3003, "step": 1012700 }, { "epoch": 6.106716030920997, "grad_norm": 2.5803301334381104, "learning_rate": 1.1954802997050028e-05, "loss": 4.3864, "step": 1012750 }, { "epoch": 6.107017522702331, "grad_norm": 6.4602556228637695, "learning_rate": 1.1946843542652767e-05, "loss": 4.3711, "step": 1012800 }, { "epoch": 6.107319014483665, "grad_norm": 4.695082664489746, "learning_rate": 1.193888662892683e-05, "loss": 4.3407, "step": 1012850 }, { "epoch": 6.1076205062649995, "grad_norm": 3.8047518730163574, "learning_rate": 1.1930932256018633e-05, "loss": 3.9651, "step": 1012900 }, { "epoch": 6.107921998046333, "grad_norm": 5.629500389099121, "learning_rate": 1.1922980424074607e-05, "loss": 4.6671, "step": 1012950 }, { "epoch": 6.108223489827667, "grad_norm": 2.854027271270752, "learning_rate": 1.1915031133241088e-05, "loss": 4.2154, "step": 1013000 }, { "epoch": 6.108524981609001, "grad_norm": 3.6069765090942383, "learning_rate": 1.1907084383664356e-05, "loss": 3.9499, "step": 1013050 }, { "epoch": 6.1088264733903355, "grad_norm": 6.032719135284424, "learning_rate": 1.1899140175490646e-05, "loss": 4.3492, "step": 1013100 }, { "epoch": 6.10912796517167, "grad_norm": 2.7575812339782715, "learning_rate": 1.1891198508866222e-05, "loss": 4.1518, "step": 1013150 }, { "epoch": 6.109429456953004, "grad_norm": 3.5287184715270996, "learning_rate": 1.1883259383937189e-05, "loss": 4.146, "step": 1013200 }, { "epoch": 6.109730948734337, "grad_norm": 1.7071905136108398, "learning_rate": 1.1875322800849641e-05, "loss": 3.8689, "step": 1013250 }, { "epoch": 6.110032440515671, "grad_norm": 5.115691661834717, "learning_rate": 1.1867388759749697e-05, "loss": 4.198, "step": 1013300 }, { "epoch": 6.110333932297006, "grad_norm": 4.016119956970215, "learning_rate": 1.1859457260783289e-05, "loss": 4.3971, "step": 1013350 }, { "epoch": 6.11063542407834, "grad_norm": 5.011976718902588, "learning_rate": 1.1851528304096453e-05, "loss": 4.1085, "step": 1013400 }, { "epoch": 6.110936915859674, "grad_norm": 3.2184267044067383, "learning_rate": 1.1843601889835103e-05, "loss": 4.2875, "step": 1013450 }, { "epoch": 6.111238407641007, "grad_norm": 4.464087963104248, "learning_rate": 1.183567801814509e-05, "loss": 4.0431, "step": 1013500 }, { "epoch": 6.1115398994223415, "grad_norm": 4.141294956207275, "learning_rate": 1.1827756689172262e-05, "loss": 3.5526, "step": 1013550 }, { "epoch": 6.111841391203676, "grad_norm": 6.1599440574646, "learning_rate": 1.181983790306239e-05, "loss": 4.071, "step": 1013600 }, { "epoch": 6.11214288298501, "grad_norm": 3.228318214416504, "learning_rate": 1.1811921659961205e-05, "loss": 4.0329, "step": 1013650 }, { "epoch": 6.112444374766344, "grad_norm": 4.615415573120117, "learning_rate": 1.1804007960014373e-05, "loss": 3.9878, "step": 1013700 }, { "epoch": 6.1127458665476775, "grad_norm": 7.114413261413574, "learning_rate": 1.1796096803367594e-05, "loss": 4.5659, "step": 1013750 }, { "epoch": 6.113047358329012, "grad_norm": 4.442835807800293, "learning_rate": 1.1788188190166387e-05, "loss": 4.415, "step": 1013800 }, { "epoch": 6.113348850110346, "grad_norm": 3.9933314323425293, "learning_rate": 1.178028212055635e-05, "loss": 3.8162, "step": 1013850 }, { "epoch": 6.11365034189168, "grad_norm": 5.657413959503174, "learning_rate": 1.1772378594683002e-05, "loss": 3.9131, "step": 1013900 }, { "epoch": 6.113951833673014, "grad_norm": 1.337388277053833, "learning_rate": 1.1764477612691726e-05, "loss": 4.1704, "step": 1013950 }, { "epoch": 6.1142533254543485, "grad_norm": 7.2655720710754395, "learning_rate": 1.1756579174728004e-05, "loss": 4.0142, "step": 1014000 }, { "epoch": 6.1142533254543485, "eval_loss": 4.843231201171875, "eval_runtime": 39.0563, "eval_samples_per_second": 13.109, "eval_steps_per_second": 6.555, "eval_tts_loss": 8.123262175730211, "step": 1014000 }, { "epoch": 6.114554817235682, "grad_norm": 5.033554553985596, "learning_rate": 1.1748683280937137e-05, "loss": 4.354, "step": 1014050 }, { "epoch": 6.114856309017016, "grad_norm": 2.513505458831787, "learning_rate": 1.1740789931464462e-05, "loss": 4.1942, "step": 1014100 }, { "epoch": 6.11515780079835, "grad_norm": 4.797536373138428, "learning_rate": 1.1732899126455226e-05, "loss": 4.2573, "step": 1014150 }, { "epoch": 6.1154592925796845, "grad_norm": 6.863955020904541, "learning_rate": 1.172501086605468e-05, "loss": 4.1709, "step": 1014200 }, { "epoch": 6.115760784361019, "grad_norm": 5.076610565185547, "learning_rate": 1.1717125150407974e-05, "loss": 4.167, "step": 1014250 }, { "epoch": 6.116062276142352, "grad_norm": 4.640174865722656, "learning_rate": 1.1709241979660228e-05, "loss": 4.0167, "step": 1014300 }, { "epoch": 6.116363767923686, "grad_norm": 3.8575847148895264, "learning_rate": 1.1701361353956557e-05, "loss": 4.0294, "step": 1014350 }, { "epoch": 6.11666525970502, "grad_norm": 5.524960517883301, "learning_rate": 1.1693483273441977e-05, "loss": 4.4915, "step": 1014400 }, { "epoch": 6.116966751486355, "grad_norm": 4.377342700958252, "learning_rate": 1.1685607738261426e-05, "loss": 4.2084, "step": 1014450 }, { "epoch": 6.117268243267689, "grad_norm": 4.603394031524658, "learning_rate": 1.1677734748559919e-05, "loss": 4.4598, "step": 1014500 }, { "epoch": 6.117569735049022, "grad_norm": 5.575933933258057, "learning_rate": 1.1669864304482274e-05, "loss": 4.4519, "step": 1014550 }, { "epoch": 6.117871226830356, "grad_norm": 6.174508094787598, "learning_rate": 1.1661996406173357e-05, "loss": 4.1925, "step": 1014600 }, { "epoch": 6.1181727186116905, "grad_norm": 4.512318134307861, "learning_rate": 1.165413105377802e-05, "loss": 4.2303, "step": 1014650 }, { "epoch": 6.118474210393025, "grad_norm": 4.517979621887207, "learning_rate": 1.1646268247440933e-05, "loss": 4.2711, "step": 1014700 }, { "epoch": 6.118775702174359, "grad_norm": 4.110952854156494, "learning_rate": 1.1638407987306875e-05, "loss": 4.6646, "step": 1014750 }, { "epoch": 6.119077193955693, "grad_norm": 5.470241546630859, "learning_rate": 1.1630550273520434e-05, "loss": 4.4665, "step": 1014800 }, { "epoch": 6.1193786857370265, "grad_norm": 7.762515544891357, "learning_rate": 1.1622695106226276e-05, "loss": 4.1768, "step": 1014850 }, { "epoch": 6.119680177518361, "grad_norm": 5.611794471740723, "learning_rate": 1.161484248556892e-05, "loss": 4.204, "step": 1014900 }, { "epoch": 6.119981669299695, "grad_norm": 3.8560707569122314, "learning_rate": 1.1606992411692918e-05, "loss": 4.4925, "step": 1014950 }, { "epoch": 6.120283161081029, "grad_norm": 3.074922800064087, "learning_rate": 1.15991448847427e-05, "loss": 4.0891, "step": 1015000 }, { "epoch": 6.120584652862363, "grad_norm": 4.181378364562988, "learning_rate": 1.1591299904862722e-05, "loss": 3.9614, "step": 1015050 }, { "epoch": 6.120886144643697, "grad_norm": 8.963436126708984, "learning_rate": 1.1583457472197365e-05, "loss": 4.4599, "step": 1015100 }, { "epoch": 6.121187636425031, "grad_norm": 6.139621734619141, "learning_rate": 1.1575617586890918e-05, "loss": 4.3381, "step": 1015150 }, { "epoch": 6.121489128206365, "grad_norm": 4.142983436584473, "learning_rate": 1.156778024908771e-05, "loss": 4.5124, "step": 1015200 }, { "epoch": 6.121790619987699, "grad_norm": 4.353391647338867, "learning_rate": 1.1559945458931964e-05, "loss": 4.1382, "step": 1015250 }, { "epoch": 6.1220921117690335, "grad_norm": 5.817462921142578, "learning_rate": 1.1552113216567827e-05, "loss": 4.2317, "step": 1015300 }, { "epoch": 6.122393603550368, "grad_norm": 3.4127800464630127, "learning_rate": 1.1544283522139453e-05, "loss": 4.0804, "step": 1015350 }, { "epoch": 6.122695095331701, "grad_norm": 7.270825386047363, "learning_rate": 1.1536456375791009e-05, "loss": 4.1832, "step": 1015400 }, { "epoch": 6.122996587113035, "grad_norm": 4.197663307189941, "learning_rate": 1.1528631777666448e-05, "loss": 4.3047, "step": 1015450 }, { "epoch": 6.123298078894369, "grad_norm": 5.3561272621154785, "learning_rate": 1.1520809727909802e-05, "loss": 4.3842, "step": 1015500 }, { "epoch": 6.123599570675704, "grad_norm": 2.2207624912261963, "learning_rate": 1.1512990226665059e-05, "loss": 4.0135, "step": 1015550 }, { "epoch": 6.123901062457038, "grad_norm": 5.424213409423828, "learning_rate": 1.1505173274076101e-05, "loss": 4.2705, "step": 1015600 }, { "epoch": 6.124202554238371, "grad_norm": 4.125248908996582, "learning_rate": 1.1497358870286766e-05, "loss": 4.3081, "step": 1015650 }, { "epoch": 6.124504046019705, "grad_norm": 5.25460958480835, "learning_rate": 1.148954701544092e-05, "loss": 4.0188, "step": 1015700 }, { "epoch": 6.1248055378010395, "grad_norm": 6.518383026123047, "learning_rate": 1.1481737709682266e-05, "loss": 4.2915, "step": 1015750 }, { "epoch": 6.125107029582374, "grad_norm": 6.705454349517822, "learning_rate": 1.1473930953154536e-05, "loss": 3.8967, "step": 1015800 }, { "epoch": 6.125408521363708, "grad_norm": 4.592440128326416, "learning_rate": 1.146612674600147e-05, "loss": 4.2263, "step": 1015850 }, { "epoch": 6.125710013145041, "grad_norm": 4.448983669281006, "learning_rate": 1.1458325088366599e-05, "loss": 3.94, "step": 1015900 }, { "epoch": 6.1260115049263755, "grad_norm": 3.6030771732330322, "learning_rate": 1.145052598039356e-05, "loss": 4.5563, "step": 1015950 }, { "epoch": 6.12631299670771, "grad_norm": 5.776621341705322, "learning_rate": 1.1442729422225888e-05, "loss": 4.2602, "step": 1016000 }, { "epoch": 6.126614488489044, "grad_norm": 5.374468803405762, "learning_rate": 1.1434935414007035e-05, "loss": 4.5562, "step": 1016050 }, { "epoch": 6.126915980270378, "grad_norm": 5.4425740242004395, "learning_rate": 1.1427143955880452e-05, "loss": 3.9631, "step": 1016100 }, { "epoch": 6.127217472051712, "grad_norm": 4.376430034637451, "learning_rate": 1.1419355047989542e-05, "loss": 4.442, "step": 1016150 }, { "epoch": 6.127518963833046, "grad_norm": 4.734912395477295, "learning_rate": 1.1411568690477608e-05, "loss": 4.2442, "step": 1016200 }, { "epoch": 6.12782045561438, "grad_norm": 4.7493438720703125, "learning_rate": 1.1403784883487982e-05, "loss": 4.3166, "step": 1016250 }, { "epoch": 6.128121947395714, "grad_norm": 3.7377400398254395, "learning_rate": 1.1396003627163935e-05, "loss": 4.0197, "step": 1016300 }, { "epoch": 6.128423439177048, "grad_norm": 7.717073917388916, "learning_rate": 1.138822492164862e-05, "loss": 4.0727, "step": 1016350 }, { "epoch": 6.1287249309583824, "grad_norm": 6.598751544952393, "learning_rate": 1.1380448767085238e-05, "loss": 4.1166, "step": 1016400 }, { "epoch": 6.129026422739716, "grad_norm": 3.824188470840454, "learning_rate": 1.137267516361684e-05, "loss": 4.2193, "step": 1016450 }, { "epoch": 6.12932791452105, "grad_norm": 4.388788223266602, "learning_rate": 1.1364904111386563e-05, "loss": 3.7736, "step": 1016500 }, { "epoch": 6.129629406302384, "grad_norm": 4.249017715454102, "learning_rate": 1.1357135610537343e-05, "loss": 3.775, "step": 1016550 }, { "epoch": 6.129930898083718, "grad_norm": 6.396420955657959, "learning_rate": 1.1349369661212198e-05, "loss": 4.1775, "step": 1016600 }, { "epoch": 6.130232389865053, "grad_norm": 6.962560176849365, "learning_rate": 1.1341606263554066e-05, "loss": 4.0268, "step": 1016650 }, { "epoch": 6.130533881646386, "grad_norm": 5.439754962921143, "learning_rate": 1.1333845417705778e-05, "loss": 4.2378, "step": 1016700 }, { "epoch": 6.13083537342772, "grad_norm": 4.227565288543701, "learning_rate": 1.132608712381019e-05, "loss": 4.6433, "step": 1016750 }, { "epoch": 6.131136865209054, "grad_norm": 5.0075907707214355, "learning_rate": 1.1318331382010071e-05, "loss": 4.3067, "step": 1016800 }, { "epoch": 6.1314383569903885, "grad_norm": 3.8944754600524902, "learning_rate": 1.1310578192448138e-05, "loss": 3.9499, "step": 1016850 }, { "epoch": 6.131739848771723, "grad_norm": 3.726411819458008, "learning_rate": 1.130282755526708e-05, "loss": 4.103, "step": 1016900 }, { "epoch": 6.132041340553057, "grad_norm": 5.6694135665893555, "learning_rate": 1.129507947060958e-05, "loss": 4.3647, "step": 1016950 }, { "epoch": 6.13234283233439, "grad_norm": 5.322018146514893, "learning_rate": 1.1287333938618176e-05, "loss": 4.5239, "step": 1017000 }, { "epoch": 6.13234283233439, "eval_loss": 4.839034080505371, "eval_runtime": 38.923, "eval_samples_per_second": 13.154, "eval_steps_per_second": 6.577, "eval_tts_loss": 8.100535386391075, "step": 1017000 }, { "epoch": 6.1326443241157245, "grad_norm": 1.4526747465133667, "learning_rate": 1.1279590959435436e-05, "loss": 3.4637, "step": 1017050 }, { "epoch": 6.132945815897059, "grad_norm": 3.1037492752075195, "learning_rate": 1.127185053320388e-05, "loss": 4.3983, "step": 1017100 }, { "epoch": 6.133247307678393, "grad_norm": 4.902060031890869, "learning_rate": 1.126411266006591e-05, "loss": 4.337, "step": 1017150 }, { "epoch": 6.133548799459727, "grad_norm": 3.7210488319396973, "learning_rate": 1.1256377340163996e-05, "loss": 4.1969, "step": 1017200 }, { "epoch": 6.13385029124106, "grad_norm": 5.133235454559326, "learning_rate": 1.1248644573640458e-05, "loss": 4.5289, "step": 1017250 }, { "epoch": 6.134151783022395, "grad_norm": 3.602562189102173, "learning_rate": 1.1240914360637582e-05, "loss": 4.0553, "step": 1017300 }, { "epoch": 6.134453274803729, "grad_norm": 6.5913166999816895, "learning_rate": 1.1233186701297654e-05, "loss": 3.927, "step": 1017350 }, { "epoch": 6.134754766585063, "grad_norm": 6.387356281280518, "learning_rate": 1.1225461595762925e-05, "loss": 4.6243, "step": 1017400 }, { "epoch": 6.135056258366397, "grad_norm": 3.9372591972351074, "learning_rate": 1.1217739044175517e-05, "loss": 3.9022, "step": 1017450 }, { "epoch": 6.1353577501477305, "grad_norm": 3.416630506515503, "learning_rate": 1.121001904667755e-05, "loss": 4.2, "step": 1017500 }, { "epoch": 6.135659241929065, "grad_norm": 2.6113646030426025, "learning_rate": 1.1202301603411156e-05, "loss": 4.178, "step": 1017550 }, { "epoch": 6.135960733710399, "grad_norm": 7.164012908935547, "learning_rate": 1.1194586714518312e-05, "loss": 4.4372, "step": 1017600 }, { "epoch": 6.136262225491733, "grad_norm": 5.04748010635376, "learning_rate": 1.1186874380140997e-05, "loss": 3.903, "step": 1017650 }, { "epoch": 6.136563717273067, "grad_norm": 3.726879119873047, "learning_rate": 1.1179164600421186e-05, "loss": 4.0859, "step": 1017700 }, { "epoch": 6.136865209054402, "grad_norm": 2.121335506439209, "learning_rate": 1.1171457375500714e-05, "loss": 4.2541, "step": 1017750 }, { "epoch": 6.137166700835735, "grad_norm": 4.907164573669434, "learning_rate": 1.1163752705521467e-05, "loss": 4.0827, "step": 1017800 }, { "epoch": 6.137468192617069, "grad_norm": 3.3705244064331055, "learning_rate": 1.1156050590625232e-05, "loss": 4.0319, "step": 1017850 }, { "epoch": 6.137769684398403, "grad_norm": 4.113058567047119, "learning_rate": 1.1148351030953729e-05, "loss": 4.3772, "step": 1017900 }, { "epoch": 6.1380711761797375, "grad_norm": 2.16796875, "learning_rate": 1.1140654026648677e-05, "loss": 4.1206, "step": 1017950 }, { "epoch": 6.138372667961072, "grad_norm": 6.0594000816345215, "learning_rate": 1.1132959577851746e-05, "loss": 4.6766, "step": 1018000 }, { "epoch": 6.138674159742405, "grad_norm": 4.0107316970825195, "learning_rate": 1.1125267684704492e-05, "loss": 3.9294, "step": 1018050 }, { "epoch": 6.138975651523739, "grad_norm": 6.944590091705322, "learning_rate": 1.11175783473485e-05, "loss": 4.0968, "step": 1018100 }, { "epoch": 6.1392771433050735, "grad_norm": 7.256692409515381, "learning_rate": 1.1109891565925305e-05, "loss": 3.7309, "step": 1018150 }, { "epoch": 6.139578635086408, "grad_norm": 4.393075942993164, "learning_rate": 1.1102207340576313e-05, "loss": 3.9622, "step": 1018200 }, { "epoch": 6.139880126867742, "grad_norm": 3.4474399089813232, "learning_rate": 1.1094525671442994e-05, "loss": 4.4665, "step": 1018250 }, { "epoch": 6.140181618649075, "grad_norm": 2.9186511039733887, "learning_rate": 1.1086846558666717e-05, "loss": 4.3788, "step": 1018300 }, { "epoch": 6.140483110430409, "grad_norm": 4.371588230133057, "learning_rate": 1.107917000238877e-05, "loss": 4.0306, "step": 1018350 }, { "epoch": 6.140784602211744, "grad_norm": 5.586213111877441, "learning_rate": 1.1071496002750473e-05, "loss": 4.309, "step": 1018400 }, { "epoch": 6.141086093993078, "grad_norm": 5.119361877441406, "learning_rate": 1.1063824559893032e-05, "loss": 4.0466, "step": 1018450 }, { "epoch": 6.141387585774412, "grad_norm": 6.740641117095947, "learning_rate": 1.1056155673957595e-05, "loss": 4.1143, "step": 1018500 }, { "epoch": 6.141689077555746, "grad_norm": 4.1519246101379395, "learning_rate": 1.1048489345085337e-05, "loss": 4.4919, "step": 1018550 }, { "epoch": 6.1419905693370795, "grad_norm": 8.992231369018555, "learning_rate": 1.104082557341736e-05, "loss": 4.2972, "step": 1018600 }, { "epoch": 6.142292061118414, "grad_norm": 3.905515670776367, "learning_rate": 1.1033164359094654e-05, "loss": 4.225, "step": 1018650 }, { "epoch": 6.142593552899748, "grad_norm": 7.331300258636475, "learning_rate": 1.1025505702258252e-05, "loss": 4.0767, "step": 1018700 }, { "epoch": 6.142895044681082, "grad_norm": 6.047948837280273, "learning_rate": 1.1017849603049112e-05, "loss": 4.0592, "step": 1018750 }, { "epoch": 6.143196536462416, "grad_norm": 3.4053804874420166, "learning_rate": 1.10101960616081e-05, "loss": 4.0642, "step": 1018800 }, { "epoch": 6.14349802824375, "grad_norm": 3.8128721714019775, "learning_rate": 1.1002545078076058e-05, "loss": 3.7861, "step": 1018850 }, { "epoch": 6.143799520025084, "grad_norm": 6.771768093109131, "learning_rate": 1.0994896652593854e-05, "loss": 4.0178, "step": 1018900 }, { "epoch": 6.144101011806418, "grad_norm": 6.046998023986816, "learning_rate": 1.0987250785302176e-05, "loss": 4.2521, "step": 1018950 }, { "epoch": 6.144402503587752, "grad_norm": 6.091573238372803, "learning_rate": 1.0979607476341761e-05, "loss": 4.4702, "step": 1019000 }, { "epoch": 6.1447039953690865, "grad_norm": 4.8137736320495605, "learning_rate": 1.0971966725853298e-05, "loss": 4.4413, "step": 1019050 }, { "epoch": 6.145005487150421, "grad_norm": 3.9704787731170654, "learning_rate": 1.0964328533977357e-05, "loss": 4.1538, "step": 1019100 }, { "epoch": 6.145306978931754, "grad_norm": 2.9323325157165527, "learning_rate": 1.0956692900854574e-05, "loss": 4.4789, "step": 1019150 }, { "epoch": 6.145608470713088, "grad_norm": 1.8274681568145752, "learning_rate": 1.0949059826625406e-05, "loss": 3.9698, "step": 1019200 }, { "epoch": 6.1459099624944225, "grad_norm": 3.115851640701294, "learning_rate": 1.0941429311430372e-05, "loss": 4.409, "step": 1019250 }, { "epoch": 6.146211454275757, "grad_norm": 4.01416015625, "learning_rate": 1.093380135540986e-05, "loss": 4.6269, "step": 1019300 }, { "epoch": 6.146512946057091, "grad_norm": 4.984346389770508, "learning_rate": 1.0926175958704308e-05, "loss": 4.0882, "step": 1019350 }, { "epoch": 6.146814437838424, "grad_norm": 4.0467939376831055, "learning_rate": 1.0918553121453988e-05, "loss": 4.3119, "step": 1019400 }, { "epoch": 6.147115929619758, "grad_norm": 4.102786540985107, "learning_rate": 1.091093284379922e-05, "loss": 4.0946, "step": 1019450 }, { "epoch": 6.147417421401093, "grad_norm": 8.756396293640137, "learning_rate": 1.0903315125880257e-05, "loss": 4.0858, "step": 1019500 }, { "epoch": 6.147718913182427, "grad_norm": 4.984177112579346, "learning_rate": 1.0895699967837257e-05, "loss": 4.184, "step": 1019550 }, { "epoch": 6.148020404963761, "grad_norm": 4.733212947845459, "learning_rate": 1.0888087369810422e-05, "loss": 4.2291, "step": 1019600 }, { "epoch": 6.148321896745094, "grad_norm": 4.0843586921691895, "learning_rate": 1.0880477331939807e-05, "loss": 4.2788, "step": 1019650 }, { "epoch": 6.1486233885264285, "grad_norm": 4.90800142288208, "learning_rate": 1.0872869854365451e-05, "loss": 3.8818, "step": 1019700 }, { "epoch": 6.148924880307763, "grad_norm": 3.812086582183838, "learning_rate": 1.0865264937227374e-05, "loss": 4.5389, "step": 1019750 }, { "epoch": 6.149226372089097, "grad_norm": 3.915921688079834, "learning_rate": 1.085766258066555e-05, "loss": 4.6498, "step": 1019800 }, { "epoch": 6.149527863870431, "grad_norm": 3.390928030014038, "learning_rate": 1.0850062784819896e-05, "loss": 4.3499, "step": 1019850 }, { "epoch": 6.149829355651765, "grad_norm": 4.9221320152282715, "learning_rate": 1.0842465549830237e-05, "loss": 4.4042, "step": 1019900 }, { "epoch": 6.150130847433099, "grad_norm": 3.487725019454956, "learning_rate": 1.0834870875836443e-05, "loss": 4.3143, "step": 1019950 }, { "epoch": 6.150432339214433, "grad_norm": 4.455696105957031, "learning_rate": 1.0827278762978236e-05, "loss": 4.3961, "step": 1020000 }, { "epoch": 6.150432339214433, "eval_loss": 4.838693618774414, "eval_runtime": 39.0328, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.559, "eval_tts_loss": 8.111754026466954, "step": 1020000 }, { "epoch": 6.150733830995767, "grad_norm": 6.874811172485352, "learning_rate": 1.0819689211395338e-05, "loss": 4.8069, "step": 1020050 }, { "epoch": 6.151035322777101, "grad_norm": 4.623138904571533, "learning_rate": 1.0812102221227437e-05, "loss": 4.0072, "step": 1020100 }, { "epoch": 6.1513368145584355, "grad_norm": 3.278404712677002, "learning_rate": 1.0804517792614204e-05, "loss": 4.0957, "step": 1020150 }, { "epoch": 6.151638306339769, "grad_norm": 6.549095630645752, "learning_rate": 1.0796935925695127e-05, "loss": 4.1735, "step": 1020200 }, { "epoch": 6.151939798121103, "grad_norm": 5.847355842590332, "learning_rate": 1.0789356620609812e-05, "loss": 4.2139, "step": 1020250 }, { "epoch": 6.152241289902437, "grad_norm": 3.3083138465881348, "learning_rate": 1.0781779877497732e-05, "loss": 4.3049, "step": 1020300 }, { "epoch": 6.1525427816837714, "grad_norm": 2.873103380203247, "learning_rate": 1.0774205696498306e-05, "loss": 4.2231, "step": 1020350 }, { "epoch": 6.152844273465106, "grad_norm": 2.894723415374756, "learning_rate": 1.0766634077750958e-05, "loss": 4.0777, "step": 1020400 }, { "epoch": 6.153145765246439, "grad_norm": 3.634031295776367, "learning_rate": 1.0759065021395008e-05, "loss": 4.1853, "step": 1020450 }, { "epoch": 6.153447257027773, "grad_norm": 6.331606864929199, "learning_rate": 1.0751498527569746e-05, "loss": 4.1679, "step": 1020500 }, { "epoch": 6.153748748809107, "grad_norm": 8.248435974121094, "learning_rate": 1.0743934596414428e-05, "loss": 4.2119, "step": 1020550 }, { "epoch": 6.154050240590442, "grad_norm": 4.033232688903809, "learning_rate": 1.073637322806829e-05, "loss": 4.0947, "step": 1020600 }, { "epoch": 6.154351732371776, "grad_norm": 4.369016647338867, "learning_rate": 1.0728814422670423e-05, "loss": 4.1926, "step": 1020650 }, { "epoch": 6.15465322415311, "grad_norm": 4.518033981323242, "learning_rate": 1.0721258180359998e-05, "loss": 4.0963, "step": 1020700 }, { "epoch": 6.154954715934443, "grad_norm": 4.73512077331543, "learning_rate": 1.0713704501276071e-05, "loss": 4.3998, "step": 1020750 }, { "epoch": 6.1552562077157775, "grad_norm": 3.828483819961548, "learning_rate": 1.0706153385557648e-05, "loss": 4.2084, "step": 1020800 }, { "epoch": 6.155557699497112, "grad_norm": 2.226224422454834, "learning_rate": 1.0698604833343648e-05, "loss": 3.6355, "step": 1020850 }, { "epoch": 6.155859191278446, "grad_norm": 5.419717311859131, "learning_rate": 1.0691058844773065e-05, "loss": 3.8109, "step": 1020900 }, { "epoch": 6.15616068305978, "grad_norm": 2.2718594074249268, "learning_rate": 1.0683515419984717e-05, "loss": 3.8565, "step": 1020950 }, { "epoch": 6.1564621748411135, "grad_norm": 3.221348762512207, "learning_rate": 1.0675974559117461e-05, "loss": 4.1077, "step": 1021000 }, { "epoch": 6.156763666622448, "grad_norm": 4.084616184234619, "learning_rate": 1.066843626231007e-05, "loss": 3.8205, "step": 1021050 }, { "epoch": 6.157065158403782, "grad_norm": 2.966855049133301, "learning_rate": 1.0660900529701266e-05, "loss": 4.3369, "step": 1021100 }, { "epoch": 6.157366650185116, "grad_norm": 4.006886959075928, "learning_rate": 1.0653367361429772e-05, "loss": 4.0142, "step": 1021150 }, { "epoch": 6.15766814196645, "grad_norm": 5.987650394439697, "learning_rate": 1.0645836757634174e-05, "loss": 4.1386, "step": 1021200 }, { "epoch": 6.157969633747784, "grad_norm": 5.397134780883789, "learning_rate": 1.0638308718453064e-05, "loss": 4.2817, "step": 1021250 }, { "epoch": 6.158271125529118, "grad_norm": 4.740568161010742, "learning_rate": 1.0630783244024998e-05, "loss": 4.4408, "step": 1021300 }, { "epoch": 6.158572617310452, "grad_norm": 5.243902206420898, "learning_rate": 1.0623260334488514e-05, "loss": 4.2784, "step": 1021350 }, { "epoch": 6.158874109091786, "grad_norm": 6.588207244873047, "learning_rate": 1.0615739989981986e-05, "loss": 4.1639, "step": 1021400 }, { "epoch": 6.15917560087312, "grad_norm": 4.464997291564941, "learning_rate": 1.0608222210643851e-05, "loss": 4.3318, "step": 1021450 }, { "epoch": 6.159477092654455, "grad_norm": 5.691761016845703, "learning_rate": 1.0600706996612484e-05, "loss": 3.9771, "step": 1021500 }, { "epoch": 6.159778584435788, "grad_norm": 5.583102226257324, "learning_rate": 1.0593194348026157e-05, "loss": 4.3528, "step": 1021550 }, { "epoch": 6.160080076217122, "grad_norm": 5.7801384925842285, "learning_rate": 1.0585684265023159e-05, "loss": 4.1851, "step": 1021600 }, { "epoch": 6.160381567998456, "grad_norm": 5.410754680633545, "learning_rate": 1.0578176747741696e-05, "loss": 3.9038, "step": 1021650 }, { "epoch": 6.160683059779791, "grad_norm": 3.970061779022217, "learning_rate": 1.0570671796319907e-05, "loss": 4.0929, "step": 1021700 }, { "epoch": 6.160984551561125, "grad_norm": 2.563154935836792, "learning_rate": 1.0563169410895916e-05, "loss": 4.103, "step": 1021750 }, { "epoch": 6.161286043342458, "grad_norm": 4.534616947174072, "learning_rate": 1.0555669591607829e-05, "loss": 4.3983, "step": 1021800 }, { "epoch": 6.161587535123792, "grad_norm": 4.406983375549316, "learning_rate": 1.054817233859362e-05, "loss": 4.1427, "step": 1021850 }, { "epoch": 6.1618890269051265, "grad_norm": 7.1680827140808105, "learning_rate": 1.0540677651991291e-05, "loss": 4.2227, "step": 1021900 }, { "epoch": 6.162190518686461, "grad_norm": 3.7275283336639404, "learning_rate": 1.0533185531938803e-05, "loss": 3.7342, "step": 1021950 }, { "epoch": 6.162492010467795, "grad_norm": 6.63512659072876, "learning_rate": 1.0525695978573994e-05, "loss": 4.0374, "step": 1022000 }, { "epoch": 6.162793502249128, "grad_norm": 6.817610263824463, "learning_rate": 1.0518208992034688e-05, "loss": 4.4398, "step": 1022050 }, { "epoch": 6.1630949940304625, "grad_norm": 2.7902822494506836, "learning_rate": 1.0510724572458723e-05, "loss": 4.3291, "step": 1022100 }, { "epoch": 6.163396485811797, "grad_norm": 2.8881208896636963, "learning_rate": 1.050324271998379e-05, "loss": 4.4303, "step": 1022150 }, { "epoch": 6.163697977593131, "grad_norm": 3.296035051345825, "learning_rate": 1.0495763434747594e-05, "loss": 4.3065, "step": 1022200 }, { "epoch": 6.163999469374465, "grad_norm": 4.044569492340088, "learning_rate": 1.0488286716887811e-05, "loss": 4.3023, "step": 1022250 }, { "epoch": 6.164300961155799, "grad_norm": 4.254007816314697, "learning_rate": 1.0480812566541996e-05, "loss": 4.4963, "step": 1022300 }, { "epoch": 6.164602452937133, "grad_norm": 6.302664279937744, "learning_rate": 1.0473340983847755e-05, "loss": 4.1448, "step": 1022350 }, { "epoch": 6.164903944718467, "grad_norm": 4.87704610824585, "learning_rate": 1.0465871968942547e-05, "loss": 4.0737, "step": 1022400 }, { "epoch": 6.165205436499801, "grad_norm": 5.961960792541504, "learning_rate": 1.0458405521963825e-05, "loss": 3.9964, "step": 1022450 }, { "epoch": 6.165506928281135, "grad_norm": 5.402919292449951, "learning_rate": 1.0450941643049e-05, "loss": 3.951, "step": 1022500 }, { "epoch": 6.165808420062469, "grad_norm": 6.808554649353027, "learning_rate": 1.0443480332335474e-05, "loss": 3.8797, "step": 1022550 }, { "epoch": 6.166109911843803, "grad_norm": 4.119390487670898, "learning_rate": 1.0436021589960525e-05, "loss": 4.5912, "step": 1022600 }, { "epoch": 6.166411403625137, "grad_norm": 4.653253555297852, "learning_rate": 1.0428565416061408e-05, "loss": 4.2853, "step": 1022650 }, { "epoch": 6.166712895406471, "grad_norm": 1.963302731513977, "learning_rate": 1.0421111810775412e-05, "loss": 4.0839, "step": 1022700 }, { "epoch": 6.167014387187805, "grad_norm": 2.90791916847229, "learning_rate": 1.041366077423963e-05, "loss": 4.0012, "step": 1022750 }, { "epoch": 6.16731587896914, "grad_norm": 4.417481899261475, "learning_rate": 1.0406212306591232e-05, "loss": 4.2483, "step": 1022800 }, { "epoch": 6.167617370750474, "grad_norm": 4.114194393157959, "learning_rate": 1.0398766407967296e-05, "loss": 4.4375, "step": 1022850 }, { "epoch": 6.167918862531807, "grad_norm": 3.7252166271209717, "learning_rate": 1.0391323078504826e-05, "loss": 3.9641, "step": 1022900 }, { "epoch": 6.168220354313141, "grad_norm": 7.556057929992676, "learning_rate": 1.0383882318340814e-05, "loss": 4.1467, "step": 1022950 }, { "epoch": 6.1685218460944755, "grad_norm": 2.323288679122925, "learning_rate": 1.0376444127612232e-05, "loss": 4.2388, "step": 1023000 }, { "epoch": 6.1685218460944755, "eval_loss": 4.834853172302246, "eval_runtime": 39.0622, "eval_samples_per_second": 13.107, "eval_steps_per_second": 6.554, "eval_tts_loss": 8.110537192952213, "step": 1023000 }, { "epoch": 6.16882333787581, "grad_norm": 3.3991661071777344, "learning_rate": 1.0369008506455923e-05, "loss": 4.2369, "step": 1023050 }, { "epoch": 6.169124829657144, "grad_norm": 4.874407768249512, "learning_rate": 1.0361575455008741e-05, "loss": 4.2003, "step": 1023100 }, { "epoch": 6.169426321438477, "grad_norm": 6.077311992645264, "learning_rate": 1.0354144973407514e-05, "loss": 4.5441, "step": 1023150 }, { "epoch": 6.1697278132198115, "grad_norm": 3.885025978088379, "learning_rate": 1.0346717061788962e-05, "loss": 4.4072, "step": 1023200 }, { "epoch": 6.170029305001146, "grad_norm": 6.367062091827393, "learning_rate": 1.033929172028976e-05, "loss": 4.0741, "step": 1023250 }, { "epoch": 6.17033079678248, "grad_norm": 4.394219875335693, "learning_rate": 1.0331868949046584e-05, "loss": 4.269, "step": 1023300 }, { "epoch": 6.170632288563814, "grad_norm": 4.344242095947266, "learning_rate": 1.0324448748196073e-05, "loss": 3.7222, "step": 1023350 }, { "epoch": 6.170933780345147, "grad_norm": 5.755746841430664, "learning_rate": 1.0317031117874736e-05, "loss": 3.9977, "step": 1023400 }, { "epoch": 6.171235272126482, "grad_norm": 4.592133045196533, "learning_rate": 1.0309616058219094e-05, "loss": 4.2893, "step": 1023450 }, { "epoch": 6.171536763907816, "grad_norm": 4.639553070068359, "learning_rate": 1.0302203569365658e-05, "loss": 4.5192, "step": 1023500 }, { "epoch": 6.17183825568915, "grad_norm": 5.468780994415283, "learning_rate": 1.0294793651450784e-05, "loss": 4.302, "step": 1023550 }, { "epoch": 6.172139747470484, "grad_norm": 4.297064304351807, "learning_rate": 1.0287386304610862e-05, "loss": 4.4345, "step": 1023600 }, { "epoch": 6.172441239251818, "grad_norm": 6.426765441894531, "learning_rate": 1.0279981528982217e-05, "loss": 4.0751, "step": 1023650 }, { "epoch": 6.172742731033152, "grad_norm": 1.0300090312957764, "learning_rate": 1.0272579324701108e-05, "loss": 4.185, "step": 1023700 }, { "epoch": 6.173044222814486, "grad_norm": 4.549749374389648, "learning_rate": 1.0265179691903775e-05, "loss": 3.9523, "step": 1023750 }, { "epoch": 6.17334571459582, "grad_norm": 4.063197135925293, "learning_rate": 1.0257782630726424e-05, "loss": 4.1893, "step": 1023800 }, { "epoch": 6.173647206377154, "grad_norm": 4.714461326599121, "learning_rate": 1.0250388141305133e-05, "loss": 3.6997, "step": 1023850 }, { "epoch": 6.173948698158489, "grad_norm": 4.799408435821533, "learning_rate": 1.0242996223776007e-05, "loss": 4.057, "step": 1023900 }, { "epoch": 6.174250189939822, "grad_norm": 3.4835493564605713, "learning_rate": 1.0235606878275138e-05, "loss": 3.9925, "step": 1023950 }, { "epoch": 6.174551681721156, "grad_norm": 5.2981038093566895, "learning_rate": 1.022822010493845e-05, "loss": 4.2667, "step": 1024000 }, { "epoch": 6.17485317350249, "grad_norm": 4.230299472808838, "learning_rate": 1.0220835903901887e-05, "loss": 4.283, "step": 1024050 }, { "epoch": 6.1751546652838245, "grad_norm": 4.800119400024414, "learning_rate": 1.0213454275301385e-05, "loss": 4.584, "step": 1024100 }, { "epoch": 6.175456157065159, "grad_norm": 4.609213352203369, "learning_rate": 1.0206075219272758e-05, "loss": 4.4429, "step": 1024150 }, { "epoch": 6.175757648846492, "grad_norm": 4.736240863800049, "learning_rate": 1.019869873595181e-05, "loss": 4.1377, "step": 1024200 }, { "epoch": 6.176059140627826, "grad_norm": 4.4526238441467285, "learning_rate": 1.0191324825474318e-05, "loss": 4.091, "step": 1024250 }, { "epoch": 6.1763606324091604, "grad_norm": 5.41721248626709, "learning_rate": 1.0183953487975971e-05, "loss": 4.1464, "step": 1024300 }, { "epoch": 6.176662124190495, "grad_norm": 4.445496082305908, "learning_rate": 1.0176584723592445e-05, "loss": 4.3531, "step": 1024350 }, { "epoch": 6.176963615971829, "grad_norm": 4.7060322761535645, "learning_rate": 1.016921853245935e-05, "loss": 4.3616, "step": 1024400 }, { "epoch": 6.177265107753163, "grad_norm": 4.038754940032959, "learning_rate": 1.0161854914712192e-05, "loss": 4.051, "step": 1024450 }, { "epoch": 6.177566599534496, "grad_norm": 6.042251110076904, "learning_rate": 1.0154493870486563e-05, "loss": 4.2863, "step": 1024500 }, { "epoch": 6.177868091315831, "grad_norm": 4.986377716064453, "learning_rate": 1.0147135399917904e-05, "loss": 4.649, "step": 1024550 }, { "epoch": 6.178169583097165, "grad_norm": 8.415621757507324, "learning_rate": 1.0139779503141626e-05, "loss": 4.1313, "step": 1024600 }, { "epoch": 6.178471074878499, "grad_norm": 3.645003318786621, "learning_rate": 1.0132426180293119e-05, "loss": 4.4821, "step": 1024650 }, { "epoch": 6.178772566659833, "grad_norm": 4.222476959228516, "learning_rate": 1.0125075431507724e-05, "loss": 4.3059, "step": 1024700 }, { "epoch": 6.1790740584411665, "grad_norm": 6.224026203155518, "learning_rate": 1.0117727256920699e-05, "loss": 4.386, "step": 1024750 }, { "epoch": 6.179375550222501, "grad_norm": 4.297092437744141, "learning_rate": 1.0110381656667255e-05, "loss": 3.8005, "step": 1024800 }, { "epoch": 6.179677042003835, "grad_norm": 5.111364841461182, "learning_rate": 1.0103038630882631e-05, "loss": 4.4157, "step": 1024850 }, { "epoch": 6.179978533785169, "grad_norm": 6.984058380126953, "learning_rate": 1.0095698179701923e-05, "loss": 4.1204, "step": 1024900 }, { "epoch": 6.180280025566503, "grad_norm": 5.473472595214844, "learning_rate": 1.0088360303260235e-05, "loss": 4.0133, "step": 1024950 }, { "epoch": 6.180581517347837, "grad_norm": 6.442324638366699, "learning_rate": 1.0081025001692628e-05, "loss": 3.9923, "step": 1025000 }, { "epoch": 6.180883009129171, "grad_norm": 3.135115146636963, "learning_rate": 1.0073692275134043e-05, "loss": 4.3485, "step": 1025050 }, { "epoch": 6.181184500910505, "grad_norm": 6.0205535888671875, "learning_rate": 1.006636212371949e-05, "loss": 4.4688, "step": 1025100 }, { "epoch": 6.181485992691839, "grad_norm": 5.685604572296143, "learning_rate": 1.005903454758386e-05, "loss": 4.1978, "step": 1025150 }, { "epoch": 6.1817874844731735, "grad_norm": 5.302348613739014, "learning_rate": 1.0051709546861996e-05, "loss": 4.2493, "step": 1025200 }, { "epoch": 6.182088976254508, "grad_norm": 6.859151840209961, "learning_rate": 1.0044387121688674e-05, "loss": 4.1025, "step": 1025250 }, { "epoch": 6.182390468035841, "grad_norm": 3.3440334796905518, "learning_rate": 1.0037067272198702e-05, "loss": 4.1679, "step": 1025300 }, { "epoch": 6.182691959817175, "grad_norm": 1.0776042938232422, "learning_rate": 1.0029749998526737e-05, "loss": 4.0117, "step": 1025350 }, { "epoch": 6.182993451598509, "grad_norm": 6.166214942932129, "learning_rate": 1.0022435300807491e-05, "loss": 4.4064, "step": 1025400 }, { "epoch": 6.183294943379844, "grad_norm": 4.582485198974609, "learning_rate": 1.0015123179175588e-05, "loss": 3.8198, "step": 1025450 }, { "epoch": 6.183596435161178, "grad_norm": 6.426245212554932, "learning_rate": 1.0007813633765538e-05, "loss": 4.0748, "step": 1025500 }, { "epoch": 6.183897926942511, "grad_norm": 6.267674446105957, "learning_rate": 1.0000506664711933e-05, "loss": 4.5462, "step": 1025550 }, { "epoch": 6.184199418723845, "grad_norm": 5.021034240722656, "learning_rate": 9.993202272149197e-06, "loss": 4.6018, "step": 1025600 }, { "epoch": 6.18450091050518, "grad_norm": 5.291874885559082, "learning_rate": 9.985900456211776e-06, "loss": 4.2097, "step": 1025650 }, { "epoch": 6.184802402286514, "grad_norm": 3.97316312789917, "learning_rate": 9.978601217034026e-06, "loss": 4.2304, "step": 1025700 }, { "epoch": 6.185103894067848, "grad_norm": 4.229088306427002, "learning_rate": 9.971304554750325e-06, "loss": 4.7326, "step": 1025750 }, { "epoch": 6.185405385849181, "grad_norm": 4.236648082733154, "learning_rate": 9.964010469494915e-06, "loss": 4.2251, "step": 1025800 }, { "epoch": 6.1857068776305155, "grad_norm": 5.07537317276001, "learning_rate": 9.956718961402038e-06, "loss": 4.3986, "step": 1025850 }, { "epoch": 6.18600836941185, "grad_norm": 6.93405294418335, "learning_rate": 9.949430030605937e-06, "loss": 4.1047, "step": 1025900 }, { "epoch": 6.186309861193184, "grad_norm": 7.401350975036621, "learning_rate": 9.942143677240672e-06, "loss": 4.3476, "step": 1025950 }, { "epoch": 6.186611352974518, "grad_norm": 2.190422296524048, "learning_rate": 9.934859901440418e-06, "loss": 4.1822, "step": 1026000 }, { "epoch": 6.186611352974518, "eval_loss": 4.833096504211426, "eval_runtime": 39.2581, "eval_samples_per_second": 13.042, "eval_steps_per_second": 6.521, "eval_tts_loss": 8.115339398267778, "step": 1026000 }, { "epoch": 6.186912844755852, "grad_norm": 0.9668341875076294, "learning_rate": 9.927578703339167e-06, "loss": 4.0908, "step": 1026050 }, { "epoch": 6.187214336537186, "grad_norm": 4.5527825355529785, "learning_rate": 9.92030008307093e-06, "loss": 3.8642, "step": 1026100 }, { "epoch": 6.18751582831852, "grad_norm": 6.2838897705078125, "learning_rate": 9.913024040769651e-06, "loss": 4.4527, "step": 1026150 }, { "epoch": 6.187817320099854, "grad_norm": 5.586788654327393, "learning_rate": 9.905750576569289e-06, "loss": 4.3171, "step": 1026200 }, { "epoch": 6.188118811881188, "grad_norm": 3.7018790245056152, "learning_rate": 9.898479690603634e-06, "loss": 3.9141, "step": 1026250 }, { "epoch": 6.1884203036625225, "grad_norm": 4.5520339012146, "learning_rate": 9.891211383006514e-06, "loss": 4.2851, "step": 1026300 }, { "epoch": 6.188721795443856, "grad_norm": 5.414066791534424, "learning_rate": 9.88394565391174e-06, "loss": 4.2046, "step": 1026350 }, { "epoch": 6.18902328722519, "grad_norm": 5.336474895477295, "learning_rate": 9.87668250345297e-06, "loss": 4.1606, "step": 1026400 }, { "epoch": 6.189324779006524, "grad_norm": 4.007259845733643, "learning_rate": 9.869421931763882e-06, "loss": 4.4649, "step": 1026450 }, { "epoch": 6.189626270787858, "grad_norm": 2.2055130004882812, "learning_rate": 9.862163938978119e-06, "loss": 4.1241, "step": 1026500 }, { "epoch": 6.189927762569193, "grad_norm": 2.9998371601104736, "learning_rate": 9.854908525229238e-06, "loss": 4.5293, "step": 1026550 }, { "epoch": 6.190229254350527, "grad_norm": 4.794929027557373, "learning_rate": 9.84765569065077e-06, "loss": 3.912, "step": 1026600 }, { "epoch": 6.19053074613186, "grad_norm": 4.322336673736572, "learning_rate": 9.84040543537617e-06, "loss": 4.202, "step": 1026650 }, { "epoch": 6.190832237913194, "grad_norm": 4.031674861907959, "learning_rate": 9.833157759538918e-06, "loss": 4.3475, "step": 1026700 }, { "epoch": 6.191133729694529, "grad_norm": 4.259881973266602, "learning_rate": 9.825912663272373e-06, "loss": 4.3253, "step": 1026750 }, { "epoch": 6.191435221475863, "grad_norm": 3.676656484603882, "learning_rate": 9.818670146709845e-06, "loss": 4.2253, "step": 1026800 }, { "epoch": 6.191736713257197, "grad_norm": 5.5120649337768555, "learning_rate": 9.811430209984661e-06, "loss": 4.4492, "step": 1026850 }, { "epoch": 6.19203820503853, "grad_norm": 3.198089361190796, "learning_rate": 9.80419285323003e-06, "loss": 4.0256, "step": 1026900 }, { "epoch": 6.1923396968198645, "grad_norm": 4.036252498626709, "learning_rate": 9.796958076579147e-06, "loss": 4.4621, "step": 1026950 }, { "epoch": 6.192641188601199, "grad_norm": 4.521757125854492, "learning_rate": 9.789725880165206e-06, "loss": 4.5606, "step": 1027000 }, { "epoch": 6.192942680382533, "grad_norm": 3.344264507293701, "learning_rate": 9.782496264121232e-06, "loss": 4.229, "step": 1027050 }, { "epoch": 6.193244172163867, "grad_norm": 4.566240310668945, "learning_rate": 9.775269228580318e-06, "loss": 4.0373, "step": 1027100 }, { "epoch": 6.1935456639452005, "grad_norm": 4.517706871032715, "learning_rate": 9.768044773675494e-06, "loss": 4.2193, "step": 1027150 }, { "epoch": 6.193847155726535, "grad_norm": 5.4992876052856445, "learning_rate": 9.760822899539683e-06, "loss": 4.4247, "step": 1027200 }, { "epoch": 6.194148647507869, "grad_norm": 4.5958170890808105, "learning_rate": 9.753603606305766e-06, "loss": 4.1303, "step": 1027250 }, { "epoch": 6.194450139289203, "grad_norm": 5.227288246154785, "learning_rate": 9.74638689410665e-06, "loss": 4.2935, "step": 1027300 }, { "epoch": 6.194751631070537, "grad_norm": 4.957185745239258, "learning_rate": 9.739172763075114e-06, "loss": 4.0688, "step": 1027350 }, { "epoch": 6.1950531228518715, "grad_norm": 5.5270538330078125, "learning_rate": 9.731961213343936e-06, "loss": 4.4131, "step": 1027400 }, { "epoch": 6.195354614633205, "grad_norm": 3.9278464317321777, "learning_rate": 9.724752245045875e-06, "loss": 4.6253, "step": 1027450 }, { "epoch": 6.195656106414539, "grad_norm": 4.569133281707764, "learning_rate": 9.717545858313525e-06, "loss": 4.1586, "step": 1027500 }, { "epoch": 6.195957598195873, "grad_norm": 5.165132999420166, "learning_rate": 9.71034205327958e-06, "loss": 4.5203, "step": 1027550 }, { "epoch": 6.196259089977207, "grad_norm": 5.274372577667236, "learning_rate": 9.7031408300766e-06, "loss": 4.0031, "step": 1027600 }, { "epoch": 6.196560581758542, "grad_norm": 5.638523101806641, "learning_rate": 9.695942188837047e-06, "loss": 4.3565, "step": 1027650 }, { "epoch": 6.196862073539875, "grad_norm": 3.0931382179260254, "learning_rate": 9.68874612969348e-06, "loss": 4.2861, "step": 1027700 }, { "epoch": 6.197163565321209, "grad_norm": 5.657242298126221, "learning_rate": 9.681552652778314e-06, "loss": 4.2838, "step": 1027750 }, { "epoch": 6.197465057102543, "grad_norm": 2.7193143367767334, "learning_rate": 9.674361758223903e-06, "loss": 4.1445, "step": 1027800 }, { "epoch": 6.197766548883878, "grad_norm": 6.861446857452393, "learning_rate": 9.667173446162613e-06, "loss": 3.8704, "step": 1027850 }, { "epoch": 6.198068040665212, "grad_norm": 4.155209541320801, "learning_rate": 9.659987716726752e-06, "loss": 4.5839, "step": 1027900 }, { "epoch": 6.198369532446545, "grad_norm": 3.8050894737243652, "learning_rate": 9.652804570048533e-06, "loss": 4.3955, "step": 1027950 }, { "epoch": 6.198671024227879, "grad_norm": 5.741901874542236, "learning_rate": 9.64562400626015e-06, "loss": 4.2322, "step": 1028000 }, { "epoch": 6.1989725160092135, "grad_norm": 2.3522841930389404, "learning_rate": 9.638446025493779e-06, "loss": 4.3603, "step": 1028050 }, { "epoch": 6.199274007790548, "grad_norm": 5.811081409454346, "learning_rate": 9.631270627881482e-06, "loss": 4.1779, "step": 1028100 }, { "epoch": 6.199575499571882, "grad_norm": 3.642357349395752, "learning_rate": 9.62409781355532e-06, "loss": 4.3906, "step": 1028150 }, { "epoch": 6.199876991353216, "grad_norm": 5.880804061889648, "learning_rate": 9.616927582647339e-06, "loss": 4.3801, "step": 1028200 }, { "epoch": 6.2001784831345494, "grad_norm": 3.6668200492858887, "learning_rate": 9.609759935289462e-06, "loss": 4.2771, "step": 1028250 }, { "epoch": 6.200479974915884, "grad_norm": 4.322079658508301, "learning_rate": 9.60259487161359e-06, "loss": 4.4327, "step": 1028300 }, { "epoch": 6.200781466697218, "grad_norm": 6.0549468994140625, "learning_rate": 9.595432391751629e-06, "loss": 4.7428, "step": 1028350 }, { "epoch": 6.201082958478552, "grad_norm": 3.843871831893921, "learning_rate": 9.588272495835358e-06, "loss": 4.2561, "step": 1028400 }, { "epoch": 6.201384450259886, "grad_norm": 3.9022128582000732, "learning_rate": 9.58111518399654e-06, "loss": 4.0421, "step": 1028450 }, { "epoch": 6.20168594204122, "grad_norm": 7.925064563751221, "learning_rate": 9.57396045636692e-06, "loss": 3.8963, "step": 1028500 }, { "epoch": 6.201987433822554, "grad_norm": 3.872749090194702, "learning_rate": 9.566808313078156e-06, "loss": 3.8461, "step": 1028550 }, { "epoch": 6.202288925603888, "grad_norm": 4.358555316925049, "learning_rate": 9.559658754261861e-06, "loss": 4.3566, "step": 1028600 }, { "epoch": 6.202590417385222, "grad_norm": 4.903500080108643, "learning_rate": 9.552511780049648e-06, "loss": 4.5953, "step": 1028650 }, { "epoch": 6.202891909166556, "grad_norm": 6.49113130569458, "learning_rate": 9.54536739057301e-06, "loss": 4.3408, "step": 1028700 }, { "epoch": 6.20319340094789, "grad_norm": 3.9750053882598877, "learning_rate": 9.538225585963477e-06, "loss": 3.8951, "step": 1028750 }, { "epoch": 6.203494892729224, "grad_norm": 5.597104072570801, "learning_rate": 9.531086366352442e-06, "loss": 4.0238, "step": 1028800 }, { "epoch": 6.203796384510558, "grad_norm": 3.856205701828003, "learning_rate": 9.523949731871267e-06, "loss": 4.1653, "step": 1028850 }, { "epoch": 6.204097876291892, "grad_norm": 4.441778182983398, "learning_rate": 9.516815682651346e-06, "loss": 4.4107, "step": 1028900 }, { "epoch": 6.204399368073227, "grad_norm": 5.115744590759277, "learning_rate": 9.509684218823977e-06, "loss": 4.2457, "step": 1028950 }, { "epoch": 6.204700859854561, "grad_norm": 4.314574241638184, "learning_rate": 9.502555340520351e-06, "loss": 4.5698, "step": 1029000 }, { "epoch": 6.204700859854561, "eval_loss": 4.831906795501709, "eval_runtime": 39.1817, "eval_samples_per_second": 13.067, "eval_steps_per_second": 6.534, "eval_tts_loss": 8.147083679587618, "step": 1029000 }, { "epoch": 6.205002351635894, "grad_norm": 12.145771980285645, "learning_rate": 9.495429047871683e-06, "loss": 4.288, "step": 1029050 }, { "epoch": 6.205303843417228, "grad_norm": 5.040600299835205, "learning_rate": 9.488305341009167e-06, "loss": 4.5081, "step": 1029100 }, { "epoch": 6.2056053351985625, "grad_norm": 3.451885938644409, "learning_rate": 9.481184220063864e-06, "loss": 4.097, "step": 1029150 }, { "epoch": 6.205906826979897, "grad_norm": 6.9745306968688965, "learning_rate": 9.47406568516682e-06, "loss": 4.7003, "step": 1029200 }, { "epoch": 6.206208318761231, "grad_norm": 3.023162603378296, "learning_rate": 9.46694973644908e-06, "loss": 3.8397, "step": 1029250 }, { "epoch": 6.206509810542564, "grad_norm": 4.899600982666016, "learning_rate": 9.45983637404154e-06, "loss": 4.1822, "step": 1029300 }, { "epoch": 6.206811302323898, "grad_norm": 6.007155418395996, "learning_rate": 9.452725598075162e-06, "loss": 4.3688, "step": 1029350 }, { "epoch": 6.207112794105233, "grad_norm": 6.581727027893066, "learning_rate": 9.445617408680822e-06, "loss": 4.1791, "step": 1029400 }, { "epoch": 6.207414285886567, "grad_norm": 7.673888683319092, "learning_rate": 9.438511805989284e-06, "loss": 3.7314, "step": 1029450 }, { "epoch": 6.207715777667901, "grad_norm": 2.603398561477661, "learning_rate": 9.431408790131345e-06, "loss": 4.2432, "step": 1029500 }, { "epoch": 6.208017269449234, "grad_norm": 4.227169990539551, "learning_rate": 9.424308361237748e-06, "loss": 4.3543, "step": 1029550 }, { "epoch": 6.208318761230569, "grad_norm": 5.002016067504883, "learning_rate": 9.417210519439139e-06, "loss": 4.6353, "step": 1029600 }, { "epoch": 6.208620253011903, "grad_norm": 3.291166067123413, "learning_rate": 9.410115264866131e-06, "loss": 3.9519, "step": 1029650 }, { "epoch": 6.208921744793237, "grad_norm": 5.9465508460998535, "learning_rate": 9.403022597649334e-06, "loss": 4.1637, "step": 1029700 }, { "epoch": 6.209223236574571, "grad_norm": 1.2755252122879028, "learning_rate": 9.395932517919246e-06, "loss": 4.0781, "step": 1029750 }, { "epoch": 6.209524728355905, "grad_norm": 5.861062049865723, "learning_rate": 9.38884502580638e-06, "loss": 4.6067, "step": 1029800 }, { "epoch": 6.209826220137239, "grad_norm": 4.179656028747559, "learning_rate": 9.381760121441146e-06, "loss": 4.2765, "step": 1029850 }, { "epoch": 6.210127711918573, "grad_norm": 4.041214942932129, "learning_rate": 9.374677804953955e-06, "loss": 4.2407, "step": 1029900 }, { "epoch": 6.210429203699907, "grad_norm": 6.457550048828125, "learning_rate": 9.367598076475158e-06, "loss": 4.3379, "step": 1029950 }, { "epoch": 6.210730695481241, "grad_norm": 10.041248321533203, "learning_rate": 9.360520936134997e-06, "loss": 4.3046, "step": 1030000 }, { "epoch": 6.2110321872625756, "grad_norm": 1.818291425704956, "learning_rate": 9.353446384063751e-06, "loss": 3.8517, "step": 1030050 }, { "epoch": 6.211333679043909, "grad_norm": 8.343231201171875, "learning_rate": 9.346374420391601e-06, "loss": 4.3872, "step": 1030100 }, { "epoch": 6.211635170825243, "grad_norm": 5.836264133453369, "learning_rate": 9.339305045248708e-06, "loss": 4.0574, "step": 1030150 }, { "epoch": 6.211936662606577, "grad_norm": 5.5241546630859375, "learning_rate": 9.332238258765202e-06, "loss": 4.2596, "step": 1030200 }, { "epoch": 6.2122381543879115, "grad_norm": 4.723347187042236, "learning_rate": 9.325174061071061e-06, "loss": 4.0846, "step": 1030250 }, { "epoch": 6.212539646169246, "grad_norm": 4.200168609619141, "learning_rate": 9.318112452296384e-06, "loss": 4.2335, "step": 1030300 }, { "epoch": 6.21284113795058, "grad_norm": 4.7680158615112305, "learning_rate": 9.31105343257108e-06, "loss": 4.2872, "step": 1030350 }, { "epoch": 6.213142629731913, "grad_norm": 5.164218902587891, "learning_rate": 9.303997002025032e-06, "loss": 4.0676, "step": 1030400 }, { "epoch": 6.213444121513247, "grad_norm": 3.9039344787597656, "learning_rate": 9.296943160788134e-06, "loss": 4.0427, "step": 1030450 }, { "epoch": 6.213745613294582, "grad_norm": 5.252764701843262, "learning_rate": 9.289891908990232e-06, "loss": 4.2177, "step": 1030500 }, { "epoch": 6.214047105075916, "grad_norm": 4.397813320159912, "learning_rate": 9.28284324676104e-06, "loss": 4.0227, "step": 1030550 }, { "epoch": 6.21434859685725, "grad_norm": 2.6117823123931885, "learning_rate": 9.275797174230303e-06, "loss": 4.4327, "step": 1030600 }, { "epoch": 6.214650088638583, "grad_norm": 5.135807991027832, "learning_rate": 9.26875369152772e-06, "loss": 4.1777, "step": 1030650 }, { "epoch": 6.214951580419918, "grad_norm": 4.053997039794922, "learning_rate": 9.261712798782867e-06, "loss": 4.2883, "step": 1030700 }, { "epoch": 6.215253072201252, "grad_norm": 2.6637744903564453, "learning_rate": 9.254674496125358e-06, "loss": 4.2446, "step": 1030750 }, { "epoch": 6.215554563982586, "grad_norm": 5.99411153793335, "learning_rate": 9.247638783684724e-06, "loss": 4.603, "step": 1030800 }, { "epoch": 6.21585605576392, "grad_norm": 3.771376848220825, "learning_rate": 9.240605661590411e-06, "loss": 4.2225, "step": 1030850 }, { "epoch": 6.2161575475452535, "grad_norm": 3.957967758178711, "learning_rate": 9.233575129971882e-06, "loss": 4.3648, "step": 1030900 }, { "epoch": 6.216459039326588, "grad_norm": 5.303887367248535, "learning_rate": 9.226547188958533e-06, "loss": 4.6602, "step": 1030950 }, { "epoch": 6.216760531107922, "grad_norm": 5.655281066894531, "learning_rate": 9.21952183867966e-06, "loss": 4.588, "step": 1031000 }, { "epoch": 6.217062022889256, "grad_norm": 6.665225982666016, "learning_rate": 9.212499079264578e-06, "loss": 4.5846, "step": 1031050 }, { "epoch": 6.21736351467059, "grad_norm": 7.600387096405029, "learning_rate": 9.205478910842584e-06, "loss": 4.0842, "step": 1031100 }, { "epoch": 6.2176650064519245, "grad_norm": 7.503591537475586, "learning_rate": 9.198461333542806e-06, "loss": 4.0362, "step": 1031150 }, { "epoch": 6.217966498233258, "grad_norm": 6.070474624633789, "learning_rate": 9.191446347494391e-06, "loss": 4.3109, "step": 1031200 }, { "epoch": 6.218267990014592, "grad_norm": 4.377626419067383, "learning_rate": 9.184433952826486e-06, "loss": 4.06, "step": 1031250 }, { "epoch": 6.218569481795926, "grad_norm": 4.548770904541016, "learning_rate": 9.177424149668106e-06, "loss": 4.3332, "step": 1031300 }, { "epoch": 6.2188709735772605, "grad_norm": 5.672216415405273, "learning_rate": 9.170416938148261e-06, "loss": 4.0653, "step": 1031350 }, { "epoch": 6.219172465358595, "grad_norm": 4.058233261108398, "learning_rate": 9.163412318395935e-06, "loss": 4.5021, "step": 1031400 }, { "epoch": 6.219473957139928, "grad_norm": 6.1607770919799805, "learning_rate": 9.156410290540005e-06, "loss": 4.4838, "step": 1031450 }, { "epoch": 6.219775448921262, "grad_norm": 3.90726375579834, "learning_rate": 9.149410854709371e-06, "loss": 4.3017, "step": 1031500 }, { "epoch": 6.220076940702596, "grad_norm": 3.633755683898926, "learning_rate": 9.142414011032812e-06, "loss": 3.7628, "step": 1031550 }, { "epoch": 6.220378432483931, "grad_norm": 4.927196502685547, "learning_rate": 9.135419759639123e-06, "loss": 3.9941, "step": 1031600 }, { "epoch": 6.220679924265265, "grad_norm": 5.8642401695251465, "learning_rate": 9.128428100656986e-06, "loss": 3.9991, "step": 1031650 }, { "epoch": 6.220981416046598, "grad_norm": 3.2351572513580322, "learning_rate": 9.121439034215117e-06, "loss": 4.0611, "step": 1031700 }, { "epoch": 6.221282907827932, "grad_norm": 4.110411167144775, "learning_rate": 9.114452560442093e-06, "loss": 4.4367, "step": 1031750 }, { "epoch": 6.221584399609267, "grad_norm": 2.3557870388031006, "learning_rate": 9.107468679466513e-06, "loss": 4.4727, "step": 1031800 }, { "epoch": 6.221885891390601, "grad_norm": 6.176474094390869, "learning_rate": 9.100487391416938e-06, "loss": 4.2341, "step": 1031850 }, { "epoch": 6.222187383171935, "grad_norm": 2.518134593963623, "learning_rate": 9.093508696421786e-06, "loss": 3.9763, "step": 1031900 }, { "epoch": 6.222488874953269, "grad_norm": 4.987135887145996, "learning_rate": 9.086532594609552e-06, "loss": 4.2423, "step": 1031950 }, { "epoch": 6.2227903667346025, "grad_norm": 1.9137458801269531, "learning_rate": 9.079559086108602e-06, "loss": 4.0008, "step": 1032000 }, { "epoch": 6.2227903667346025, "eval_loss": 4.8299760818481445, "eval_runtime": 39.2136, "eval_samples_per_second": 13.057, "eval_steps_per_second": 6.528, "eval_tts_loss": 8.113969727089144, "step": 1032000 }, { "epoch": 6.223091858515937, "grad_norm": 5.405186176300049, "learning_rate": 9.072588171047229e-06, "loss": 4.2904, "step": 1032050 }, { "epoch": 6.223393350297271, "grad_norm": 5.6356048583984375, "learning_rate": 9.065619849553752e-06, "loss": 4.3335, "step": 1032100 }, { "epoch": 6.223694842078605, "grad_norm": 5.315788269042969, "learning_rate": 9.058654121756447e-06, "loss": 4.3298, "step": 1032150 }, { "epoch": 6.223996333859939, "grad_norm": 8.07382869720459, "learning_rate": 9.05169098778345e-06, "loss": 4.223, "step": 1032200 }, { "epoch": 6.224297825641273, "grad_norm": 2.99153733253479, "learning_rate": 9.044730447762938e-06, "loss": 3.9642, "step": 1032250 }, { "epoch": 6.224599317422607, "grad_norm": 5.634726047515869, "learning_rate": 9.037772501823043e-06, "loss": 4.0706, "step": 1032300 }, { "epoch": 6.224900809203941, "grad_norm": 3.6676547527313232, "learning_rate": 9.030817150091763e-06, "loss": 4.191, "step": 1032350 }, { "epoch": 6.225202300985275, "grad_norm": 5.350943565368652, "learning_rate": 9.02386439269711e-06, "loss": 4.3727, "step": 1032400 }, { "epoch": 6.2255037927666095, "grad_norm": 3.2789273262023926, "learning_rate": 9.016914229767069e-06, "loss": 3.9125, "step": 1032450 }, { "epoch": 6.225805284547943, "grad_norm": 4.372374057769775, "learning_rate": 9.009966661429503e-06, "loss": 4.4801, "step": 1032500 }, { "epoch": 6.226106776329277, "grad_norm": 4.29016637802124, "learning_rate": 9.003021687812289e-06, "loss": 4.5905, "step": 1032550 }, { "epoch": 6.226408268110611, "grad_norm": 6.3774027824401855, "learning_rate": 8.99607930904328e-06, "loss": 4.2634, "step": 1032600 }, { "epoch": 6.226709759891945, "grad_norm": 6.115050315856934, "learning_rate": 8.989139525250189e-06, "loss": 4.2975, "step": 1032650 }, { "epoch": 6.22701125167328, "grad_norm": 5.251617908477783, "learning_rate": 8.982202336560746e-06, "loss": 4.2115, "step": 1032700 }, { "epoch": 6.227312743454614, "grad_norm": 10.394024848937988, "learning_rate": 8.975267743102648e-06, "loss": 4.3505, "step": 1032750 }, { "epoch": 6.227614235235947, "grad_norm": 4.535055160522461, "learning_rate": 8.968335745003496e-06, "loss": 4.5233, "step": 1032800 }, { "epoch": 6.227915727017281, "grad_norm": 3.856729507446289, "learning_rate": 8.961406342390836e-06, "loss": 3.8289, "step": 1032850 }, { "epoch": 6.228217218798616, "grad_norm": 4.890798091888428, "learning_rate": 8.954479535392251e-06, "loss": 3.9414, "step": 1032900 }, { "epoch": 6.22851871057995, "grad_norm": 4.70316743850708, "learning_rate": 8.947555324135152e-06, "loss": 4.1051, "step": 1032950 }, { "epoch": 6.228820202361284, "grad_norm": 3.572298765182495, "learning_rate": 8.940633708747008e-06, "loss": 4.2163, "step": 1033000 }, { "epoch": 6.229121694142617, "grad_norm": 2.242886543273926, "learning_rate": 8.933714689355232e-06, "loss": 4.1212, "step": 1033050 }, { "epoch": 6.2294231859239515, "grad_norm": 5.085737228393555, "learning_rate": 8.926798266087088e-06, "loss": 4.3727, "step": 1033100 }, { "epoch": 6.229724677705286, "grad_norm": 5.462203025817871, "learning_rate": 8.919884439069925e-06, "loss": 4.2961, "step": 1033150 }, { "epoch": 6.23002616948662, "grad_norm": 3.9939069747924805, "learning_rate": 8.912973208430956e-06, "loss": 4.4442, "step": 1033200 }, { "epoch": 6.230327661267954, "grad_norm": 7.519163608551025, "learning_rate": 8.90606457429735e-06, "loss": 4.1722, "step": 1033250 }, { "epoch": 6.230629153049287, "grad_norm": 3.9738876819610596, "learning_rate": 8.899158536796286e-06, "loss": 4.1125, "step": 1033300 }, { "epoch": 6.230930644830622, "grad_norm": 1.941313624382019, "learning_rate": 8.892255096054829e-06, "loss": 4.1923, "step": 1033350 }, { "epoch": 6.231232136611956, "grad_norm": 4.538352966308594, "learning_rate": 8.885354252200078e-06, "loss": 4.416, "step": 1033400 }, { "epoch": 6.23153362839329, "grad_norm": 4.660863399505615, "learning_rate": 8.87845600535898e-06, "loss": 4.3311, "step": 1033450 }, { "epoch": 6.231835120174624, "grad_norm": 5.21674108505249, "learning_rate": 8.871560355658535e-06, "loss": 4.444, "step": 1033500 }, { "epoch": 6.2321366119559585, "grad_norm": 4.402937412261963, "learning_rate": 8.86466730322561e-06, "loss": 4.2551, "step": 1033550 }, { "epoch": 6.232438103737292, "grad_norm": 3.1355485916137695, "learning_rate": 8.857776848187048e-06, "loss": 4.3672, "step": 1033600 }, { "epoch": 6.232739595518626, "grad_norm": 4.246630668640137, "learning_rate": 8.850888990669685e-06, "loss": 4.2179, "step": 1033650 }, { "epoch": 6.23304108729996, "grad_norm": 4.934570789337158, "learning_rate": 8.84400373080032e-06, "loss": 4.1912, "step": 1033700 }, { "epoch": 6.233342579081294, "grad_norm": 6.493074417114258, "learning_rate": 8.837121068705583e-06, "loss": 4.1566, "step": 1033750 }, { "epoch": 6.233644070862629, "grad_norm": 5.562867641448975, "learning_rate": 8.83024100451219e-06, "loss": 4.2389, "step": 1033800 }, { "epoch": 6.233945562643962, "grad_norm": 4.609105587005615, "learning_rate": 8.823363538346773e-06, "loss": 4.0329, "step": 1033850 }, { "epoch": 6.234247054425296, "grad_norm": 7.373548984527588, "learning_rate": 8.816488670335847e-06, "loss": 3.8436, "step": 1033900 }, { "epoch": 6.23454854620663, "grad_norm": 6.745499610900879, "learning_rate": 8.809616400605996e-06, "loss": 4.1643, "step": 1033950 }, { "epoch": 6.2348500379879646, "grad_norm": 9.00551700592041, "learning_rate": 8.802746729283665e-06, "loss": 4.1803, "step": 1034000 }, { "epoch": 6.235151529769299, "grad_norm": 5.460381984710693, "learning_rate": 8.795879656495275e-06, "loss": 4.0866, "step": 1034050 }, { "epoch": 6.235453021550633, "grad_norm": 5.977598190307617, "learning_rate": 8.789015182367187e-06, "loss": 4.26, "step": 1034100 }, { "epoch": 6.235754513331966, "grad_norm": 8.334343910217285, "learning_rate": 8.782153307025803e-06, "loss": 4.1834, "step": 1034150 }, { "epoch": 6.2360560051133005, "grad_norm": 4.841203689575195, "learning_rate": 8.775294030597318e-06, "loss": 3.9888, "step": 1034200 }, { "epoch": 6.236357496894635, "grad_norm": 5.0341691970825195, "learning_rate": 8.76843735320802e-06, "loss": 4.2041, "step": 1034250 }, { "epoch": 6.236658988675969, "grad_norm": 5.189927577972412, "learning_rate": 8.761583274984086e-06, "loss": 4.0723, "step": 1034300 }, { "epoch": 6.236960480457303, "grad_norm": 5.806893348693848, "learning_rate": 8.754731796051666e-06, "loss": 4.2912, "step": 1034350 }, { "epoch": 6.237261972238636, "grad_norm": 3.9084460735321045, "learning_rate": 8.74788291653683e-06, "loss": 4.1523, "step": 1034400 }, { "epoch": 6.237563464019971, "grad_norm": 5.226700782775879, "learning_rate": 8.741036636565641e-06, "loss": 4.4811, "step": 1034450 }, { "epoch": 6.237864955801305, "grad_norm": 6.2336626052856445, "learning_rate": 8.734192956264063e-06, "loss": 4.1789, "step": 1034500 }, { "epoch": 6.238166447582639, "grad_norm": 3.6057727336883545, "learning_rate": 8.727351875758065e-06, "loss": 4.1894, "step": 1034550 }, { "epoch": 6.238467939363973, "grad_norm": 4.098724365234375, "learning_rate": 8.720513395173561e-06, "loss": 4.5238, "step": 1034600 }, { "epoch": 6.238769431145307, "grad_norm": 4.103440761566162, "learning_rate": 8.713677514636387e-06, "loss": 4.1187, "step": 1034650 }, { "epoch": 6.239070922926641, "grad_norm": 4.37543249130249, "learning_rate": 8.70684423427237e-06, "loss": 4.0816, "step": 1034700 }, { "epoch": 6.239372414707975, "grad_norm": 5.237444877624512, "learning_rate": 8.700013554207247e-06, "loss": 4.6132, "step": 1034750 }, { "epoch": 6.239673906489309, "grad_norm": 2.6161139011383057, "learning_rate": 8.693185474566716e-06, "loss": 4.3053, "step": 1034800 }, { "epoch": 6.239975398270643, "grad_norm": 4.316936016082764, "learning_rate": 8.686359995476427e-06, "loss": 4.1255, "step": 1034850 }, { "epoch": 6.240276890051978, "grad_norm": 2.9653401374816895, "learning_rate": 8.679537117062062e-06, "loss": 4.0612, "step": 1034900 }, { "epoch": 6.240578381833311, "grad_norm": 4.601945877075195, "learning_rate": 8.672716839449107e-06, "loss": 4.4251, "step": 1034950 }, { "epoch": 6.240879873614645, "grad_norm": 6.130619525909424, "learning_rate": 8.665899162763124e-06, "loss": 4.2312, "step": 1035000 }, { "epoch": 6.240879873614645, "eval_loss": 4.826171875, "eval_runtime": 39.1411, "eval_samples_per_second": 13.081, "eval_steps_per_second": 6.54, "eval_tts_loss": 8.10722230072351, "step": 1035000 }, { "epoch": 6.241181365395979, "grad_norm": 6.64235782623291, "learning_rate": 8.659084087129581e-06, "loss": 4.2868, "step": 1035050 }, { "epoch": 6.2414828571773135, "grad_norm": 5.275937080383301, "learning_rate": 8.652271612673878e-06, "loss": 4.3005, "step": 1035100 }, { "epoch": 6.241784348958648, "grad_norm": 6.178396701812744, "learning_rate": 8.645461739521414e-06, "loss": 4.4238, "step": 1035150 }, { "epoch": 6.242085840739981, "grad_norm": 5.340651035308838, "learning_rate": 8.638654467797523e-06, "loss": 4.2134, "step": 1035200 }, { "epoch": 6.242387332521315, "grad_norm": 3.693875551223755, "learning_rate": 8.631849797627421e-06, "loss": 4.0767, "step": 1035250 }, { "epoch": 6.2426888243026495, "grad_norm": 5.63420295715332, "learning_rate": 8.625047729136391e-06, "loss": 4.0676, "step": 1035300 }, { "epoch": 6.242990316083984, "grad_norm": 2.5330705642700195, "learning_rate": 8.618248262449618e-06, "loss": 4.2827, "step": 1035350 }, { "epoch": 6.243291807865318, "grad_norm": 8.428057670593262, "learning_rate": 8.611451397692214e-06, "loss": 4.2623, "step": 1035400 }, { "epoch": 6.243593299646651, "grad_norm": 5.803028106689453, "learning_rate": 8.604657134989267e-06, "loss": 4.3485, "step": 1035450 }, { "epoch": 6.243894791427985, "grad_norm": 3.867065906524658, "learning_rate": 8.597865474465843e-06, "loss": 3.8878, "step": 1035500 }, { "epoch": 6.24419628320932, "grad_norm": 3.2928662300109863, "learning_rate": 8.591076416246923e-06, "loss": 4.4234, "step": 1035550 }, { "epoch": 6.244497774990654, "grad_norm": 4.13867712020874, "learning_rate": 8.584289960457408e-06, "loss": 4.157, "step": 1035600 }, { "epoch": 6.244799266771988, "grad_norm": 6.056456565856934, "learning_rate": 8.577506107222249e-06, "loss": 4.4062, "step": 1035650 }, { "epoch": 6.245100758553322, "grad_norm": 4.617682933807373, "learning_rate": 8.570724856666244e-06, "loss": 4.3898, "step": 1035700 }, { "epoch": 6.245402250334656, "grad_norm": 4.972387790679932, "learning_rate": 8.563946208914229e-06, "loss": 4.018, "step": 1035750 }, { "epoch": 6.24570374211599, "grad_norm": 2.8412139415740967, "learning_rate": 8.557170164090954e-06, "loss": 4.32, "step": 1035800 }, { "epoch": 6.246005233897324, "grad_norm": 5.0980939865112305, "learning_rate": 8.550396722321101e-06, "loss": 4.1099, "step": 1035850 }, { "epoch": 6.246306725678658, "grad_norm": 1.9550269842147827, "learning_rate": 8.543625883729354e-06, "loss": 4.1157, "step": 1035900 }, { "epoch": 6.246608217459992, "grad_norm": 4.469468593597412, "learning_rate": 8.536857648440281e-06, "loss": 4.1862, "step": 1035950 }, { "epoch": 6.246909709241326, "grad_norm": 5.198384761810303, "learning_rate": 8.530092016578483e-06, "loss": 4.5243, "step": 1036000 }, { "epoch": 6.24721120102266, "grad_norm": 6.767685413360596, "learning_rate": 8.52332898826844e-06, "loss": 4.3623, "step": 1036050 }, { "epoch": 6.247512692803994, "grad_norm": 6.613608360290527, "learning_rate": 8.51656856363464e-06, "loss": 4.0274, "step": 1036100 }, { "epoch": 6.247814184585328, "grad_norm": 5.1597113609313965, "learning_rate": 8.509810742801466e-06, "loss": 3.9313, "step": 1036150 }, { "epoch": 6.2481156763666625, "grad_norm": 4.445883274078369, "learning_rate": 8.503055525893316e-06, "loss": 3.8133, "step": 1036200 }, { "epoch": 6.248417168147996, "grad_norm": 5.045104503631592, "learning_rate": 8.496302913034524e-06, "loss": 4.3426, "step": 1036250 }, { "epoch": 6.24871865992933, "grad_norm": 4.092502593994141, "learning_rate": 8.489552904349312e-06, "loss": 4.4328, "step": 1036300 }, { "epoch": 6.249020151710664, "grad_norm": 5.7991790771484375, "learning_rate": 8.48280549996196e-06, "loss": 4.0221, "step": 1036350 }, { "epoch": 6.2493216434919985, "grad_norm": 4.154432773590088, "learning_rate": 8.476060699996617e-06, "loss": 4.2452, "step": 1036400 }, { "epoch": 6.249623135273333, "grad_norm": 4.4541916847229, "learning_rate": 8.469318504577387e-06, "loss": 4.1175, "step": 1036450 }, { "epoch": 6.249924627054667, "grad_norm": 5.124157428741455, "learning_rate": 8.462578913828372e-06, "loss": 4.1498, "step": 1036500 }, { "epoch": 6.250226118836, "grad_norm": 5.473306655883789, "learning_rate": 8.455841927873636e-06, "loss": 4.2157, "step": 1036550 }, { "epoch": 6.250527610617334, "grad_norm": 2.4464545249938965, "learning_rate": 8.449107546837114e-06, "loss": 4.2564, "step": 1036600 }, { "epoch": 6.250829102398669, "grad_norm": 4.135465621948242, "learning_rate": 8.442375770842758e-06, "loss": 4.1479, "step": 1036650 }, { "epoch": 6.251130594180003, "grad_norm": 4.044125556945801, "learning_rate": 8.435646600014484e-06, "loss": 4.2987, "step": 1036700 }, { "epoch": 6.251432085961337, "grad_norm": 4.580842971801758, "learning_rate": 8.42892003447611e-06, "loss": 4.0685, "step": 1036750 }, { "epoch": 6.25173357774267, "grad_norm": 5.035514831542969, "learning_rate": 8.422196074351406e-06, "loss": 4.4148, "step": 1036800 }, { "epoch": 6.252035069524005, "grad_norm": 4.378883361816406, "learning_rate": 8.415474719764153e-06, "loss": 4.1411, "step": 1036850 }, { "epoch": 6.252336561305339, "grad_norm": 5.130667686462402, "learning_rate": 8.408755970838054e-06, "loss": 4.3792, "step": 1036900 }, { "epoch": 6.252638053086673, "grad_norm": 4.051558494567871, "learning_rate": 8.40203982769671e-06, "loss": 4.3012, "step": 1036950 }, { "epoch": 6.252939544868007, "grad_norm": 4.279202938079834, "learning_rate": 8.395326290463756e-06, "loss": 4.448, "step": 1037000 }, { "epoch": 6.2532410366493405, "grad_norm": 5.462879180908203, "learning_rate": 8.388615359262757e-06, "loss": 4.3381, "step": 1037050 }, { "epoch": 6.253542528430675, "grad_norm": 5.220454692840576, "learning_rate": 8.381907034217217e-06, "loss": 4.179, "step": 1037100 }, { "epoch": 6.253844020212009, "grad_norm": 5.535382270812988, "learning_rate": 8.375201315450536e-06, "loss": 4.3464, "step": 1037150 }, { "epoch": 6.254145511993343, "grad_norm": 4.942020893096924, "learning_rate": 8.368498203086199e-06, "loss": 4.2457, "step": 1037200 }, { "epoch": 6.254447003774677, "grad_norm": 4.544367790222168, "learning_rate": 8.361797697247507e-06, "loss": 4.0975, "step": 1037250 }, { "epoch": 6.2547484955560115, "grad_norm": 3.3444998264312744, "learning_rate": 8.355099798057812e-06, "loss": 4.1891, "step": 1037300 }, { "epoch": 6.255049987337345, "grad_norm": 4.318490505218506, "learning_rate": 8.348404505640382e-06, "loss": 3.9741, "step": 1037350 }, { "epoch": 6.255351479118679, "grad_norm": 4.5982537269592285, "learning_rate": 8.341711820118403e-06, "loss": 4.5005, "step": 1037400 }, { "epoch": 6.255652970900013, "grad_norm": 6.18479585647583, "learning_rate": 8.335021741615038e-06, "loss": 4.2436, "step": 1037450 }, { "epoch": 6.2559544626813475, "grad_norm": 4.173498630523682, "learning_rate": 8.328334270253478e-06, "loss": 4.3302, "step": 1037500 }, { "epoch": 6.256255954462682, "grad_norm": 5.034013748168945, "learning_rate": 8.32164940615674e-06, "loss": 4.1198, "step": 1037550 }, { "epoch": 6.256557446244015, "grad_norm": 3.508131742477417, "learning_rate": 8.314967149447837e-06, "loss": 4.5647, "step": 1037600 }, { "epoch": 6.256858938025349, "grad_norm": 5.500007629394531, "learning_rate": 8.308287500249794e-06, "loss": 4.1706, "step": 1037650 }, { "epoch": 6.257160429806683, "grad_norm": 6.3542046546936035, "learning_rate": 8.301610458685493e-06, "loss": 4.419, "step": 1037700 }, { "epoch": 6.257461921588018, "grad_norm": 6.296797752380371, "learning_rate": 8.294936024877818e-06, "loss": 4.1452, "step": 1037750 }, { "epoch": 6.257763413369352, "grad_norm": 5.120035648345947, "learning_rate": 8.288264198949674e-06, "loss": 4.0036, "step": 1037800 }, { "epoch": 6.258064905150686, "grad_norm": 5.0032548904418945, "learning_rate": 8.281594981023743e-06, "loss": 4.3161, "step": 1037850 }, { "epoch": 6.258366396932019, "grad_norm": 4.709199905395508, "learning_rate": 8.274928371222861e-06, "loss": 4.5278, "step": 1037900 }, { "epoch": 6.2586678887133536, "grad_norm": 4.152355670928955, "learning_rate": 8.268264369669647e-06, "loss": 4.3923, "step": 1037950 }, { "epoch": 6.258969380494688, "grad_norm": 5.2501091957092285, "learning_rate": 8.261602976486754e-06, "loss": 4.5761, "step": 1038000 }, { "epoch": 6.258969380494688, "eval_loss": 4.821338653564453, "eval_runtime": 38.9885, "eval_samples_per_second": 13.132, "eval_steps_per_second": 6.566, "eval_tts_loss": 8.145760574780775, "step": 1038000 }, { "epoch": 6.259270872276022, "grad_norm": 5.112664222717285, "learning_rate": 8.254944191796797e-06, "loss": 4.1991, "step": 1038050 }, { "epoch": 6.259572364057356, "grad_norm": 6.07472038269043, "learning_rate": 8.248288015722315e-06, "loss": 3.7978, "step": 1038100 }, { "epoch": 6.2598738558386895, "grad_norm": 5.979792594909668, "learning_rate": 8.241634448385792e-06, "loss": 4.3815, "step": 1038150 }, { "epoch": 6.260175347620024, "grad_norm": 4.086608409881592, "learning_rate": 8.23498348990968e-06, "loss": 4.301, "step": 1038200 }, { "epoch": 6.260476839401358, "grad_norm": 5.479361534118652, "learning_rate": 8.228335140416414e-06, "loss": 3.9897, "step": 1038250 }, { "epoch": 6.260778331182692, "grad_norm": 6.096217632293701, "learning_rate": 8.221689400028297e-06, "loss": 4.0911, "step": 1038300 }, { "epoch": 6.261079822964026, "grad_norm": 7.625225067138672, "learning_rate": 8.215046268867681e-06, "loss": 4.1802, "step": 1038350 }, { "epoch": 6.26138131474536, "grad_norm": 3.9349517822265625, "learning_rate": 8.208405747056784e-06, "loss": 4.227, "step": 1038400 }, { "epoch": 6.261682806526694, "grad_norm": 4.971241474151611, "learning_rate": 8.201767834717827e-06, "loss": 4.1902, "step": 1038450 }, { "epoch": 6.261984298308028, "grad_norm": 4.903191089630127, "learning_rate": 8.195132531972976e-06, "loss": 4.4261, "step": 1038500 }, { "epoch": 6.262285790089362, "grad_norm": 4.991570472717285, "learning_rate": 8.188499838944367e-06, "loss": 4.3474, "step": 1038550 }, { "epoch": 6.2625872818706965, "grad_norm": 2.642958402633667, "learning_rate": 8.181869755754023e-06, "loss": 4.4152, "step": 1038600 }, { "epoch": 6.26288877365203, "grad_norm": 7.09555721282959, "learning_rate": 8.175242282523975e-06, "loss": 4.5389, "step": 1038650 }, { "epoch": 6.263190265433364, "grad_norm": 8.129949569702148, "learning_rate": 8.16861741937621e-06, "loss": 4.1875, "step": 1038700 }, { "epoch": 6.263491757214698, "grad_norm": 6.344634056091309, "learning_rate": 8.16199516643265e-06, "loss": 4.5719, "step": 1038750 }, { "epoch": 6.263793248996032, "grad_norm": 2.666029691696167, "learning_rate": 8.155375523815127e-06, "loss": 4.1177, "step": 1038800 }, { "epoch": 6.264094740777367, "grad_norm": 6.892940521240234, "learning_rate": 8.14875849164553e-06, "loss": 4.1938, "step": 1038850 }, { "epoch": 6.264396232558701, "grad_norm": 5.10064172744751, "learning_rate": 8.142144070045559e-06, "loss": 4.1287, "step": 1038900 }, { "epoch": 6.264697724340034, "grad_norm": 5.162724494934082, "learning_rate": 8.135532259137e-06, "loss": 3.7113, "step": 1038950 }, { "epoch": 6.264999216121368, "grad_norm": 2.579457998275757, "learning_rate": 8.128923059041526e-06, "loss": 4.4925, "step": 1039000 }, { "epoch": 6.2653007079027025, "grad_norm": 6.948094844818115, "learning_rate": 8.122316469880752e-06, "loss": 4.1119, "step": 1039050 }, { "epoch": 6.265602199684037, "grad_norm": 7.126020908355713, "learning_rate": 8.1157124917763e-06, "loss": 4.3734, "step": 1039100 }, { "epoch": 6.265903691465371, "grad_norm": 5.393867015838623, "learning_rate": 8.109111124849672e-06, "loss": 4.2614, "step": 1039150 }, { "epoch": 6.266205183246704, "grad_norm": 4.131092548370361, "learning_rate": 8.102512369222336e-06, "loss": 4.3269, "step": 1039200 }, { "epoch": 6.2665066750280385, "grad_norm": 4.118861675262451, "learning_rate": 8.095916225015765e-06, "loss": 4.3978, "step": 1039250 }, { "epoch": 6.266808166809373, "grad_norm": 4.3352742195129395, "learning_rate": 8.089322692351374e-06, "loss": 4.3497, "step": 1039300 }, { "epoch": 6.267109658590707, "grad_norm": 3.91162109375, "learning_rate": 8.082731771350453e-06, "loss": 4.3809, "step": 1039350 }, { "epoch": 6.267411150372041, "grad_norm": 7.9135942459106445, "learning_rate": 8.076143462134339e-06, "loss": 4.3819, "step": 1039400 }, { "epoch": 6.267712642153375, "grad_norm": 4.232983589172363, "learning_rate": 8.06955776482428e-06, "loss": 4.2187, "step": 1039450 }, { "epoch": 6.268014133934709, "grad_norm": 3.647603988647461, "learning_rate": 8.062974679541451e-06, "loss": 3.6826, "step": 1039500 }, { "epoch": 6.268315625716043, "grad_norm": 3.1255416870117188, "learning_rate": 8.056394206407018e-06, "loss": 3.8012, "step": 1039550 }, { "epoch": 6.268617117497377, "grad_norm": 3.0931766033172607, "learning_rate": 8.049816345542104e-06, "loss": 4.2814, "step": 1039600 }, { "epoch": 6.268918609278711, "grad_norm": 5.415017127990723, "learning_rate": 8.043241097067726e-06, "loss": 4.0498, "step": 1039650 }, { "epoch": 6.2692201010600455, "grad_norm": 4.776140213012695, "learning_rate": 8.036668461104905e-06, "loss": 4.2443, "step": 1039700 }, { "epoch": 6.269521592841379, "grad_norm": 5.525318145751953, "learning_rate": 8.030098437774646e-06, "loss": 4.3051, "step": 1039750 }, { "epoch": 6.269823084622713, "grad_norm": 1.883322834968567, "learning_rate": 8.023531027197783e-06, "loss": 4.3375, "step": 1039800 }, { "epoch": 6.270124576404047, "grad_norm": 6.425246238708496, "learning_rate": 8.01696622949522e-06, "loss": 4.7066, "step": 1039850 }, { "epoch": 6.270426068185381, "grad_norm": 7.614790439605713, "learning_rate": 8.01040404478781e-06, "loss": 4.0001, "step": 1039900 }, { "epoch": 6.270727559966716, "grad_norm": 11.651505470275879, "learning_rate": 8.003844473196276e-06, "loss": 4.112, "step": 1039950 }, { "epoch": 6.27102905174805, "grad_norm": 3.6683576107025146, "learning_rate": 7.997287514841316e-06, "loss": 4.0081, "step": 1040000 }, { "epoch": 6.271330543529383, "grad_norm": 5.892425060272217, "learning_rate": 7.990733169843638e-06, "loss": 4.3499, "step": 1040050 }, { "epoch": 6.271632035310717, "grad_norm": 7.46647834777832, "learning_rate": 7.984181438323877e-06, "loss": 4.0127, "step": 1040100 }, { "epoch": 6.2719335270920515, "grad_norm": 4.474137783050537, "learning_rate": 7.977632320402571e-06, "loss": 3.9752, "step": 1040150 }, { "epoch": 6.272235018873386, "grad_norm": 2.9952704906463623, "learning_rate": 7.971085816200256e-06, "loss": 4.0517, "step": 1040200 }, { "epoch": 6.27253651065472, "grad_norm": 5.034274101257324, "learning_rate": 7.964541925837454e-06, "loss": 4.3427, "step": 1040250 }, { "epoch": 6.272838002436053, "grad_norm": 4.870392799377441, "learning_rate": 7.958000649434564e-06, "loss": 3.9464, "step": 1040300 }, { "epoch": 6.2731394942173875, "grad_norm": 6.2111639976501465, "learning_rate": 7.951461987111929e-06, "loss": 4.0539, "step": 1040350 }, { "epoch": 6.273440985998722, "grad_norm": 6.3677544593811035, "learning_rate": 7.944925938989948e-06, "loss": 4.3857, "step": 1040400 }, { "epoch": 6.273742477780056, "grad_norm": 9.722972869873047, "learning_rate": 7.93839250518888e-06, "loss": 4.1168, "step": 1040450 }, { "epoch": 6.27404396956139, "grad_norm": 6.511646270751953, "learning_rate": 7.931861685828938e-06, "loss": 3.7563, "step": 1040500 }, { "epoch": 6.274345461342723, "grad_norm": 6.6359357833862305, "learning_rate": 7.925333481030382e-06, "loss": 4.2497, "step": 1040550 }, { "epoch": 6.274646953124058, "grad_norm": 2.637925863265991, "learning_rate": 7.918807890913281e-06, "loss": 4.6406, "step": 1040600 }, { "epoch": 6.274948444905392, "grad_norm": 4.4683918952941895, "learning_rate": 7.912284915597755e-06, "loss": 3.9172, "step": 1040650 }, { "epoch": 6.275249936686726, "grad_norm": 4.042970657348633, "learning_rate": 7.905764555203891e-06, "loss": 4.4883, "step": 1040700 }, { "epoch": 6.27555142846806, "grad_norm": 4.083834171295166, "learning_rate": 7.899246809851645e-06, "loss": 4.1839, "step": 1040750 }, { "epoch": 6.275852920249394, "grad_norm": 1.89656662940979, "learning_rate": 7.892731679660953e-06, "loss": 4.2379, "step": 1040800 }, { "epoch": 6.276154412030728, "grad_norm": 4.006491661071777, "learning_rate": 7.88621916475175e-06, "loss": 4.3051, "step": 1040850 }, { "epoch": 6.276455903812062, "grad_norm": 5.14767599105835, "learning_rate": 7.879709265243861e-06, "loss": 4.2494, "step": 1040900 }, { "epoch": 6.276757395593396, "grad_norm": 2.1869914531707764, "learning_rate": 7.873201981257089e-06, "loss": 4.2991, "step": 1040950 }, { "epoch": 6.27705888737473, "grad_norm": 6.357082366943359, "learning_rate": 7.866697312911252e-06, "loss": 4.6704, "step": 1041000 }, { "epoch": 6.27705888737473, "eval_loss": 4.821949481964111, "eval_runtime": 39.1206, "eval_samples_per_second": 13.088, "eval_steps_per_second": 6.544, "eval_tts_loss": 8.14311330942914, "step": 1041000 }, { "epoch": 6.277360379156065, "grad_norm": 4.9667840003967285, "learning_rate": 7.86019526032599e-06, "loss": 3.9248, "step": 1041050 }, { "epoch": 6.277661870937398, "grad_norm": 4.506594657897949, "learning_rate": 7.85369582362101e-06, "loss": 4.2664, "step": 1041100 }, { "epoch": 6.277963362718732, "grad_norm": 6.333621978759766, "learning_rate": 7.847199002915895e-06, "loss": 4.0631, "step": 1041150 }, { "epoch": 6.278264854500066, "grad_norm": 2.826619863510132, "learning_rate": 7.840704798330217e-06, "loss": 4.2981, "step": 1041200 }, { "epoch": 6.2785663462814005, "grad_norm": 6.443743705749512, "learning_rate": 7.83421320998348e-06, "loss": 4.4345, "step": 1041250 }, { "epoch": 6.278867838062735, "grad_norm": 4.035067081451416, "learning_rate": 7.827724237995193e-06, "loss": 4.3212, "step": 1041300 }, { "epoch": 6.279169329844068, "grad_norm": 3.1052346229553223, "learning_rate": 7.82123788248472e-06, "loss": 4.1234, "step": 1041350 }, { "epoch": 6.279470821625402, "grad_norm": 4.361624717712402, "learning_rate": 7.81475414357149e-06, "loss": 4.3916, "step": 1041400 }, { "epoch": 6.2797723134067365, "grad_norm": 9.584028244018555, "learning_rate": 7.8082730213748e-06, "loss": 4.115, "step": 1041450 }, { "epoch": 6.280073805188071, "grad_norm": 5.56730842590332, "learning_rate": 7.801794516013926e-06, "loss": 4.4482, "step": 1041500 }, { "epoch": 6.280375296969405, "grad_norm": 3.9114437103271484, "learning_rate": 7.795318627608089e-06, "loss": 4.3007, "step": 1041550 }, { "epoch": 6.280676788750739, "grad_norm": 5.516161918640137, "learning_rate": 7.788845356276492e-06, "loss": 4.1986, "step": 1041600 }, { "epoch": 6.280978280532072, "grad_norm": 6.041079044342041, "learning_rate": 7.782374702138222e-06, "loss": 4.4149, "step": 1041650 }, { "epoch": 6.281279772313407, "grad_norm": 6.120043754577637, "learning_rate": 7.775906665312387e-06, "loss": 4.7879, "step": 1041700 }, { "epoch": 6.281581264094741, "grad_norm": 4.7862548828125, "learning_rate": 7.769441245918057e-06, "loss": 3.9018, "step": 1041750 }, { "epoch": 6.281882755876075, "grad_norm": 5.842787265777588, "learning_rate": 7.762978444074153e-06, "loss": 3.9939, "step": 1041800 }, { "epoch": 6.282184247657409, "grad_norm": 4.17337703704834, "learning_rate": 7.756518259899663e-06, "loss": 4.2863, "step": 1041850 }, { "epoch": 6.2824857394387426, "grad_norm": 5.82761812210083, "learning_rate": 7.750060693513477e-06, "loss": 4.3688, "step": 1041900 }, { "epoch": 6.282787231220077, "grad_norm": 4.117506980895996, "learning_rate": 7.743605745034414e-06, "loss": 4.0568, "step": 1041950 }, { "epoch": 6.283088723001411, "grad_norm": 4.961038589477539, "learning_rate": 7.737153414581265e-06, "loss": 4.4802, "step": 1042000 }, { "epoch": 6.283390214782745, "grad_norm": 6.9962944984436035, "learning_rate": 7.730703702272816e-06, "loss": 4.3037, "step": 1042050 }, { "epoch": 6.283691706564079, "grad_norm": 4.524539470672607, "learning_rate": 7.724256608227708e-06, "loss": 4.4178, "step": 1042100 }, { "epoch": 6.283993198345413, "grad_norm": 5.753998756408691, "learning_rate": 7.71781213256461e-06, "loss": 4.4039, "step": 1042150 }, { "epoch": 6.284294690126747, "grad_norm": 4.476604461669922, "learning_rate": 7.711370275402163e-06, "loss": 4.0341, "step": 1042200 }, { "epoch": 6.284596181908081, "grad_norm": 5.403101921081543, "learning_rate": 7.704931036858869e-06, "loss": 4.1288, "step": 1042250 }, { "epoch": 6.284897673689415, "grad_norm": 7.690392017364502, "learning_rate": 7.69849441705327e-06, "loss": 4.3045, "step": 1042300 }, { "epoch": 6.2851991654707495, "grad_norm": 4.161014080047607, "learning_rate": 7.692060416103801e-06, "loss": 4.2928, "step": 1042350 }, { "epoch": 6.285500657252083, "grad_norm": 5.946112155914307, "learning_rate": 7.68562903412887e-06, "loss": 4.4025, "step": 1042400 }, { "epoch": 6.285802149033417, "grad_norm": 6.719057083129883, "learning_rate": 7.679200271246832e-06, "loss": 4.1181, "step": 1042450 }, { "epoch": 6.286103640814751, "grad_norm": 5.821659088134766, "learning_rate": 7.672774127576026e-06, "loss": 4.4223, "step": 1042500 }, { "epoch": 6.2864051325960855, "grad_norm": 6.5150651931762695, "learning_rate": 7.66635060323469e-06, "loss": 4.4722, "step": 1042550 }, { "epoch": 6.28670662437742, "grad_norm": 4.930360317230225, "learning_rate": 7.659929698341043e-06, "loss": 4.2476, "step": 1042600 }, { "epoch": 6.287008116158754, "grad_norm": 4.60756778717041, "learning_rate": 7.653511413013281e-06, "loss": 3.9995, "step": 1042650 }, { "epoch": 6.287309607940087, "grad_norm": 5.0558180809021, "learning_rate": 7.647095747369487e-06, "loss": 4.1746, "step": 1042700 }, { "epoch": 6.287611099721421, "grad_norm": 3.147412061691284, "learning_rate": 7.640682701527768e-06, "loss": 4.1131, "step": 1042750 }, { "epoch": 6.287912591502756, "grad_norm": 4.374709129333496, "learning_rate": 7.63427227560613e-06, "loss": 4.419, "step": 1042800 }, { "epoch": 6.28821408328409, "grad_norm": 8.513162612915039, "learning_rate": 7.6278644697225115e-06, "loss": 4.2984, "step": 1042850 }, { "epoch": 6.288515575065424, "grad_norm": 5.84855842590332, "learning_rate": 7.6214592839948686e-06, "loss": 4.055, "step": 1042900 }, { "epoch": 6.288817066846757, "grad_norm": 5.5409064292907715, "learning_rate": 7.615056718541124e-06, "loss": 4.2764, "step": 1042950 }, { "epoch": 6.2891185586280915, "grad_norm": 3.851337432861328, "learning_rate": 7.608656773479033e-06, "loss": 4.2951, "step": 1043000 }, { "epoch": 6.289420050409426, "grad_norm": 6.864336013793945, "learning_rate": 7.602259448926401e-06, "loss": 4.4031, "step": 1043050 }, { "epoch": 6.28972154219076, "grad_norm": 5.454432487487793, "learning_rate": 7.595864745001001e-06, "loss": 4.3248, "step": 1043100 }, { "epoch": 6.290023033972094, "grad_norm": 4.577899932861328, "learning_rate": 7.589472661820489e-06, "loss": 4.2064, "step": 1043150 }, { "epoch": 6.290324525753428, "grad_norm": 4.279107570648193, "learning_rate": 7.583083199502488e-06, "loss": 3.9651, "step": 1043200 }, { "epoch": 6.290626017534762, "grad_norm": 5.7907538414001465, "learning_rate": 7.576696358164619e-06, "loss": 4.335, "step": 1043250 }, { "epoch": 6.290927509316096, "grad_norm": 2.601686954498291, "learning_rate": 7.570312137924373e-06, "loss": 4.3286, "step": 1043300 }, { "epoch": 6.29122900109743, "grad_norm": 3.2127184867858887, "learning_rate": 7.563930538899287e-06, "loss": 4.163, "step": 1043350 }, { "epoch": 6.291530492878764, "grad_norm": 5.463676929473877, "learning_rate": 7.557551561206787e-06, "loss": 4.1292, "step": 1043400 }, { "epoch": 6.2918319846600985, "grad_norm": 4.5982346534729, "learning_rate": 7.551175204964294e-06, "loss": 4.3208, "step": 1043450 }, { "epoch": 6.292133476441432, "grad_norm": 5.3979902267456055, "learning_rate": 7.54480147028913e-06, "loss": 4.3638, "step": 1043500 }, { "epoch": 6.292434968222766, "grad_norm": 5.526094436645508, "learning_rate": 7.538430357298586e-06, "loss": 4.2517, "step": 1043550 }, { "epoch": 6.2927364600041, "grad_norm": 7.0713725090026855, "learning_rate": 7.532061866109951e-06, "loss": 4.1739, "step": 1043600 }, { "epoch": 6.2930379517854345, "grad_norm": 7.391059398651123, "learning_rate": 7.5256959968403804e-06, "loss": 4.3948, "step": 1043650 }, { "epoch": 6.293339443566769, "grad_norm": 7.602294445037842, "learning_rate": 7.5193327496070484e-06, "loss": 3.8687, "step": 1043700 }, { "epoch": 6.293640935348103, "grad_norm": 7.842248916625977, "learning_rate": 7.512972124527095e-06, "loss": 4.2164, "step": 1043750 }, { "epoch": 6.293942427129436, "grad_norm": 1.6290501356124878, "learning_rate": 7.506614121717524e-06, "loss": 4.04, "step": 1043800 }, { "epoch": 6.29424391891077, "grad_norm": 6.925841331481934, "learning_rate": 7.500258741295378e-06, "loss": 4.3226, "step": 1043850 }, { "epoch": 6.294545410692105, "grad_norm": 3.0774030685424805, "learning_rate": 7.493905983377629e-06, "loss": 4.2043, "step": 1043900 }, { "epoch": 6.294846902473439, "grad_norm": 1.8829371929168701, "learning_rate": 7.487555848081167e-06, "loss": 4.3725, "step": 1043950 }, { "epoch": 6.295148394254773, "grad_norm": 4.642388820648193, "learning_rate": 7.481208335522848e-06, "loss": 3.9474, "step": 1044000 }, { "epoch": 6.295148394254773, "eval_loss": 4.819313049316406, "eval_runtime": 39.2576, "eval_samples_per_second": 13.042, "eval_steps_per_second": 6.521, "eval_tts_loss": 8.130558699118035, "step": 1044000 }, { "epoch": 6.295449886036106, "grad_norm": 6.694522380828857, "learning_rate": 7.474863445819529e-06, "loss": 4.3916, "step": 1044050 }, { "epoch": 6.2957513778174405, "grad_norm": 5.624479293823242, "learning_rate": 7.468521179087933e-06, "loss": 4.3426, "step": 1044100 }, { "epoch": 6.296052869598775, "grad_norm": 5.470904350280762, "learning_rate": 7.4621815354448e-06, "loss": 4.0905, "step": 1044150 }, { "epoch": 6.296354361380109, "grad_norm": 3.801086187362671, "learning_rate": 7.45584451500682e-06, "loss": 4.2599, "step": 1044200 }, { "epoch": 6.296655853161443, "grad_norm": 6.0704240798950195, "learning_rate": 7.449510117890584e-06, "loss": 4.5128, "step": 1044250 }, { "epoch": 6.2969573449427765, "grad_norm": 5.846676826477051, "learning_rate": 7.443178344212697e-06, "loss": 3.911, "step": 1044300 }, { "epoch": 6.297258836724111, "grad_norm": 2.872262477874756, "learning_rate": 7.436849194089684e-06, "loss": 4.2438, "step": 1044350 }, { "epoch": 6.297560328505445, "grad_norm": 5.1705427169799805, "learning_rate": 7.430522667637984e-06, "loss": 4.0487, "step": 1044400 }, { "epoch": 6.297861820286779, "grad_norm": 2.547452688217163, "learning_rate": 7.424198764974054e-06, "loss": 3.7047, "step": 1044450 }, { "epoch": 6.298163312068113, "grad_norm": 5.5312371253967285, "learning_rate": 7.417877486214318e-06, "loss": 4.359, "step": 1044500 }, { "epoch": 6.298464803849447, "grad_norm": 7.668111324310303, "learning_rate": 7.4115588314750345e-06, "loss": 4.3272, "step": 1044550 }, { "epoch": 6.298766295630781, "grad_norm": 4.491943836212158, "learning_rate": 7.405242800872524e-06, "loss": 4.3967, "step": 1044600 }, { "epoch": 6.299067787412115, "grad_norm": 6.398695468902588, "learning_rate": 7.3989293945230455e-06, "loss": 4.2032, "step": 1044650 }, { "epoch": 6.299369279193449, "grad_norm": 4.168522834777832, "learning_rate": 7.392618612542789e-06, "loss": 4.1752, "step": 1044700 }, { "epoch": 6.2996707709747835, "grad_norm": 4.903036117553711, "learning_rate": 7.386310455047828e-06, "loss": 3.9484, "step": 1044750 }, { "epoch": 6.299972262756118, "grad_norm": 5.264554500579834, "learning_rate": 7.3800049221543365e-06, "loss": 4.4678, "step": 1044800 }, { "epoch": 6.300273754537451, "grad_norm": 8.031550407409668, "learning_rate": 7.373702013978305e-06, "loss": 4.1636, "step": 1044850 }, { "epoch": 6.300575246318785, "grad_norm": 4.1407999992370605, "learning_rate": 7.367401730635758e-06, "loss": 4.0683, "step": 1044900 }, { "epoch": 6.300876738100119, "grad_norm": 5.140815734863281, "learning_rate": 7.361104072242635e-06, "loss": 4.6554, "step": 1044950 }, { "epoch": 6.301178229881454, "grad_norm": 5.683574199676514, "learning_rate": 7.354809038914827e-06, "loss": 4.2126, "step": 1045000 }, { "epoch": 6.301479721662788, "grad_norm": 4.867024898529053, "learning_rate": 7.3485166307681915e-06, "loss": 4.0502, "step": 1045050 }, { "epoch": 6.301781213444121, "grad_norm": 5.6213788986206055, "learning_rate": 7.342226847918553e-06, "loss": 4.3809, "step": 1045100 }, { "epoch": 6.302082705225455, "grad_norm": 4.2437238693237305, "learning_rate": 7.335939690481651e-06, "loss": 4.4298, "step": 1045150 }, { "epoch": 6.3023841970067895, "grad_norm": 6.725069522857666, "learning_rate": 7.3296551585731614e-06, "loss": 3.7928, "step": 1045200 }, { "epoch": 6.302685688788124, "grad_norm": 7.365272045135498, "learning_rate": 7.323373252308789e-06, "loss": 4.3402, "step": 1045250 }, { "epoch": 6.302987180569458, "grad_norm": 3.106804609298706, "learning_rate": 7.317093971804094e-06, "loss": 4.337, "step": 1045300 }, { "epoch": 6.303288672350792, "grad_norm": 5.66282320022583, "learning_rate": 7.310817317174666e-06, "loss": 4.5552, "step": 1045350 }, { "epoch": 6.3035901641321255, "grad_norm": 3.8691458702087402, "learning_rate": 7.304543288536047e-06, "loss": 4.4102, "step": 1045400 }, { "epoch": 6.30389165591346, "grad_norm": 6.008033752441406, "learning_rate": 7.298271886003643e-06, "loss": 4.228, "step": 1045450 }, { "epoch": 6.304193147694794, "grad_norm": 1.4978095293045044, "learning_rate": 7.292003109692912e-06, "loss": 3.9858, "step": 1045500 }, { "epoch": 6.304494639476128, "grad_norm": 5.845057487487793, "learning_rate": 7.2857369597192295e-06, "loss": 3.9062, "step": 1045550 }, { "epoch": 6.304796131257462, "grad_norm": 7.894407272338867, "learning_rate": 7.279473436197853e-06, "loss": 4.1666, "step": 1045600 }, { "epoch": 6.305097623038796, "grad_norm": 3.658480167388916, "learning_rate": 7.2732125392441055e-06, "loss": 4.0499, "step": 1045650 }, { "epoch": 6.30539911482013, "grad_norm": 4.030773639678955, "learning_rate": 7.266954268973213e-06, "loss": 4.7508, "step": 1045700 }, { "epoch": 6.305700606601464, "grad_norm": 7.809083461761475, "learning_rate": 7.260698625500333e-06, "loss": 4.2653, "step": 1045750 }, { "epoch": 6.306002098382798, "grad_norm": 5.979448318481445, "learning_rate": 7.254445608940573e-06, "loss": 4.4905, "step": 1045800 }, { "epoch": 6.3063035901641324, "grad_norm": 5.430985450744629, "learning_rate": 7.2481952194090585e-06, "loss": 4.4507, "step": 1045850 }, { "epoch": 6.306605081945466, "grad_norm": 6.039052486419678, "learning_rate": 7.241947457020797e-06, "loss": 4.2224, "step": 1045900 }, { "epoch": 6.3069065737268, "grad_norm": 5.869096755981445, "learning_rate": 7.235702321890746e-06, "loss": 4.4396, "step": 1045950 }, { "epoch": 6.307208065508134, "grad_norm": 5.675436496734619, "learning_rate": 7.229459814133881e-06, "loss": 4.1912, "step": 1046000 }, { "epoch": 6.307509557289468, "grad_norm": 8.091249465942383, "learning_rate": 7.223219933865043e-06, "loss": 4.2727, "step": 1046050 }, { "epoch": 6.307811049070803, "grad_norm": 4.159482955932617, "learning_rate": 7.216982681199091e-06, "loss": 4.4801, "step": 1046100 }, { "epoch": 6.308112540852136, "grad_norm": 4.984770774841309, "learning_rate": 7.210748056250815e-06, "loss": 4.1646, "step": 1046150 }, { "epoch": 6.30841403263347, "grad_norm": 4.24045467376709, "learning_rate": 7.20451605913494e-06, "loss": 3.7726, "step": 1046200 }, { "epoch": 6.308715524414804, "grad_norm": 3.6913156509399414, "learning_rate": 7.198286689966176e-06, "loss": 4.2436, "step": 1046250 }, { "epoch": 6.3090170161961385, "grad_norm": 6.5933732986450195, "learning_rate": 7.192059948859163e-06, "loss": 4.7285, "step": 1046300 }, { "epoch": 6.309318507977473, "grad_norm": 5.379948139190674, "learning_rate": 7.185835835928494e-06, "loss": 4.0692, "step": 1046350 }, { "epoch": 6.309619999758807, "grad_norm": 1.9409079551696777, "learning_rate": 7.1796143512886925e-06, "loss": 3.9073, "step": 1046400 }, { "epoch": 6.30992149154014, "grad_norm": 5.35869026184082, "learning_rate": 7.173395495054285e-06, "loss": 3.9694, "step": 1046450 }, { "epoch": 6.3102229833214745, "grad_norm": 1.7541236877441406, "learning_rate": 7.167179267339696e-06, "loss": 4.2832, "step": 1046500 }, { "epoch": 6.310524475102809, "grad_norm": 6.189035415649414, "learning_rate": 7.160965668259333e-06, "loss": 3.9077, "step": 1046550 }, { "epoch": 6.310825966884143, "grad_norm": 2.785322666168213, "learning_rate": 7.154754697927573e-06, "loss": 4.2004, "step": 1046600 }, { "epoch": 6.311127458665477, "grad_norm": 4.854398727416992, "learning_rate": 7.14854635645869e-06, "loss": 4.058, "step": 1046650 }, { "epoch": 6.31142895044681, "grad_norm": 6.5083417892456055, "learning_rate": 7.142340643966976e-06, "loss": 4.5502, "step": 1046700 }, { "epoch": 6.311730442228145, "grad_norm": 4.789698600769043, "learning_rate": 7.1361375605665905e-06, "loss": 4.2545, "step": 1046750 }, { "epoch": 6.312031934009479, "grad_norm": 2.9367659091949463, "learning_rate": 7.1299371063717245e-06, "loss": 4.0816, "step": 1046800 }, { "epoch": 6.312333425790813, "grad_norm": 4.729342460632324, "learning_rate": 7.123739281496471e-06, "loss": 4.4337, "step": 1046850 }, { "epoch": 6.312634917572147, "grad_norm": 3.1284728050231934, "learning_rate": 7.117544086054888e-06, "loss": 3.9284, "step": 1046900 }, { "epoch": 6.312936409353481, "grad_norm": 5.053402423858643, "learning_rate": 7.111351520161035e-06, "loss": 4.3755, "step": 1046950 }, { "epoch": 6.313237901134815, "grad_norm": 3.597607374191284, "learning_rate": 7.10516158392882e-06, "loss": 4.2569, "step": 1047000 }, { "epoch": 6.313237901134815, "eval_loss": 4.8182220458984375, "eval_runtime": 39.1028, "eval_samples_per_second": 13.094, "eval_steps_per_second": 6.547, "eval_tts_loss": 8.147772127282668, "step": 1047000 }, { "epoch": 6.313539392916149, "grad_norm": 7.062894344329834, "learning_rate": 7.098974277472202e-06, "loss": 4.6194, "step": 1047050 }, { "epoch": 6.313840884697483, "grad_norm": 4.2493896484375, "learning_rate": 7.092789600905025e-06, "loss": 4.2534, "step": 1047100 }, { "epoch": 6.314142376478817, "grad_norm": 2.1070263385772705, "learning_rate": 7.086607554341111e-06, "loss": 4.1506, "step": 1047150 }, { "epoch": 6.314443868260152, "grad_norm": 6.010256290435791, "learning_rate": 7.080428137894223e-06, "loss": 4.2324, "step": 1047200 }, { "epoch": 6.314745360041485, "grad_norm": 4.145165920257568, "learning_rate": 7.074251351678134e-06, "loss": 4.4135, "step": 1047250 }, { "epoch": 6.315046851822819, "grad_norm": 1.1095826625823975, "learning_rate": 7.068077195806454e-06, "loss": 3.9979, "step": 1047300 }, { "epoch": 6.315348343604153, "grad_norm": 8.272675514221191, "learning_rate": 7.061905670392826e-06, "loss": 4.2387, "step": 1047350 }, { "epoch": 6.3156498353854875, "grad_norm": 4.150600433349609, "learning_rate": 7.055736775550874e-06, "loss": 4.6952, "step": 1047400 }, { "epoch": 6.315951327166822, "grad_norm": 4.585010051727295, "learning_rate": 7.049570511394076e-06, "loss": 3.9536, "step": 1047450 }, { "epoch": 6.316252818948156, "grad_norm": 5.100522994995117, "learning_rate": 7.0434068780359395e-06, "loss": 4.5879, "step": 1047500 }, { "epoch": 6.316554310729489, "grad_norm": 6.019232749938965, "learning_rate": 7.037245875589908e-06, "loss": 4.4266, "step": 1047550 }, { "epoch": 6.3168558025108235, "grad_norm": 3.109654426574707, "learning_rate": 7.031087504169308e-06, "loss": 4.5391, "step": 1047600 }, { "epoch": 6.317157294292158, "grad_norm": 5.388368129730225, "learning_rate": 7.024931763887532e-06, "loss": 4.0551, "step": 1047650 }, { "epoch": 6.317458786073492, "grad_norm": 6.675605773925781, "learning_rate": 7.018778654857854e-06, "loss": 4.2765, "step": 1047700 }, { "epoch": 6.317760277854826, "grad_norm": 5.445741653442383, "learning_rate": 7.012628177193503e-06, "loss": 4.3396, "step": 1047750 }, { "epoch": 6.318061769636159, "grad_norm": 5.735639572143555, "learning_rate": 7.006480331007669e-06, "loss": 4.2377, "step": 1047800 }, { "epoch": 6.318363261417494, "grad_norm": 1.5404293537139893, "learning_rate": 7.00033511641353e-06, "loss": 4.2662, "step": 1047850 }, { "epoch": 6.318664753198828, "grad_norm": 2.060291290283203, "learning_rate": 6.9941925335241455e-06, "loss": 3.9806, "step": 1047900 }, { "epoch": 6.318966244980162, "grad_norm": 3.9214091300964355, "learning_rate": 6.988052582452541e-06, "loss": 4.1603, "step": 1047950 }, { "epoch": 6.319267736761496, "grad_norm": 10.392807006835938, "learning_rate": 6.981915263311777e-06, "loss": 3.5767, "step": 1048000 }, { "epoch": 6.3195692285428295, "grad_norm": 4.9414286613464355, "learning_rate": 6.975780576214729e-06, "loss": 4.3461, "step": 1048050 }, { "epoch": 6.319870720324164, "grad_norm": 5.958329200744629, "learning_rate": 6.96964852127434e-06, "loss": 4.1648, "step": 1048100 }, { "epoch": 6.320172212105498, "grad_norm": 4.041356086730957, "learning_rate": 6.96351909860347e-06, "loss": 4.5066, "step": 1048150 }, { "epoch": 6.320473703886832, "grad_norm": 6.978280067443848, "learning_rate": 6.9573923083148965e-06, "loss": 4.2501, "step": 1048200 }, { "epoch": 6.320775195668166, "grad_norm": 5.372739315032959, "learning_rate": 6.951268150521394e-06, "loss": 4.1969, "step": 1048250 }, { "epoch": 6.3210766874495, "grad_norm": 4.7686004638671875, "learning_rate": 6.94514662533564e-06, "loss": 4.3513, "step": 1048300 }, { "epoch": 6.321378179230834, "grad_norm": 3.5078365802764893, "learning_rate": 6.939027732870344e-06, "loss": 3.9053, "step": 1048350 }, { "epoch": 6.321679671012168, "grad_norm": 6.266551971435547, "learning_rate": 6.932911473238051e-06, "loss": 4.64, "step": 1048400 }, { "epoch": 6.321981162793502, "grad_norm": 4.17225980758667, "learning_rate": 6.926797846551385e-06, "loss": 4.3464, "step": 1048450 }, { "epoch": 6.3222826545748365, "grad_norm": 6.153175354003906, "learning_rate": 6.9206868529228075e-06, "loss": 4.5353, "step": 1048500 }, { "epoch": 6.322584146356171, "grad_norm": 3.5343878269195557, "learning_rate": 6.914578492464795e-06, "loss": 4.4234, "step": 1048550 }, { "epoch": 6.322885638137504, "grad_norm": 3.900467872619629, "learning_rate": 6.9084727652897905e-06, "loss": 4.3135, "step": 1048600 }, { "epoch": 6.323187129918838, "grad_norm": 5.582424163818359, "learning_rate": 6.902369671510122e-06, "loss": 4.0137, "step": 1048650 }, { "epoch": 6.3234886217001725, "grad_norm": 7.629096031188965, "learning_rate": 6.896269211238164e-06, "loss": 4.2763, "step": 1048700 }, { "epoch": 6.323790113481507, "grad_norm": 4.19211483001709, "learning_rate": 6.890171384586129e-06, "loss": 4.2255, "step": 1048750 }, { "epoch": 6.324091605262841, "grad_norm": 2.93489146232605, "learning_rate": 6.8840761916662605e-06, "loss": 4.3618, "step": 1048800 }, { "epoch": 6.324393097044174, "grad_norm": 5.5649333000183105, "learning_rate": 6.8779836325907165e-06, "loss": 4.1214, "step": 1048850 }, { "epoch": 6.324694588825508, "grad_norm": 3.589451551437378, "learning_rate": 6.871893707471677e-06, "loss": 4.5898, "step": 1048900 }, { "epoch": 6.324996080606843, "grad_norm": 1.3993228673934937, "learning_rate": 6.865806416421132e-06, "loss": 4.2284, "step": 1048950 }, { "epoch": 6.325297572388177, "grad_norm": 7.128439903259277, "learning_rate": 6.859721759551179e-06, "loss": 3.9654, "step": 1049000 }, { "epoch": 6.325599064169511, "grad_norm": 2.2377192974090576, "learning_rate": 6.853639736973793e-06, "loss": 4.0712, "step": 1049050 }, { "epoch": 6.325900555950845, "grad_norm": 4.1743059158325195, "learning_rate": 6.847560348800868e-06, "loss": 4.1797, "step": 1049100 }, { "epoch": 6.3262020477321785, "grad_norm": 6.086449146270752, "learning_rate": 6.8414835951443145e-06, "loss": 4.3421, "step": 1049150 }, { "epoch": 6.326503539513513, "grad_norm": 1.1861517429351807, "learning_rate": 6.83540947611596e-06, "loss": 4.0134, "step": 1049200 }, { "epoch": 6.326805031294847, "grad_norm": 4.320295810699463, "learning_rate": 6.829337991827566e-06, "loss": 4.4008, "step": 1049250 }, { "epoch": 6.327106523076181, "grad_norm": 5.399519920349121, "learning_rate": 6.823269142390892e-06, "loss": 4.2208, "step": 1049300 }, { "epoch": 6.327408014857515, "grad_norm": 4.404662132263184, "learning_rate": 6.817202927917647e-06, "loss": 4.1919, "step": 1049350 }, { "epoch": 6.327709506638849, "grad_norm": 5.703934192657471, "learning_rate": 6.8111393485194285e-06, "loss": 3.9961, "step": 1049400 }, { "epoch": 6.328010998420183, "grad_norm": 4.554591178894043, "learning_rate": 6.805078404307845e-06, "loss": 4.6897, "step": 1049450 }, { "epoch": 6.328312490201517, "grad_norm": 4.505899429321289, "learning_rate": 6.799020095394475e-06, "loss": 4.5474, "step": 1049500 }, { "epoch": 6.328613981982851, "grad_norm": 4.302942752838135, "learning_rate": 6.792964421890762e-06, "loss": 3.9137, "step": 1049550 }, { "epoch": 6.3289154737641855, "grad_norm": 4.418415546417236, "learning_rate": 6.786911383908167e-06, "loss": 4.2583, "step": 1049600 }, { "epoch": 6.329216965545519, "grad_norm": 7.198336124420166, "learning_rate": 6.780860981558117e-06, "loss": 3.9605, "step": 1049650 }, { "epoch": 6.329518457326853, "grad_norm": 4.468069553375244, "learning_rate": 6.774813214951907e-06, "loss": 4.2737, "step": 1049700 }, { "epoch": 6.329819949108187, "grad_norm": 5.177638053894043, "learning_rate": 6.7687680842008656e-06, "loss": 4.3332, "step": 1049750 }, { "epoch": 6.3301214408895214, "grad_norm": 1.9870504140853882, "learning_rate": 6.762725589416268e-06, "loss": 4.2073, "step": 1049800 }, { "epoch": 6.330422932670856, "grad_norm": 6.047516345977783, "learning_rate": 6.75668573070926e-06, "loss": 3.98, "step": 1049850 }, { "epoch": 6.330724424452189, "grad_norm": 7.93405818939209, "learning_rate": 6.75064850819107e-06, "loss": 4.3097, "step": 1049900 }, { "epoch": 6.331025916233523, "grad_norm": 4.585689067840576, "learning_rate": 6.744613921972775e-06, "loss": 4.1004, "step": 1049950 }, { "epoch": 6.331327408014857, "grad_norm": 4.683448314666748, "learning_rate": 6.738581972165385e-06, "loss": 4.5774, "step": 1050000 }, { "epoch": 6.331327408014857, "eval_loss": 4.818714141845703, "eval_runtime": 38.987, "eval_samples_per_second": 13.133, "eval_steps_per_second": 6.566, "eval_tts_loss": 8.11889060249392, "step": 1050000 }, { "epoch": 6.331628899796192, "grad_norm": 4.200116157531738, "learning_rate": 6.7325526588799475e-06, "loss": 4.2726, "step": 1050050 }, { "epoch": 6.331930391577526, "grad_norm": 1.7598942518234253, "learning_rate": 6.726525982227437e-06, "loss": 4.3365, "step": 1050100 }, { "epoch": 6.33223188335886, "grad_norm": 3.8313374519348145, "learning_rate": 6.720501942318768e-06, "loss": 4.3089, "step": 1050150 }, { "epoch": 6.332533375140193, "grad_norm": 6.27899169921875, "learning_rate": 6.714480539264783e-06, "loss": 4.2001, "step": 1050200 }, { "epoch": 6.3328348669215275, "grad_norm": 3.8501381874084473, "learning_rate": 6.70846177317631e-06, "loss": 4.5822, "step": 1050250 }, { "epoch": 6.333136358702862, "grad_norm": 5.496181488037109, "learning_rate": 6.702445644164111e-06, "loss": 4.4597, "step": 1050300 }, { "epoch": 6.333437850484196, "grad_norm": 4.361030578613281, "learning_rate": 6.696432152338882e-06, "loss": 4.1903, "step": 1050350 }, { "epoch": 6.33373934226553, "grad_norm": 5.181246280670166, "learning_rate": 6.690421297811316e-06, "loss": 4.4037, "step": 1050400 }, { "epoch": 6.3340408340468635, "grad_norm": 9.567925453186035, "learning_rate": 6.684413080692058e-06, "loss": 4.2719, "step": 1050450 }, { "epoch": 6.334342325828198, "grad_norm": 5.022897720336914, "learning_rate": 6.6784075010916205e-06, "loss": 4.0336, "step": 1050500 }, { "epoch": 6.334643817609532, "grad_norm": 5.799408912658691, "learning_rate": 6.672404559120581e-06, "loss": 4.3803, "step": 1050550 }, { "epoch": 6.334945309390866, "grad_norm": 4.98936128616333, "learning_rate": 6.666404254889401e-06, "loss": 4.4623, "step": 1050600 }, { "epoch": 6.3352468011722, "grad_norm": 6.546629905700684, "learning_rate": 6.660406588508494e-06, "loss": 4.4379, "step": 1050650 }, { "epoch": 6.3355482929535345, "grad_norm": 4.838316440582275, "learning_rate": 6.654411560088252e-06, "loss": 4.0603, "step": 1050700 }, { "epoch": 6.335849784734868, "grad_norm": 5.660417556762695, "learning_rate": 6.648419169739022e-06, "loss": 4.2036, "step": 1050750 }, { "epoch": 6.336151276516202, "grad_norm": 5.870877265930176, "learning_rate": 6.642429417571033e-06, "loss": 4.4295, "step": 1050800 }, { "epoch": 6.336452768297536, "grad_norm": 6.012696743011475, "learning_rate": 6.6364423036945445e-06, "loss": 4.0611, "step": 1050850 }, { "epoch": 6.33675426007887, "grad_norm": 1.978733777999878, "learning_rate": 6.6304578282197544e-06, "loss": 4.3288, "step": 1050900 }, { "epoch": 6.337055751860205, "grad_norm": 6.401663303375244, "learning_rate": 6.624475991256789e-06, "loss": 4.243, "step": 1050950 }, { "epoch": 6.337357243641538, "grad_norm": 3.918480396270752, "learning_rate": 6.618496792915729e-06, "loss": 4.4819, "step": 1051000 }, { "epoch": 6.337658735422872, "grad_norm": 4.187556743621826, "learning_rate": 6.612520233306634e-06, "loss": 4.0281, "step": 1051050 }, { "epoch": 6.337960227204206, "grad_norm": 6.319941997528076, "learning_rate": 6.606546312539484e-06, "loss": 4.2711, "step": 1051100 }, { "epoch": 6.338261718985541, "grad_norm": 4.09271240234375, "learning_rate": 6.600575030724192e-06, "loss": 3.9671, "step": 1051150 }, { "epoch": 6.338563210766875, "grad_norm": 1.5887994766235352, "learning_rate": 6.594606387970702e-06, "loss": 4.2097, "step": 1051200 }, { "epoch": 6.338864702548209, "grad_norm": 6.602896690368652, "learning_rate": 6.58864038438881e-06, "loss": 4.0816, "step": 1051250 }, { "epoch": 6.339166194329542, "grad_norm": 4.258200645446777, "learning_rate": 6.582677020088328e-06, "loss": 4.1348, "step": 1051300 }, { "epoch": 6.3394676861108765, "grad_norm": 6.136647701263428, "learning_rate": 6.576716295179018e-06, "loss": 4.2641, "step": 1051350 }, { "epoch": 6.339769177892211, "grad_norm": 3.6463613510131836, "learning_rate": 6.5707582097705604e-06, "loss": 4.1753, "step": 1051400 }, { "epoch": 6.340070669673545, "grad_norm": 4.747345924377441, "learning_rate": 6.564802763972632e-06, "loss": 3.9248, "step": 1051450 }, { "epoch": 6.340372161454879, "grad_norm": 8.032851219177246, "learning_rate": 6.558849957894813e-06, "loss": 4.2174, "step": 1051500 }, { "epoch": 6.3406736532362125, "grad_norm": 7.297952175140381, "learning_rate": 6.552899791646649e-06, "loss": 4.354, "step": 1051550 }, { "epoch": 6.340975145017547, "grad_norm": 4.275542259216309, "learning_rate": 6.546952265337652e-06, "loss": 4.0815, "step": 1051600 }, { "epoch": 6.341276636798881, "grad_norm": 5.722171783447266, "learning_rate": 6.5410073790772845e-06, "loss": 4.2799, "step": 1051650 }, { "epoch": 6.341578128580215, "grad_norm": 4.649531841278076, "learning_rate": 6.535065132974943e-06, "loss": 4.2189, "step": 1051700 }, { "epoch": 6.341879620361549, "grad_norm": 5.336044788360596, "learning_rate": 6.529125527139989e-06, "loss": 4.3322, "step": 1051750 }, { "epoch": 6.342181112142883, "grad_norm": 4.846982955932617, "learning_rate": 6.523188561681752e-06, "loss": 4.0414, "step": 1051800 }, { "epoch": 6.342482603924217, "grad_norm": 6.075271129608154, "learning_rate": 6.517254236709463e-06, "loss": 3.8555, "step": 1051850 }, { "epoch": 6.342784095705551, "grad_norm": 4.087514400482178, "learning_rate": 6.511322552332365e-06, "loss": 4.0085, "step": 1051900 }, { "epoch": 6.343085587486885, "grad_norm": 6.808719158172607, "learning_rate": 6.505393508659607e-06, "loss": 4.5293, "step": 1051950 }, { "epoch": 6.343387079268219, "grad_norm": 8.924551963806152, "learning_rate": 6.4994671058002824e-06, "loss": 4.3414, "step": 1052000 }, { "epoch": 6.343688571049553, "grad_norm": 4.815306186676025, "learning_rate": 6.493543343863489e-06, "loss": 4.2485, "step": 1052050 }, { "epoch": 6.343990062830887, "grad_norm": 5.548521518707275, "learning_rate": 6.487622222958255e-06, "loss": 4.1592, "step": 1052100 }, { "epoch": 6.344291554612221, "grad_norm": 6.172640800476074, "learning_rate": 6.481703743193495e-06, "loss": 4.3313, "step": 1052150 }, { "epoch": 6.344593046393555, "grad_norm": 7.359450817108154, "learning_rate": 6.47578790467817e-06, "loss": 3.9774, "step": 1052200 }, { "epoch": 6.34489453817489, "grad_norm": 3.3442580699920654, "learning_rate": 6.469874707521178e-06, "loss": 4.2964, "step": 1052250 }, { "epoch": 6.345196029956224, "grad_norm": 5.691324710845947, "learning_rate": 6.463964151831313e-06, "loss": 4.5189, "step": 1052300 }, { "epoch": 6.345497521737557, "grad_norm": 6.207416534423828, "learning_rate": 6.458056237717324e-06, "loss": 4.4195, "step": 1052350 }, { "epoch": 6.345799013518891, "grad_norm": 4.714034557342529, "learning_rate": 6.452150965287989e-06, "loss": 3.9682, "step": 1052400 }, { "epoch": 6.3461005053002255, "grad_norm": 5.136186599731445, "learning_rate": 6.446248334651938e-06, "loss": 4.5216, "step": 1052450 }, { "epoch": 6.34640199708156, "grad_norm": 1.7907476425170898, "learning_rate": 6.440348345917834e-06, "loss": 3.9285, "step": 1052500 }, { "epoch": 6.346703488862894, "grad_norm": 4.180742263793945, "learning_rate": 6.4344509991942575e-06, "loss": 4.1695, "step": 1052550 }, { "epoch": 6.347004980644227, "grad_norm": 2.274880886077881, "learning_rate": 6.42855629458972e-06, "loss": 4.0531, "step": 1052600 }, { "epoch": 6.3473064724255615, "grad_norm": 4.636981964111328, "learning_rate": 6.422664232212721e-06, "loss": 4.2094, "step": 1052650 }, { "epoch": 6.347607964206896, "grad_norm": 7.5634765625, "learning_rate": 6.416774812171704e-06, "loss": 4.1283, "step": 1052700 }, { "epoch": 6.34790945598823, "grad_norm": 3.6750078201293945, "learning_rate": 6.410888034575018e-06, "loss": 4.3222, "step": 1052750 }, { "epoch": 6.348210947769564, "grad_norm": 7.077736854553223, "learning_rate": 6.405003899531025e-06, "loss": 3.9405, "step": 1052800 }, { "epoch": 6.348512439550898, "grad_norm": 4.047644138336182, "learning_rate": 6.399122407148038e-06, "loss": 3.7325, "step": 1052850 }, { "epoch": 6.348813931332232, "grad_norm": 3.2954788208007812, "learning_rate": 6.393243557534255e-06, "loss": 4.5045, "step": 1052900 }, { "epoch": 6.349115423113566, "grad_norm": 5.196213245391846, "learning_rate": 6.387367350797873e-06, "loss": 4.3094, "step": 1052950 }, { "epoch": 6.3494169148949, "grad_norm": 7.178204536437988, "learning_rate": 6.381493787047088e-06, "loss": 4.269, "step": 1053000 }, { "epoch": 6.3494169148949, "eval_loss": 4.816197395324707, "eval_runtime": 39.2207, "eval_samples_per_second": 13.054, "eval_steps_per_second": 6.527, "eval_tts_loss": 8.137965890052442, "step": 1053000 }, { "epoch": 6.349718406676234, "grad_norm": 4.129894256591797, "learning_rate": 6.375622866389929e-06, "loss": 4.3388, "step": 1053050 }, { "epoch": 6.350019898457568, "grad_norm": 6.393435955047607, "learning_rate": 6.3697545889344785e-06, "loss": 3.9498, "step": 1053100 }, { "epoch": 6.350321390238902, "grad_norm": 4.980090618133545, "learning_rate": 6.363888954788731e-06, "loss": 4.2134, "step": 1053150 }, { "epoch": 6.350622882020236, "grad_norm": 7.18690299987793, "learning_rate": 6.358025964060603e-06, "loss": 4.5019, "step": 1053200 }, { "epoch": 6.35092437380157, "grad_norm": 3.7866365909576416, "learning_rate": 6.352165616858024e-06, "loss": 4.6335, "step": 1053250 }, { "epoch": 6.351225865582904, "grad_norm": 7.2568793296813965, "learning_rate": 6.346307913288856e-06, "loss": 4.3457, "step": 1053300 }, { "epoch": 6.351527357364239, "grad_norm": 4.470643997192383, "learning_rate": 6.340452853460864e-06, "loss": 4.2215, "step": 1053350 }, { "epoch": 6.351828849145572, "grad_norm": 5.762772560119629, "learning_rate": 6.33460043748183e-06, "loss": 4.3261, "step": 1053400 }, { "epoch": 6.352130340926906, "grad_norm": 4.40238618850708, "learning_rate": 6.328750665459481e-06, "loss": 4.1739, "step": 1053450 }, { "epoch": 6.35243183270824, "grad_norm": 3.6916561126708984, "learning_rate": 6.322903537501434e-06, "loss": 4.0879, "step": 1053500 }, { "epoch": 6.3527333244895745, "grad_norm": 4.013850212097168, "learning_rate": 6.317059053715284e-06, "loss": 4.3223, "step": 1053550 }, { "epoch": 6.353034816270909, "grad_norm": 7.4458394050598145, "learning_rate": 6.31121721420863e-06, "loss": 4.2641, "step": 1053600 }, { "epoch": 6.353336308052242, "grad_norm": 6.154860019683838, "learning_rate": 6.305378019088986e-06, "loss": 3.955, "step": 1053650 }, { "epoch": 6.353637799833576, "grad_norm": 4.353725433349609, "learning_rate": 6.299541468463765e-06, "loss": 4.2449, "step": 1053700 }, { "epoch": 6.3539392916149104, "grad_norm": 2.429810047149658, "learning_rate": 6.2937075624404305e-06, "loss": 4.381, "step": 1053750 }, { "epoch": 6.354240783396245, "grad_norm": 6.309067726135254, "learning_rate": 6.287876301126332e-06, "loss": 4.2822, "step": 1053800 }, { "epoch": 6.354542275177579, "grad_norm": 10.49724006652832, "learning_rate": 6.282047684628783e-06, "loss": 4.1718, "step": 1053850 }, { "epoch": 6.354843766958913, "grad_norm": 1.2432395219802856, "learning_rate": 6.27622171305503e-06, "loss": 4.4261, "step": 1053900 }, { "epoch": 6.355145258740246, "grad_norm": 4.486776351928711, "learning_rate": 6.270398386512337e-06, "loss": 3.7899, "step": 1053950 }, { "epoch": 6.355446750521581, "grad_norm": 5.492956161499023, "learning_rate": 6.2645777051078375e-06, "loss": 4.1464, "step": 1054000 }, { "epoch": 6.355748242302915, "grad_norm": 1.7909048795700073, "learning_rate": 6.258759668948643e-06, "loss": 4.3304, "step": 1054050 }, { "epoch": 6.356049734084249, "grad_norm": 6.0105814933776855, "learning_rate": 6.252944278141886e-06, "loss": 4.1334, "step": 1054100 }, { "epoch": 6.356351225865583, "grad_norm": 5.305566310882568, "learning_rate": 6.247131532794514e-06, "loss": 4.1178, "step": 1054150 }, { "epoch": 6.3566527176469165, "grad_norm": 5.450228214263916, "learning_rate": 6.241321433013558e-06, "loss": 4.5279, "step": 1054200 }, { "epoch": 6.356954209428251, "grad_norm": 7.64824104309082, "learning_rate": 6.235513978905932e-06, "loss": 4.1232, "step": 1054250 }, { "epoch": 6.357255701209585, "grad_norm": 4.253852367401123, "learning_rate": 6.2297091705785174e-06, "loss": 4.1678, "step": 1054300 }, { "epoch": 6.357557192990919, "grad_norm": 4.220731735229492, "learning_rate": 6.223907008138113e-06, "loss": 4.1555, "step": 1054350 }, { "epoch": 6.357858684772253, "grad_norm": 4.984231472015381, "learning_rate": 6.218107491691532e-06, "loss": 4.517, "step": 1054400 }, { "epoch": 6.358160176553588, "grad_norm": 4.342042922973633, "learning_rate": 6.2123106213454735e-06, "loss": 4.1402, "step": 1054450 }, { "epoch": 6.358461668334921, "grad_norm": 4.539694786071777, "learning_rate": 6.206516397206651e-06, "loss": 4.239, "step": 1054500 }, { "epoch": 6.358763160116255, "grad_norm": 6.76288366317749, "learning_rate": 6.200724819381714e-06, "loss": 3.9189, "step": 1054550 }, { "epoch": 6.359064651897589, "grad_norm": 4.678059101104736, "learning_rate": 6.194935887977192e-06, "loss": 4.2358, "step": 1054600 }, { "epoch": 6.3593661436789235, "grad_norm": 5.2094573974609375, "learning_rate": 6.189149603099686e-06, "loss": 4.1299, "step": 1054650 }, { "epoch": 6.359667635460258, "grad_norm": 4.464444637298584, "learning_rate": 6.183365964855641e-06, "loss": 3.7583, "step": 1054700 }, { "epoch": 6.359969127241591, "grad_norm": 5.344927787780762, "learning_rate": 6.177584973351507e-06, "loss": 4.0692, "step": 1054750 }, { "epoch": 6.360270619022925, "grad_norm": 6.648351669311523, "learning_rate": 6.171806628693649e-06, "loss": 4.0559, "step": 1054800 }, { "epoch": 6.360572110804259, "grad_norm": 4.786155700683594, "learning_rate": 6.16603093098848e-06, "loss": 4.5322, "step": 1054850 }, { "epoch": 6.360873602585594, "grad_norm": 5.486185073852539, "learning_rate": 6.160257880342217e-06, "loss": 4.2684, "step": 1054900 }, { "epoch": 6.361175094366928, "grad_norm": 6.360518455505371, "learning_rate": 6.154487476861142e-06, "loss": 4.413, "step": 1054950 }, { "epoch": 6.361476586148262, "grad_norm": 6.006208896636963, "learning_rate": 6.148719720651468e-06, "loss": 4.2876, "step": 1055000 }, { "epoch": 6.361778077929595, "grad_norm": 4.901707172393799, "learning_rate": 6.142954611819295e-06, "loss": 4.2492, "step": 1055050 }, { "epoch": 6.36207956971093, "grad_norm": 4.4807515144348145, "learning_rate": 6.137192150470771e-06, "loss": 4.4134, "step": 1055100 }, { "epoch": 6.362381061492264, "grad_norm": 5.441586494445801, "learning_rate": 6.1314323367119116e-06, "loss": 4.3599, "step": 1055150 }, { "epoch": 6.362682553273598, "grad_norm": 4.5908613204956055, "learning_rate": 6.125675170648714e-06, "loss": 4.1728, "step": 1055200 }, { "epoch": 6.362984045054932, "grad_norm": 4.52139139175415, "learning_rate": 6.119920652387161e-06, "loss": 4.2734, "step": 1055250 }, { "epoch": 6.3632855368362655, "grad_norm": 4.068215370178223, "learning_rate": 6.114168782033135e-06, "loss": 4.215, "step": 1055300 }, { "epoch": 6.3635870286176, "grad_norm": 4.133562088012695, "learning_rate": 6.108419559692501e-06, "loss": 4.1984, "step": 1055350 }, { "epoch": 6.363888520398934, "grad_norm": 5.514767169952393, "learning_rate": 6.102672985471041e-06, "loss": 4.2805, "step": 1055400 }, { "epoch": 6.364190012180268, "grad_norm": 4.163960933685303, "learning_rate": 6.09692905947457e-06, "loss": 4.134, "step": 1055450 }, { "epoch": 6.364491503961602, "grad_norm": 7.0888776779174805, "learning_rate": 6.091187781808738e-06, "loss": 4.283, "step": 1055500 }, { "epoch": 6.364792995742936, "grad_norm": 4.6653666496276855, "learning_rate": 6.085449152579225e-06, "loss": 4.2242, "step": 1055550 }, { "epoch": 6.36509448752427, "grad_norm": 3.3452794551849365, "learning_rate": 6.079713171891648e-06, "loss": 4.39, "step": 1055600 }, { "epoch": 6.365395979305604, "grad_norm": 6.303167343139648, "learning_rate": 6.073979839851556e-06, "loss": 4.2963, "step": 1055650 }, { "epoch": 6.365697471086938, "grad_norm": 3.535884380340576, "learning_rate": 6.068249156564481e-06, "loss": 4.2032, "step": 1055700 }, { "epoch": 6.3659989628682725, "grad_norm": 3.5868735313415527, "learning_rate": 6.062521122135872e-06, "loss": 4.1413, "step": 1055750 }, { "epoch": 6.366300454649606, "grad_norm": 5.219151020050049, "learning_rate": 6.056795736671144e-06, "loss": 4.2083, "step": 1055800 }, { "epoch": 6.36660194643094, "grad_norm": 5.044733047485352, "learning_rate": 6.051073000275697e-06, "loss": 4.2743, "step": 1055850 }, { "epoch": 6.366903438212274, "grad_norm": 8.855685234069824, "learning_rate": 6.04535291305483e-06, "loss": 4.2079, "step": 1055900 }, { "epoch": 6.367204929993608, "grad_norm": 4.948491096496582, "learning_rate": 6.039635475113774e-06, "loss": 4.396, "step": 1055950 }, { "epoch": 6.367506421774943, "grad_norm": 5.873838424682617, "learning_rate": 6.033920686557797e-06, "loss": 4.3474, "step": 1056000 }, { "epoch": 6.367506421774943, "eval_loss": 4.813442230224609, "eval_runtime": 39.0554, "eval_samples_per_second": 13.11, "eval_steps_per_second": 6.555, "eval_tts_loss": 8.145860680048871, "step": 1056000 }, { "epoch": 6.367807913556277, "grad_norm": 1.336488127708435, "learning_rate": 6.028208547492064e-06, "loss": 4.3141, "step": 1056050 }, { "epoch": 6.36810940533761, "grad_norm": 5.802981853485107, "learning_rate": 6.02249905802169e-06, "loss": 4.5252, "step": 1056100 }, { "epoch": 6.368410897118944, "grad_norm": 7.028835773468018, "learning_rate": 6.0167922182517425e-06, "loss": 4.2123, "step": 1056150 }, { "epoch": 6.368712388900279, "grad_norm": 5.3039326667785645, "learning_rate": 6.011088028287287e-06, "loss": 4.2908, "step": 1056200 }, { "epoch": 6.369013880681613, "grad_norm": 3.9803426265716553, "learning_rate": 6.0053864882332545e-06, "loss": 3.8526, "step": 1056250 }, { "epoch": 6.369315372462947, "grad_norm": 4.858489036560059, "learning_rate": 5.999687598194614e-06, "loss": 4.2235, "step": 1056300 }, { "epoch": 6.36961686424428, "grad_norm": 5.918729305267334, "learning_rate": 5.993991358276212e-06, "loss": 4.3153, "step": 1056350 }, { "epoch": 6.3699183560256145, "grad_norm": 4.70469856262207, "learning_rate": 5.9882977685828835e-06, "loss": 4.028, "step": 1056400 }, { "epoch": 6.370219847806949, "grad_norm": 6.252125263214111, "learning_rate": 5.9826068292194106e-06, "loss": 4.4853, "step": 1056450 }, { "epoch": 6.370521339588283, "grad_norm": 5.690666675567627, "learning_rate": 5.97691854029056e-06, "loss": 4.1976, "step": 1056500 }, { "epoch": 6.370822831369617, "grad_norm": 6.807355880737305, "learning_rate": 5.971232901900963e-06, "loss": 4.482, "step": 1056550 }, { "epoch": 6.371124323150951, "grad_norm": 2.96532940864563, "learning_rate": 5.965549914155288e-06, "loss": 4.2848, "step": 1056600 }, { "epoch": 6.371425814932285, "grad_norm": 4.036548614501953, "learning_rate": 5.95986957715815e-06, "loss": 4.1741, "step": 1056650 }, { "epoch": 6.371727306713619, "grad_norm": 5.061432361602783, "learning_rate": 5.954191891014032e-06, "loss": 4.1323, "step": 1056700 }, { "epoch": 6.372028798494953, "grad_norm": 4.979496479034424, "learning_rate": 5.948516855827451e-06, "loss": 4.2078, "step": 1056750 }, { "epoch": 6.372330290276287, "grad_norm": 2.8007500171661377, "learning_rate": 5.942844471702857e-06, "loss": 4.4083, "step": 1056800 }, { "epoch": 6.3726317820576215, "grad_norm": 6.678362846374512, "learning_rate": 5.937174738744616e-06, "loss": 4.2314, "step": 1056850 }, { "epoch": 6.372933273838955, "grad_norm": 5.538977146148682, "learning_rate": 5.931507657057077e-06, "loss": 4.3338, "step": 1056900 }, { "epoch": 6.373234765620289, "grad_norm": 6.908508777618408, "learning_rate": 5.925843226744542e-06, "loss": 4.3554, "step": 1056950 }, { "epoch": 6.373536257401623, "grad_norm": 6.490169048309326, "learning_rate": 5.920181447911276e-06, "loss": 4.413, "step": 1057000 }, { "epoch": 6.373837749182957, "grad_norm": 4.175704002380371, "learning_rate": 5.914522320661464e-06, "loss": 4.1332, "step": 1057050 }, { "epoch": 6.374139240964292, "grad_norm": 4.077162742614746, "learning_rate": 5.90886584509922e-06, "loss": 4.0099, "step": 1057100 }, { "epoch": 6.374440732745625, "grad_norm": 1.4626895189285278, "learning_rate": 5.903212021328696e-06, "loss": 4.0694, "step": 1057150 }, { "epoch": 6.374742224526959, "grad_norm": 4.057933807373047, "learning_rate": 5.897560849453909e-06, "loss": 4.3479, "step": 1057200 }, { "epoch": 6.375043716308293, "grad_norm": 4.162336349487305, "learning_rate": 5.891912329578858e-06, "loss": 4.1524, "step": 1057250 }, { "epoch": 6.375345208089628, "grad_norm": 5.719404220581055, "learning_rate": 5.8862664618075266e-06, "loss": 4.4819, "step": 1057300 }, { "epoch": 6.375646699870962, "grad_norm": 6.434555530548096, "learning_rate": 5.880623246243782e-06, "loss": 4.3332, "step": 1057350 }, { "epoch": 6.375948191652295, "grad_norm": 4.095320701599121, "learning_rate": 5.874982682991491e-06, "loss": 4.0272, "step": 1057400 }, { "epoch": 6.376249683433629, "grad_norm": 5.4763360023498535, "learning_rate": 5.869344772154489e-06, "loss": 4.3064, "step": 1057450 }, { "epoch": 6.3765511752149635, "grad_norm": 4.338238716125488, "learning_rate": 5.863709513836506e-06, "loss": 4.0297, "step": 1057500 }, { "epoch": 6.376852666996298, "grad_norm": 5.146057605743408, "learning_rate": 5.858076908141246e-06, "loss": 4.3329, "step": 1057550 }, { "epoch": 6.377154158777632, "grad_norm": 5.775318145751953, "learning_rate": 5.852446955172407e-06, "loss": 4.1868, "step": 1057600 }, { "epoch": 6.377455650558966, "grad_norm": 8.255678176879883, "learning_rate": 5.846819655033541e-06, "loss": 4.4183, "step": 1057650 }, { "epoch": 6.3777571423402994, "grad_norm": 3.7704591751098633, "learning_rate": 5.841195007828231e-06, "loss": 4.2025, "step": 1057700 }, { "epoch": 6.378058634121634, "grad_norm": 6.913575649261475, "learning_rate": 5.8355730136600275e-06, "loss": 4.0159, "step": 1057750 }, { "epoch": 6.378360125902968, "grad_norm": 6.161825656890869, "learning_rate": 5.829953672632365e-06, "loss": 4.1852, "step": 1057800 }, { "epoch": 6.378661617684302, "grad_norm": 4.356902122497559, "learning_rate": 5.82433698484866e-06, "loss": 4.1109, "step": 1057850 }, { "epoch": 6.378963109465636, "grad_norm": 6.347421169281006, "learning_rate": 5.818722950412297e-06, "loss": 4.2192, "step": 1057900 }, { "epoch": 6.37926460124697, "grad_norm": 4.742737293243408, "learning_rate": 5.813111569426559e-06, "loss": 4.1199, "step": 1057950 }, { "epoch": 6.379566093028304, "grad_norm": 7.6729021072387695, "learning_rate": 5.807502841994749e-06, "loss": 4.5497, "step": 1058000 }, { "epoch": 6.379867584809638, "grad_norm": 3.981088399887085, "learning_rate": 5.8018967682200825e-06, "loss": 4.3891, "step": 1058050 }, { "epoch": 6.380169076590972, "grad_norm": 4.211678504943848, "learning_rate": 5.796293348205694e-06, "loss": 4.4072, "step": 1058100 }, { "epoch": 6.380470568372306, "grad_norm": 5.384971618652344, "learning_rate": 5.790692582054751e-06, "loss": 3.9285, "step": 1058150 }, { "epoch": 6.380772060153641, "grad_norm": 7.826410293579102, "learning_rate": 5.785094469870321e-06, "loss": 3.912, "step": 1058200 }, { "epoch": 6.381073551934974, "grad_norm": 5.669681072235107, "learning_rate": 5.779499011755439e-06, "loss": 4.2833, "step": 1058250 }, { "epoch": 6.381375043716308, "grad_norm": 4.554218769073486, "learning_rate": 5.773906207813039e-06, "loss": 4.1954, "step": 1058300 }, { "epoch": 6.381676535497642, "grad_norm": 6.120279312133789, "learning_rate": 5.7683160581460885e-06, "loss": 4.0791, "step": 1058350 }, { "epoch": 6.381978027278977, "grad_norm": 8.055153846740723, "learning_rate": 5.762728562857422e-06, "loss": 4.36, "step": 1058400 }, { "epoch": 6.382279519060311, "grad_norm": 6.727406978607178, "learning_rate": 5.757143722049923e-06, "loss": 4.5185, "step": 1058450 }, { "epoch": 6.382581010841644, "grad_norm": 6.2928643226623535, "learning_rate": 5.751561535826343e-06, "loss": 4.3725, "step": 1058500 }, { "epoch": 6.382882502622978, "grad_norm": 6.125033378601074, "learning_rate": 5.745982004289418e-06, "loss": 4.2384, "step": 1058550 }, { "epoch": 6.3831839944043125, "grad_norm": 3.8674325942993164, "learning_rate": 5.740405127541814e-06, "loss": 4.2302, "step": 1058600 }, { "epoch": 6.383485486185647, "grad_norm": 8.86469841003418, "learning_rate": 5.7348309056862164e-06, "loss": 4.1014, "step": 1058650 }, { "epoch": 6.383786977966981, "grad_norm": 5.197315692901611, "learning_rate": 5.729259338825176e-06, "loss": 4.494, "step": 1058700 }, { "epoch": 6.384088469748315, "grad_norm": 5.799222469329834, "learning_rate": 5.723690427061212e-06, "loss": 4.4984, "step": 1058750 }, { "epoch": 6.384389961529648, "grad_norm": 2.6077511310577393, "learning_rate": 5.718124170496857e-06, "loss": 4.3846, "step": 1058800 }, { "epoch": 6.384691453310983, "grad_norm": 5.830441474914551, "learning_rate": 5.7125605692344974e-06, "loss": 3.7219, "step": 1058850 }, { "epoch": 6.384992945092317, "grad_norm": 4.7931952476501465, "learning_rate": 5.706999623376568e-06, "loss": 4.6758, "step": 1058900 }, { "epoch": 6.385294436873651, "grad_norm": 5.904758453369141, "learning_rate": 5.701441333025403e-06, "loss": 4.3719, "step": 1058950 }, { "epoch": 6.385595928654985, "grad_norm": 2.2925665378570557, "learning_rate": 5.695885698283253e-06, "loss": 4.0198, "step": 1059000 }, { "epoch": 6.385595928654985, "eval_loss": 4.808598518371582, "eval_runtime": 39.2371, "eval_samples_per_second": 13.049, "eval_steps_per_second": 6.524, "eval_tts_loss": 8.128494161921022, "step": 1059000 }, { "epoch": 6.385897420436319, "grad_norm": 2.0635616779327393, "learning_rate": 5.690332719252422e-06, "loss": 3.8079, "step": 1059050 }, { "epoch": 6.386198912217653, "grad_norm": 6.921584606170654, "learning_rate": 5.684782396035093e-06, "loss": 4.061, "step": 1059100 }, { "epoch": 6.386500403998987, "grad_norm": 8.854055404663086, "learning_rate": 5.6792347287333505e-06, "loss": 4.2559, "step": 1059150 }, { "epoch": 6.386801895780321, "grad_norm": 4.137092113494873, "learning_rate": 5.673689717449348e-06, "loss": 4.3786, "step": 1059200 }, { "epoch": 6.387103387561655, "grad_norm": 3.776258707046509, "learning_rate": 5.6681473622851355e-06, "loss": 4.3506, "step": 1059250 }, { "epoch": 6.387404879342989, "grad_norm": 3.380295991897583, "learning_rate": 5.6626076633426834e-06, "loss": 4.2201, "step": 1059300 }, { "epoch": 6.387706371124323, "grad_norm": 3.623380422592163, "learning_rate": 5.657070620723958e-06, "loss": 4.0296, "step": 1059350 }, { "epoch": 6.388007862905657, "grad_norm": 5.107817649841309, "learning_rate": 5.651536234530879e-06, "loss": 4.47, "step": 1059400 }, { "epoch": 6.388309354686991, "grad_norm": 3.9228627681732178, "learning_rate": 5.6460045048652815e-06, "loss": 4.2799, "step": 1059450 }, { "epoch": 6.3886108464683256, "grad_norm": 6.383609294891357, "learning_rate": 5.6404754318289335e-06, "loss": 4.2193, "step": 1059500 }, { "epoch": 6.388912338249659, "grad_norm": 3.941152334213257, "learning_rate": 5.634949015523654e-06, "loss": 4.1363, "step": 1059550 }, { "epoch": 6.389213830030993, "grad_norm": 7.8037800788879395, "learning_rate": 5.6294252560510955e-06, "loss": 4.4477, "step": 1059600 }, { "epoch": 6.389515321812327, "grad_norm": 1.8421320915222168, "learning_rate": 5.623904153512943e-06, "loss": 4.2996, "step": 1059650 }, { "epoch": 6.3898168135936615, "grad_norm": 7.308849811553955, "learning_rate": 5.618385708010814e-06, "loss": 4.159, "step": 1059700 }, { "epoch": 6.390118305374996, "grad_norm": 6.477833271026611, "learning_rate": 5.612869919646229e-06, "loss": 4.1642, "step": 1059750 }, { "epoch": 6.39041979715633, "grad_norm": 5.46065092086792, "learning_rate": 5.607356788520723e-06, "loss": 4.1229, "step": 1059800 }, { "epoch": 6.390721288937663, "grad_norm": 4.4178466796875, "learning_rate": 5.601846314735781e-06, "loss": 4.3337, "step": 1059850 }, { "epoch": 6.391022780718997, "grad_norm": 6.944418430328369, "learning_rate": 5.5963384983927884e-06, "loss": 4.2613, "step": 1059900 }, { "epoch": 6.391324272500332, "grad_norm": 4.8031158447265625, "learning_rate": 5.590833339593098e-06, "loss": 4.1014, "step": 1059950 }, { "epoch": 6.391625764281666, "grad_norm": 5.318720817565918, "learning_rate": 5.585330838438046e-06, "loss": 4.1653, "step": 1060000 }, { "epoch": 6.391927256063, "grad_norm": 4.25435209274292, "learning_rate": 5.579830995028866e-06, "loss": 4.526, "step": 1060050 }, { "epoch": 6.392228747844333, "grad_norm": 9.254100799560547, "learning_rate": 5.574333809466814e-06, "loss": 4.1444, "step": 1060100 }, { "epoch": 6.392530239625668, "grad_norm": 4.611473083496094, "learning_rate": 5.568839281853038e-06, "loss": 4.0724, "step": 1060150 }, { "epoch": 6.392831731407002, "grad_norm": 7.994354724884033, "learning_rate": 5.56334741228866e-06, "loss": 4.4091, "step": 1060200 }, { "epoch": 6.393133223188336, "grad_norm": 3.0958800315856934, "learning_rate": 5.557858200874765e-06, "loss": 3.9495, "step": 1060250 }, { "epoch": 6.39343471496967, "grad_norm": 4.459410190582275, "learning_rate": 5.552371647712339e-06, "loss": 4.0844, "step": 1060300 }, { "epoch": 6.393736206751004, "grad_norm": 3.8742451667785645, "learning_rate": 5.546887752902402e-06, "loss": 4.4027, "step": 1060350 }, { "epoch": 6.394037698532338, "grad_norm": 4.845800399780273, "learning_rate": 5.541406516545821e-06, "loss": 4.5327, "step": 1060400 }, { "epoch": 6.394339190313672, "grad_norm": 6.094567775726318, "learning_rate": 5.535927938743501e-06, "loss": 3.8808, "step": 1060450 }, { "epoch": 6.394640682095006, "grad_norm": 5.6054911613464355, "learning_rate": 5.530452019596277e-06, "loss": 4.1229, "step": 1060500 }, { "epoch": 6.39494217387634, "grad_norm": 5.250927925109863, "learning_rate": 5.524978759204918e-06, "loss": 4.187, "step": 1060550 }, { "epoch": 6.3952436656576745, "grad_norm": 3.79430890083313, "learning_rate": 5.519508157670144e-06, "loss": 3.9344, "step": 1060600 }, { "epoch": 6.395545157439008, "grad_norm": 3.9942774772644043, "learning_rate": 5.514040215092641e-06, "loss": 4.1905, "step": 1060650 }, { "epoch": 6.395846649220342, "grad_norm": 4.26417875289917, "learning_rate": 5.508574931573029e-06, "loss": 4.1147, "step": 1060700 }, { "epoch": 6.396148141001676, "grad_norm": 4.39953088760376, "learning_rate": 5.5031123072118935e-06, "loss": 3.8514, "step": 1060750 }, { "epoch": 6.3964496327830105, "grad_norm": 2.9053828716278076, "learning_rate": 5.497652342109787e-06, "loss": 4.1961, "step": 1060800 }, { "epoch": 6.396751124564345, "grad_norm": 4.344259738922119, "learning_rate": 5.492195036367147e-06, "loss": 4.2349, "step": 1060850 }, { "epoch": 6.397052616345678, "grad_norm": 4.983435153961182, "learning_rate": 5.486740390084443e-06, "loss": 4.1637, "step": 1060900 }, { "epoch": 6.397354108127012, "grad_norm": 1.9502562284469604, "learning_rate": 5.481288403362061e-06, "loss": 4.0602, "step": 1060950 }, { "epoch": 6.397655599908346, "grad_norm": 5.1113080978393555, "learning_rate": 5.475839076300304e-06, "loss": 4.4782, "step": 1061000 }, { "epoch": 6.397957091689681, "grad_norm": 3.9576590061187744, "learning_rate": 5.470392408999508e-06, "loss": 4.0227, "step": 1061050 }, { "epoch": 6.398258583471015, "grad_norm": 7.023009777069092, "learning_rate": 5.464948401559877e-06, "loss": 4.3124, "step": 1061100 }, { "epoch": 6.398560075252348, "grad_norm": 10.573248863220215, "learning_rate": 5.459507054081597e-06, "loss": 4.3959, "step": 1061150 }, { "epoch": 6.398861567033682, "grad_norm": 2.4571382999420166, "learning_rate": 5.454068366664821e-06, "loss": 4.1788, "step": 1061200 }, { "epoch": 6.399163058815017, "grad_norm": 4.443305492401123, "learning_rate": 5.448632339409653e-06, "loss": 4.3281, "step": 1061250 }, { "epoch": 6.399464550596351, "grad_norm": 7.119052886962891, "learning_rate": 5.443198972416113e-06, "loss": 4.5737, "step": 1061300 }, { "epoch": 6.399766042377685, "grad_norm": 1.9635154008865356, "learning_rate": 5.437768265784187e-06, "loss": 4.6005, "step": 1061350 }, { "epoch": 6.400067534159019, "grad_norm": 5.368845462799072, "learning_rate": 5.4323402196138625e-06, "loss": 4.3983, "step": 1061400 }, { "epoch": 6.4003690259403525, "grad_norm": 4.293065547943115, "learning_rate": 5.4269148340049915e-06, "loss": 4.141, "step": 1061450 }, { "epoch": 6.400670517721687, "grad_norm": 5.575182914733887, "learning_rate": 5.4214921090574455e-06, "loss": 4.3069, "step": 1061500 }, { "epoch": 6.400972009503021, "grad_norm": 4.556519031524658, "learning_rate": 5.416072044871012e-06, "loss": 4.5588, "step": 1061550 }, { "epoch": 6.401273501284355, "grad_norm": 6.758466720581055, "learning_rate": 5.4106546415454255e-06, "loss": 3.8722, "step": 1061600 }, { "epoch": 6.401574993065689, "grad_norm": 5.570623874664307, "learning_rate": 5.405239899180391e-06, "loss": 4.3005, "step": 1061650 }, { "epoch": 6.401876484847023, "grad_norm": 7.428353786468506, "learning_rate": 5.399827817875596e-06, "loss": 4.4497, "step": 1061700 }, { "epoch": 6.402177976628357, "grad_norm": 6.313944339752197, "learning_rate": 5.394418397730593e-06, "loss": 4.3284, "step": 1061750 }, { "epoch": 6.402479468409691, "grad_norm": 4.185037136077881, "learning_rate": 5.38901163884497e-06, "loss": 4.0341, "step": 1061800 }, { "epoch": 6.402780960191025, "grad_norm": 5.666525363922119, "learning_rate": 5.383607541318213e-06, "loss": 4.4571, "step": 1061850 }, { "epoch": 6.4030824519723595, "grad_norm": 3.7539563179016113, "learning_rate": 5.3782061052497934e-06, "loss": 4.5552, "step": 1061900 }, { "epoch": 6.403383943753694, "grad_norm": 7.903051376342773, "learning_rate": 5.372807330739082e-06, "loss": 3.9689, "step": 1061950 }, { "epoch": 6.403685435535027, "grad_norm": 4.418202877044678, "learning_rate": 5.3674112178854815e-06, "loss": 3.7344, "step": 1062000 }, { "epoch": 6.403685435535027, "eval_loss": 4.808934211730957, "eval_runtime": 38.9453, "eval_samples_per_second": 13.147, "eval_steps_per_second": 6.573, "eval_tts_loss": 8.159434817842632, "step": 1062000 }, { "epoch": 6.403986927316361, "grad_norm": 4.46511173248291, "learning_rate": 5.362017766788262e-06, "loss": 4.3663, "step": 1062050 }, { "epoch": 6.404288419097695, "grad_norm": 5.808865070343018, "learning_rate": 5.35662697754668e-06, "loss": 4.2627, "step": 1062100 }, { "epoch": 6.40458991087903, "grad_norm": 5.187259197235107, "learning_rate": 5.351238850259986e-06, "loss": 4.0001, "step": 1062150 }, { "epoch": 6.404891402660364, "grad_norm": 3.910247325897217, "learning_rate": 5.345853385027305e-06, "loss": 4.1608, "step": 1062200 }, { "epoch": 6.405192894441697, "grad_norm": 1.7968999147415161, "learning_rate": 5.34047058194777e-06, "loss": 4.3529, "step": 1062250 }, { "epoch": 6.405494386223031, "grad_norm": 4.848423480987549, "learning_rate": 5.335090441120437e-06, "loss": 4.2378, "step": 1062300 }, { "epoch": 6.405795878004366, "grad_norm": 3.9097533226013184, "learning_rate": 5.329712962644295e-06, "loss": 4.3396, "step": 1062350 }, { "epoch": 6.4060973697857, "grad_norm": 4.754103660583496, "learning_rate": 5.324338146618329e-06, "loss": 3.6926, "step": 1062400 }, { "epoch": 6.406398861567034, "grad_norm": 6.163593292236328, "learning_rate": 5.318965993141477e-06, "loss": 4.1073, "step": 1062450 }, { "epoch": 6.406700353348368, "grad_norm": 5.865122318267822, "learning_rate": 5.3135965023125785e-06, "loss": 4.1979, "step": 1062500 }, { "epoch": 6.4070018451297015, "grad_norm": 6.237422466278076, "learning_rate": 5.3082296742304354e-06, "loss": 4.3331, "step": 1062550 }, { "epoch": 6.407303336911036, "grad_norm": 5.483893394470215, "learning_rate": 5.30286550899387e-06, "loss": 4.0165, "step": 1062600 }, { "epoch": 6.40760482869237, "grad_norm": 3.0104451179504395, "learning_rate": 5.2975040067015695e-06, "loss": 3.8176, "step": 1062650 }, { "epoch": 6.407906320473704, "grad_norm": 5.945208549499512, "learning_rate": 5.292145167452188e-06, "loss": 4.2567, "step": 1062700 }, { "epoch": 6.408207812255038, "grad_norm": 4.198639392852783, "learning_rate": 5.286788991344398e-06, "loss": 3.6912, "step": 1062750 }, { "epoch": 6.408509304036372, "grad_norm": 3.6688284873962402, "learning_rate": 5.2814354784767035e-06, "loss": 3.989, "step": 1062800 }, { "epoch": 6.408810795817706, "grad_norm": 5.283944606781006, "learning_rate": 5.276084628947658e-06, "loss": 4.1233, "step": 1062850 }, { "epoch": 6.40911228759904, "grad_norm": 5.850869178771973, "learning_rate": 5.270736442855783e-06, "loss": 4.0877, "step": 1062900 }, { "epoch": 6.409413779380374, "grad_norm": 4.566244602203369, "learning_rate": 5.2653909202994335e-06, "loss": 4.569, "step": 1062950 }, { "epoch": 6.4097152711617085, "grad_norm": 4.6368889808654785, "learning_rate": 5.2600480613770146e-06, "loss": 3.9489, "step": 1063000 }, { "epoch": 6.410016762943042, "grad_norm": 3.2052674293518066, "learning_rate": 5.25470786618688e-06, "loss": 4.1352, "step": 1063050 }, { "epoch": 6.410318254724376, "grad_norm": 4.893918037414551, "learning_rate": 5.2493703348272855e-06, "loss": 4.1398, "step": 1063100 }, { "epoch": 6.41061974650571, "grad_norm": 5.926498889923096, "learning_rate": 5.244035467396434e-06, "loss": 4.2816, "step": 1063150 }, { "epoch": 6.410921238287044, "grad_norm": 5.153188228607178, "learning_rate": 5.23870326399255e-06, "loss": 4.4196, "step": 1063200 }, { "epoch": 6.411222730068379, "grad_norm": 4.7614216804504395, "learning_rate": 5.233373724713735e-06, "loss": 4.3934, "step": 1063250 }, { "epoch": 6.411524221849712, "grad_norm": 5.392566204071045, "learning_rate": 5.22804684965808e-06, "loss": 4.3032, "step": 1063300 }, { "epoch": 6.411825713631046, "grad_norm": 1.909756064414978, "learning_rate": 5.222722638923638e-06, "loss": 4.0079, "step": 1063350 }, { "epoch": 6.41212720541238, "grad_norm": 3.8309106826782227, "learning_rate": 5.217401092608364e-06, "loss": 3.6157, "step": 1063400 }, { "epoch": 6.4124286971937146, "grad_norm": 5.395161151885986, "learning_rate": 5.212082210810231e-06, "loss": 4.432, "step": 1063450 }, { "epoch": 6.412730188975049, "grad_norm": 6.65133810043335, "learning_rate": 5.206765993627094e-06, "loss": 4.1829, "step": 1063500 }, { "epoch": 6.413031680756383, "grad_norm": 7.053497314453125, "learning_rate": 5.201452441156773e-06, "loss": 4.0792, "step": 1063550 }, { "epoch": 6.413333172537716, "grad_norm": 4.328364372253418, "learning_rate": 5.196141553497091e-06, "loss": 3.8982, "step": 1063600 }, { "epoch": 6.4136346643190505, "grad_norm": 5.014252185821533, "learning_rate": 5.190833330745786e-06, "loss": 4.1149, "step": 1063650 }, { "epoch": 6.413936156100385, "grad_norm": 4.135033130645752, "learning_rate": 5.185527773000531e-06, "loss": 3.8487, "step": 1063700 }, { "epoch": 6.414237647881719, "grad_norm": 3.3938355445861816, "learning_rate": 5.18022488035898e-06, "loss": 4.1884, "step": 1063750 }, { "epoch": 6.414539139663053, "grad_norm": 11.475470542907715, "learning_rate": 5.174924652918738e-06, "loss": 4.1368, "step": 1063800 }, { "epoch": 6.414840631444386, "grad_norm": 5.175677299499512, "learning_rate": 5.169627090777328e-06, "loss": 4.1316, "step": 1063850 }, { "epoch": 6.415142123225721, "grad_norm": 6.37137508392334, "learning_rate": 5.164332194032222e-06, "loss": 4.2303, "step": 1063900 }, { "epoch": 6.415443615007055, "grad_norm": 2.3174972534179688, "learning_rate": 5.159039962780892e-06, "loss": 4.4238, "step": 1063950 }, { "epoch": 6.415745106788389, "grad_norm": 4.333175182342529, "learning_rate": 5.15375039712076e-06, "loss": 4.0885, "step": 1064000 }, { "epoch": 6.416046598569723, "grad_norm": 8.695418357849121, "learning_rate": 5.148463497149113e-06, "loss": 4.3104, "step": 1064050 }, { "epoch": 6.4163480903510575, "grad_norm": 5.639506816864014, "learning_rate": 5.143179262963293e-06, "loss": 4.3161, "step": 1064100 }, { "epoch": 6.416649582132391, "grad_norm": 2.073101043701172, "learning_rate": 5.1378976946605534e-06, "loss": 3.6705, "step": 1064150 }, { "epoch": 6.416951073913725, "grad_norm": 2.923203468322754, "learning_rate": 5.132618792338067e-06, "loss": 4.2701, "step": 1064200 }, { "epoch": 6.417252565695059, "grad_norm": 5.592584609985352, "learning_rate": 5.12734255609299e-06, "loss": 4.0751, "step": 1064250 }, { "epoch": 6.417554057476393, "grad_norm": 2.673657178878784, "learning_rate": 5.1220689860224426e-06, "loss": 4.4611, "step": 1064300 }, { "epoch": 6.417855549257728, "grad_norm": 6.134788990020752, "learning_rate": 5.11679808222345e-06, "loss": 4.1016, "step": 1064350 }, { "epoch": 6.418157041039061, "grad_norm": 5.42886209487915, "learning_rate": 5.111529844793033e-06, "loss": 4.3223, "step": 1064400 }, { "epoch": 6.418458532820395, "grad_norm": 3.2752022743225098, "learning_rate": 5.1062642738281475e-06, "loss": 3.9114, "step": 1064450 }, { "epoch": 6.418760024601729, "grad_norm": 4.800781726837158, "learning_rate": 5.101001369425683e-06, "loss": 4.1137, "step": 1064500 }, { "epoch": 6.4190615163830635, "grad_norm": 1.050727367401123, "learning_rate": 5.0957411316824955e-06, "loss": 4.1378, "step": 1064550 }, { "epoch": 6.419363008164398, "grad_norm": 4.453561305999756, "learning_rate": 5.0904835606954245e-06, "loss": 4.2889, "step": 1064600 }, { "epoch": 6.419664499945731, "grad_norm": 6.908295154571533, "learning_rate": 5.085228656561224e-06, "loss": 4.1845, "step": 1064650 }, { "epoch": 6.419965991727065, "grad_norm": 4.821804046630859, "learning_rate": 5.0799764193765525e-06, "loss": 3.9438, "step": 1064700 }, { "epoch": 6.4202674835083995, "grad_norm": 3.9234750270843506, "learning_rate": 5.074726849238115e-06, "loss": 4.1981, "step": 1064750 }, { "epoch": 6.420568975289734, "grad_norm": 5.659223556518555, "learning_rate": 5.0694799462425e-06, "loss": 4.301, "step": 1064800 }, { "epoch": 6.420870467071068, "grad_norm": 3.673966884613037, "learning_rate": 5.064235710486281e-06, "loss": 4.3277, "step": 1064850 }, { "epoch": 6.421171958852401, "grad_norm": 5.449695110321045, "learning_rate": 5.058994142065998e-06, "loss": 4.0432, "step": 1064900 }, { "epoch": 6.421473450633735, "grad_norm": 4.255095481872559, "learning_rate": 5.053755241078056e-06, "loss": 4.3973, "step": 1064950 }, { "epoch": 6.42177494241507, "grad_norm": 8.429993629455566, "learning_rate": 5.048519007618912e-06, "loss": 4.2235, "step": 1065000 }, { "epoch": 6.42177494241507, "eval_loss": 4.8050432205200195, "eval_runtime": 39.2101, "eval_samples_per_second": 13.058, "eval_steps_per_second": 6.529, "eval_tts_loss": 8.136091955808878, "step": 1065000 }, { "epoch": 6.422076434196404, "grad_norm": 5.603877067565918, "learning_rate": 5.043285441784905e-06, "loss": 4.0123, "step": 1065050 }, { "epoch": 6.422377925977738, "grad_norm": 6.418548107147217, "learning_rate": 5.0380545436723765e-06, "loss": 4.2022, "step": 1065100 }, { "epoch": 6.422679417759072, "grad_norm": 4.241071701049805, "learning_rate": 5.0328263133775806e-06, "loss": 4.2119, "step": 1065150 }, { "epoch": 6.422980909540406, "grad_norm": 7.467248439788818, "learning_rate": 5.027600750996741e-06, "loss": 4.3837, "step": 1065200 }, { "epoch": 6.42328240132174, "grad_norm": 5.157438278198242, "learning_rate": 5.022377856626015e-06, "loss": 4.3739, "step": 1065250 }, { "epoch": 6.423583893103074, "grad_norm": 4.655951976776123, "learning_rate": 5.017157630361524e-06, "loss": 3.9549, "step": 1065300 }, { "epoch": 6.423885384884408, "grad_norm": 4.313179969787598, "learning_rate": 5.011940072299375e-06, "loss": 4.1194, "step": 1065350 }, { "epoch": 6.424186876665742, "grad_norm": 5.253937721252441, "learning_rate": 5.0067251825355255e-06, "loss": 3.8736, "step": 1065400 }, { "epoch": 6.424488368447076, "grad_norm": 5.063797473907471, "learning_rate": 5.001512961165998e-06, "loss": 4.5404, "step": 1065450 }, { "epoch": 6.42478986022841, "grad_norm": 4.2885026931762695, "learning_rate": 4.996303408286717e-06, "loss": 4.222, "step": 1065500 }, { "epoch": 6.425091352009744, "grad_norm": 7.843709468841553, "learning_rate": 4.991096523993521e-06, "loss": 4.2731, "step": 1065550 }, { "epoch": 6.425392843791078, "grad_norm": 4.150062084197998, "learning_rate": 4.98589230838225e-06, "loss": 4.213, "step": 1065600 }, { "epoch": 6.4256943355724125, "grad_norm": 4.493993759155273, "learning_rate": 4.980690761548695e-06, "loss": 4.1726, "step": 1065650 }, { "epoch": 6.425995827353747, "grad_norm": 5.625463962554932, "learning_rate": 4.975491883588561e-06, "loss": 4.0589, "step": 1065700 }, { "epoch": 6.42629731913508, "grad_norm": 3.134838104248047, "learning_rate": 4.970295674597541e-06, "loss": 4.2157, "step": 1065750 }, { "epoch": 6.426598810916414, "grad_norm": 5.071561813354492, "learning_rate": 4.965102134671272e-06, "loss": 4.2, "step": 1065800 }, { "epoch": 6.4269003026977485, "grad_norm": 6.134539604187012, "learning_rate": 4.9599112639053304e-06, "loss": 4.4053, "step": 1065850 }, { "epoch": 6.427201794479083, "grad_norm": 4.351550102233887, "learning_rate": 4.9547230623952215e-06, "loss": 4.1609, "step": 1065900 }, { "epoch": 6.427503286260417, "grad_norm": 3.8696341514587402, "learning_rate": 4.949537530236453e-06, "loss": 4.2439, "step": 1065950 }, { "epoch": 6.42780477804175, "grad_norm": 3.8035295009613037, "learning_rate": 4.944354667524431e-06, "loss": 4.4399, "step": 1066000 }, { "epoch": 6.428106269823084, "grad_norm": 6.85897970199585, "learning_rate": 4.9391744743545625e-06, "loss": 4.3299, "step": 1066050 }, { "epoch": 6.428407761604419, "grad_norm": 5.492850303649902, "learning_rate": 4.9339969508221885e-06, "loss": 4.2544, "step": 1066100 }, { "epoch": 6.428709253385753, "grad_norm": 6.195430755615234, "learning_rate": 4.928822097022567e-06, "loss": 4.3368, "step": 1066150 }, { "epoch": 6.429010745167087, "grad_norm": 6.592685222625732, "learning_rate": 4.923649913050936e-06, "loss": 3.9881, "step": 1066200 }, { "epoch": 6.429312236948421, "grad_norm": 7.0160064697265625, "learning_rate": 4.918480399002523e-06, "loss": 4.1027, "step": 1066250 }, { "epoch": 6.429613728729755, "grad_norm": 3.783984899520874, "learning_rate": 4.913313554972431e-06, "loss": 4.0051, "step": 1066300 }, { "epoch": 6.429915220511089, "grad_norm": 4.394341945648193, "learning_rate": 4.908149381055737e-06, "loss": 4.0458, "step": 1066350 }, { "epoch": 6.430216712292423, "grad_norm": 5.600076198577881, "learning_rate": 4.902987877347514e-06, "loss": 4.3148, "step": 1066400 }, { "epoch": 6.430518204073757, "grad_norm": 4.868718147277832, "learning_rate": 4.897829043942702e-06, "loss": 4.3177, "step": 1066450 }, { "epoch": 6.430819695855091, "grad_norm": 3.862081527709961, "learning_rate": 4.892672880936294e-06, "loss": 4.0677, "step": 1066500 }, { "epoch": 6.431121187636425, "grad_norm": 3.534482002258301, "learning_rate": 4.8875193884231775e-06, "loss": 4.2192, "step": 1066550 }, { "epoch": 6.431422679417759, "grad_norm": 5.615919589996338, "learning_rate": 4.882368566498163e-06, "loss": 4.5701, "step": 1066600 }, { "epoch": 6.431724171199093, "grad_norm": 4.142369747161865, "learning_rate": 4.877220415256072e-06, "loss": 4.2375, "step": 1066650 }, { "epoch": 6.432025662980427, "grad_norm": 5.086777687072754, "learning_rate": 4.872074934791631e-06, "loss": 4.1014, "step": 1066700 }, { "epoch": 6.4323271547617615, "grad_norm": 4.972324371337891, "learning_rate": 4.866932125199546e-06, "loss": 4.2399, "step": 1066750 }, { "epoch": 6.432628646543095, "grad_norm": 5.973069667816162, "learning_rate": 4.861791986574443e-06, "loss": 4.0265, "step": 1066800 }, { "epoch": 6.432930138324429, "grad_norm": 5.9630231857299805, "learning_rate": 4.856654519010961e-06, "loss": 4.2631, "step": 1066850 }, { "epoch": 6.433231630105763, "grad_norm": 4.279388904571533, "learning_rate": 4.851519722603592e-06, "loss": 3.7415, "step": 1066900 }, { "epoch": 6.4335331218870975, "grad_norm": 2.6502349376678467, "learning_rate": 4.8463875974468765e-06, "loss": 3.8072, "step": 1066950 }, { "epoch": 6.433834613668432, "grad_norm": 11.449715614318848, "learning_rate": 4.841258143635257e-06, "loss": 4.0268, "step": 1067000 }, { "epoch": 6.434136105449765, "grad_norm": 4.383812427520752, "learning_rate": 4.83613136126314e-06, "loss": 3.9628, "step": 1067050 }, { "epoch": 6.434437597231099, "grad_norm": 5.255950450897217, "learning_rate": 4.831007250424851e-06, "loss": 4.7194, "step": 1067100 }, { "epoch": 6.434739089012433, "grad_norm": 3.3606884479522705, "learning_rate": 4.825885811214697e-06, "loss": 4.3426, "step": 1067150 }, { "epoch": 6.435040580793768, "grad_norm": 4.848697185516357, "learning_rate": 4.82076704372697e-06, "loss": 4.0264, "step": 1067200 }, { "epoch": 6.435342072575102, "grad_norm": 4.2395148277282715, "learning_rate": 4.81565094805581e-06, "loss": 4.4099, "step": 1067250 }, { "epoch": 6.435643564356436, "grad_norm": 9.029749870300293, "learning_rate": 4.8105375242954105e-06, "loss": 4.5231, "step": 1067300 }, { "epoch": 6.435945056137769, "grad_norm": 6.051209449768066, "learning_rate": 4.8054267725398945e-06, "loss": 4.2339, "step": 1067350 }, { "epoch": 6.4362465479191036, "grad_norm": 5.1718974113464355, "learning_rate": 4.800318692883271e-06, "loss": 4.0484, "step": 1067400 }, { "epoch": 6.436548039700438, "grad_norm": 4.45321798324585, "learning_rate": 4.7952132854195965e-06, "loss": 4.0522, "step": 1067450 }, { "epoch": 6.436849531481772, "grad_norm": 5.749359607696533, "learning_rate": 4.790110550242798e-06, "loss": 4.4012, "step": 1067500 }, { "epoch": 6.437151023263106, "grad_norm": 4.563320159912109, "learning_rate": 4.785010487446766e-06, "loss": 4.3457, "step": 1067550 }, { "epoch": 6.4374525150444395, "grad_norm": 6.228012561798096, "learning_rate": 4.779913097125393e-06, "loss": 4.1579, "step": 1067600 }, { "epoch": 6.437754006825774, "grad_norm": 2.2598114013671875, "learning_rate": 4.7748183793724856e-06, "loss": 3.9564, "step": 1067650 }, { "epoch": 6.438055498607108, "grad_norm": 1.8764240741729736, "learning_rate": 4.769726334281804e-06, "loss": 4.1518, "step": 1067700 }, { "epoch": 6.438356990388442, "grad_norm": 6.542296886444092, "learning_rate": 4.7646369619470385e-06, "loss": 4.0806, "step": 1067750 }, { "epoch": 6.438658482169776, "grad_norm": 6.24224328994751, "learning_rate": 4.759550262461881e-06, "loss": 3.8153, "step": 1067800 }, { "epoch": 6.4389599739511105, "grad_norm": 3.8459763526916504, "learning_rate": 4.754466235919924e-06, "loss": 4.206, "step": 1067850 }, { "epoch": 6.439261465732444, "grad_norm": 4.69619607925415, "learning_rate": 4.749384882414742e-06, "loss": 4.1263, "step": 1067900 }, { "epoch": 6.439562957513778, "grad_norm": 5.631230354309082, "learning_rate": 4.74430620203986e-06, "loss": 4.3031, "step": 1067950 }, { "epoch": 6.439864449295112, "grad_norm": 7.786375522613525, "learning_rate": 4.739230194888705e-06, "loss": 4.2568, "step": 1068000 }, { "epoch": 6.439864449295112, "eval_loss": 4.801137924194336, "eval_runtime": 39.0296, "eval_samples_per_second": 13.118, "eval_steps_per_second": 6.559, "eval_tts_loss": 8.1360477826064, "step": 1068000 }, { "epoch": 6.4401659410764465, "grad_norm": 5.613545894622803, "learning_rate": 4.734156861054733e-06, "loss": 4.3081, "step": 1068050 }, { "epoch": 6.440467432857781, "grad_norm": 2.337871789932251, "learning_rate": 4.729086200631305e-06, "loss": 3.8913, "step": 1068100 }, { "epoch": 6.440768924639114, "grad_norm": 5.182715892791748, "learning_rate": 4.724018213711711e-06, "loss": 4.6925, "step": 1068150 }, { "epoch": 6.441070416420448, "grad_norm": 4.8583526611328125, "learning_rate": 4.718952900389278e-06, "loss": 4.5041, "step": 1068200 }, { "epoch": 6.441371908201782, "grad_norm": 4.938063144683838, "learning_rate": 4.71389026075718e-06, "loss": 3.7218, "step": 1068250 }, { "epoch": 6.441673399983117, "grad_norm": 4.806929111480713, "learning_rate": 4.708830294908578e-06, "loss": 4.3477, "step": 1068300 }, { "epoch": 6.441974891764451, "grad_norm": 4.561183929443359, "learning_rate": 4.703773002936612e-06, "loss": 4.6148, "step": 1068350 }, { "epoch": 6.442276383545784, "grad_norm": 6.446333408355713, "learning_rate": 4.698718384934375e-06, "loss": 4.3629, "step": 1068400 }, { "epoch": 6.442577875327118, "grad_norm": 3.1789445877075195, "learning_rate": 4.693666440994842e-06, "loss": 4.367, "step": 1068450 }, { "epoch": 6.4428793671084525, "grad_norm": 4.463071823120117, "learning_rate": 4.688617171211023e-06, "loss": 4.4065, "step": 1068500 }, { "epoch": 6.443180858889787, "grad_norm": 7.309439182281494, "learning_rate": 4.68357057567586e-06, "loss": 4.3907, "step": 1068550 }, { "epoch": 6.443482350671121, "grad_norm": 8.124037742614746, "learning_rate": 4.678526654482178e-06, "loss": 4.4433, "step": 1068600 }, { "epoch": 6.443783842452454, "grad_norm": 4.036209583282471, "learning_rate": 4.673485407722838e-06, "loss": 4.1575, "step": 1068650 }, { "epoch": 6.4440853342337885, "grad_norm": 5.917115211486816, "learning_rate": 4.66844683549063e-06, "loss": 4.2253, "step": 1068700 }, { "epoch": 6.444386826015123, "grad_norm": 5.421265125274658, "learning_rate": 4.663410937878215e-06, "loss": 4.5571, "step": 1068750 }, { "epoch": 6.444688317796457, "grad_norm": 3.91257381439209, "learning_rate": 4.6583777149783345e-06, "loss": 3.853, "step": 1068800 }, { "epoch": 6.444989809577791, "grad_norm": 8.27042293548584, "learning_rate": 4.653347166883598e-06, "loss": 4.2785, "step": 1068850 }, { "epoch": 6.445291301359125, "grad_norm": 5.889756679534912, "learning_rate": 4.648319293686581e-06, "loss": 4.1674, "step": 1068900 }, { "epoch": 6.445592793140459, "grad_norm": 4.865573406219482, "learning_rate": 4.643294095479827e-06, "loss": 4.3019, "step": 1068950 }, { "epoch": 6.445894284921793, "grad_norm": 4.804474353790283, "learning_rate": 4.638271572355811e-06, "loss": 3.9691, "step": 1069000 }, { "epoch": 6.446195776703127, "grad_norm": 4.184410572052002, "learning_rate": 4.633251724406978e-06, "loss": 4.1986, "step": 1069050 }, { "epoch": 6.446497268484461, "grad_norm": 5.756594181060791, "learning_rate": 4.628234551725668e-06, "loss": 3.8349, "step": 1069100 }, { "epoch": 6.4467987602657955, "grad_norm": 4.359170436859131, "learning_rate": 4.6232200544042754e-06, "loss": 4.2789, "step": 1069150 }, { "epoch": 6.447100252047129, "grad_norm": 4.410272121429443, "learning_rate": 4.618208232535025e-06, "loss": 4.3238, "step": 1069200 }, { "epoch": 6.447401743828463, "grad_norm": 6.810302734375, "learning_rate": 4.613199086210179e-06, "loss": 4.3162, "step": 1069250 }, { "epoch": 6.447703235609797, "grad_norm": 4.7126007080078125, "learning_rate": 4.608192615521944e-06, "loss": 4.5209, "step": 1069300 }, { "epoch": 6.448004727391131, "grad_norm": 4.673712253570557, "learning_rate": 4.6031888205624315e-06, "loss": 3.9763, "step": 1069350 }, { "epoch": 6.448306219172466, "grad_norm": 6.930112361907959, "learning_rate": 4.59818770142375e-06, "loss": 4.3713, "step": 1069400 }, { "epoch": 6.4486077109538, "grad_norm": 7.814266204833984, "learning_rate": 4.593189258197927e-06, "loss": 4.1914, "step": 1069450 }, { "epoch": 6.448909202735133, "grad_norm": 4.675078392028809, "learning_rate": 4.5881934909769215e-06, "loss": 4.4211, "step": 1069500 }, { "epoch": 6.449210694516467, "grad_norm": 4.129077434539795, "learning_rate": 4.58320039985271e-06, "loss": 4.0842, "step": 1069550 }, { "epoch": 6.4495121862978015, "grad_norm": 7.7053728103637695, "learning_rate": 4.578209984917186e-06, "loss": 4.2608, "step": 1069600 }, { "epoch": 6.449813678079136, "grad_norm": 3.8988049030303955, "learning_rate": 4.573222246262159e-06, "loss": 4.2327, "step": 1069650 }, { "epoch": 6.45011516986047, "grad_norm": 4.549405574798584, "learning_rate": 4.568237183979456e-06, "loss": 4.4873, "step": 1069700 }, { "epoch": 6.450416661641803, "grad_norm": 3.6354622840881348, "learning_rate": 4.563254798160804e-06, "loss": 4.1765, "step": 1069750 }, { "epoch": 6.4507181534231375, "grad_norm": 4.755942344665527, "learning_rate": 4.558275088897895e-06, "loss": 3.8697, "step": 1069800 }, { "epoch": 6.451019645204472, "grad_norm": 7.626578330993652, "learning_rate": 4.55329805628239e-06, "loss": 4.1122, "step": 1069850 }, { "epoch": 6.451321136985806, "grad_norm": 4.804696083068848, "learning_rate": 4.5483237004058644e-06, "loss": 4.26, "step": 1069900 }, { "epoch": 6.45162262876714, "grad_norm": 5.918002128601074, "learning_rate": 4.543352021359864e-06, "loss": 4.0851, "step": 1069950 }, { "epoch": 6.451924120548474, "grad_norm": 2.800182580947876, "learning_rate": 4.538383019235881e-06, "loss": 4.2138, "step": 1070000 }, { "epoch": 6.452225612329808, "grad_norm": 5.296895503997803, "learning_rate": 4.533416694125392e-06, "loss": 3.839, "step": 1070050 }, { "epoch": 6.452527104111142, "grad_norm": 5.945093154907227, "learning_rate": 4.528453046119773e-06, "loss": 4.5592, "step": 1070100 }, { "epoch": 6.452828595892476, "grad_norm": 4.817762851715088, "learning_rate": 4.523492075310353e-06, "loss": 4.3185, "step": 1070150 }, { "epoch": 6.45313008767381, "grad_norm": 4.926286697387695, "learning_rate": 4.518533781788492e-06, "loss": 3.7245, "step": 1070200 }, { "epoch": 6.4534315794551445, "grad_norm": 6.6781487464904785, "learning_rate": 4.513578165645398e-06, "loss": 4.3403, "step": 1070250 }, { "epoch": 6.453733071236478, "grad_norm": 6.404109001159668, "learning_rate": 4.508625226972268e-06, "loss": 4.1901, "step": 1070300 }, { "epoch": 6.454034563017812, "grad_norm": 4.7104315757751465, "learning_rate": 4.50367496586026e-06, "loss": 4.575, "step": 1070350 }, { "epoch": 6.454336054799146, "grad_norm": 4.459618091583252, "learning_rate": 4.498727382400486e-06, "loss": 4.0141, "step": 1070400 }, { "epoch": 6.45463754658048, "grad_norm": 4.191274166107178, "learning_rate": 4.493782476683971e-06, "loss": 4.132, "step": 1070450 }, { "epoch": 6.454939038361815, "grad_norm": 4.3921799659729, "learning_rate": 4.488840248801762e-06, "loss": 4.3643, "step": 1070500 }, { "epoch": 6.455240530143148, "grad_norm": 6.497936725616455, "learning_rate": 4.4839006988448e-06, "loss": 4.1931, "step": 1070550 }, { "epoch": 6.455542021924482, "grad_norm": 6.190058708190918, "learning_rate": 4.47896382690398e-06, "loss": 4.414, "step": 1070600 }, { "epoch": 6.455843513705816, "grad_norm": 2.7001426219940186, "learning_rate": 4.474029633070147e-06, "loss": 4.0844, "step": 1070650 }, { "epoch": 6.4561450054871505, "grad_norm": 7.194271564483643, "learning_rate": 4.469098117434144e-06, "loss": 4.3579, "step": 1070700 }, { "epoch": 6.456446497268485, "grad_norm": 4.428004264831543, "learning_rate": 4.464169280086683e-06, "loss": 4.0336, "step": 1070750 }, { "epoch": 6.456747989049818, "grad_norm": 9.569971084594727, "learning_rate": 4.4592431211184885e-06, "loss": 4.5942, "step": 1070800 }, { "epoch": 6.457049480831152, "grad_norm": 2.8530542850494385, "learning_rate": 4.454319640620258e-06, "loss": 4.1971, "step": 1070850 }, { "epoch": 6.4573509726124865, "grad_norm": 5.660614013671875, "learning_rate": 4.44939883868255e-06, "loss": 3.9258, "step": 1070900 }, { "epoch": 6.457652464393821, "grad_norm": 98.32249450683594, "learning_rate": 4.444480715395943e-06, "loss": 4.3369, "step": 1070950 }, { "epoch": 6.457953956175155, "grad_norm": 2.7639200687408447, "learning_rate": 4.439565270850964e-06, "loss": 4.2356, "step": 1071000 }, { "epoch": 6.457953956175155, "eval_loss": 4.8028974533081055, "eval_runtime": 39.3359, "eval_samples_per_second": 13.016, "eval_steps_per_second": 6.508, "eval_tts_loss": 8.162800615134165, "step": 1071000 }, { "epoch": 6.458255447956489, "grad_norm": 2.714669704437256, "learning_rate": 4.434652505138059e-06, "loss": 4.0763, "step": 1071050 }, { "epoch": 6.458556939737822, "grad_norm": 1.973831057548523, "learning_rate": 4.429742418347637e-06, "loss": 4.1068, "step": 1071100 }, { "epoch": 6.458858431519157, "grad_norm": 2.8994393348693848, "learning_rate": 4.42483501057006e-06, "loss": 4.2412, "step": 1071150 }, { "epoch": 6.459159923300491, "grad_norm": 5.402451992034912, "learning_rate": 4.419930281895656e-06, "loss": 4.05, "step": 1071200 }, { "epoch": 6.459461415081825, "grad_norm": 4.071234226226807, "learning_rate": 4.415028232414669e-06, "loss": 4.3027, "step": 1071250 }, { "epoch": 6.459762906863159, "grad_norm": 7.159228324890137, "learning_rate": 4.410128862217344e-06, "loss": 4.1186, "step": 1071300 }, { "epoch": 6.460064398644493, "grad_norm": 5.208163738250732, "learning_rate": 4.405232171393807e-06, "loss": 4.4248, "step": 1071350 }, { "epoch": 6.460365890425827, "grad_norm": 3.9545838832855225, "learning_rate": 4.400338160034222e-06, "loss": 4.1037, "step": 1071400 }, { "epoch": 6.460667382207161, "grad_norm": 4.552319049835205, "learning_rate": 4.395446828228616e-06, "loss": 4.2714, "step": 1071450 }, { "epoch": 6.460968873988495, "grad_norm": 7.320040225982666, "learning_rate": 4.390558176067016e-06, "loss": 4.0243, "step": 1071500 }, { "epoch": 6.461270365769829, "grad_norm": 5.623331069946289, "learning_rate": 4.385672203639401e-06, "loss": 4.311, "step": 1071550 }, { "epoch": 6.461571857551164, "grad_norm": 2.6739914417266846, "learning_rate": 4.380788911035699e-06, "loss": 4.054, "step": 1071600 }, { "epoch": 6.461873349332497, "grad_norm": 4.635369300842285, "learning_rate": 4.375908298345754e-06, "loss": 4.1339, "step": 1071650 }, { "epoch": 6.462174841113831, "grad_norm": 4.260566234588623, "learning_rate": 4.371030365659412e-06, "loss": 4.259, "step": 1071700 }, { "epoch": 6.462476332895165, "grad_norm": 3.9687392711639404, "learning_rate": 4.366155113066433e-06, "loss": 4.3802, "step": 1071750 }, { "epoch": 6.4627778246764995, "grad_norm": 5.090049743652344, "learning_rate": 4.361282540656513e-06, "loss": 3.8742, "step": 1071800 }, { "epoch": 6.463079316457834, "grad_norm": 7.7588019371032715, "learning_rate": 4.356412648519381e-06, "loss": 4.1529, "step": 1071850 }, { "epoch": 6.463380808239167, "grad_norm": 5.24312686920166, "learning_rate": 4.35154543674463e-06, "loss": 4.3639, "step": 1071900 }, { "epoch": 6.463682300020501, "grad_norm": 3.616966724395752, "learning_rate": 4.346680905421806e-06, "loss": 4.1868, "step": 1071950 }, { "epoch": 6.4639837918018355, "grad_norm": 4.094785690307617, "learning_rate": 4.341819054640472e-06, "loss": 4.1493, "step": 1072000 }, { "epoch": 6.46428528358317, "grad_norm": 6.030837059020996, "learning_rate": 4.336959884490104e-06, "loss": 4.4256, "step": 1072050 }, { "epoch": 6.464586775364504, "grad_norm": 4.774924278259277, "learning_rate": 4.332103395060116e-06, "loss": 4.075, "step": 1072100 }, { "epoch": 6.464888267145837, "grad_norm": 4.79457426071167, "learning_rate": 4.327249586439868e-06, "loss": 4.0537, "step": 1072150 }, { "epoch": 6.465189758927171, "grad_norm": 4.254610538482666, "learning_rate": 4.322398458718723e-06, "loss": 4.2518, "step": 1072200 }, { "epoch": 6.465491250708506, "grad_norm": 3.603618621826172, "learning_rate": 4.317550011985943e-06, "loss": 4.1838, "step": 1072250 }, { "epoch": 6.46579274248984, "grad_norm": 5.05917501449585, "learning_rate": 4.312704246330739e-06, "loss": 4.08, "step": 1072300 }, { "epoch": 6.466094234271174, "grad_norm": 5.963038921356201, "learning_rate": 4.307861161842324e-06, "loss": 4.1073, "step": 1072350 }, { "epoch": 6.466395726052507, "grad_norm": 3.265324115753174, "learning_rate": 4.303020758609793e-06, "loss": 4.0244, "step": 1072400 }, { "epoch": 6.4666972178338415, "grad_norm": 4.537345886230469, "learning_rate": 4.298183036722258e-06, "loss": 4.4222, "step": 1072450 }, { "epoch": 6.466998709615176, "grad_norm": 7.400811195373535, "learning_rate": 4.2933479962687315e-06, "loss": 4.3739, "step": 1072500 }, { "epoch": 6.46730020139651, "grad_norm": 1.7905967235565186, "learning_rate": 4.288515637338191e-06, "loss": 4.2963, "step": 1072550 }, { "epoch": 6.467601693177844, "grad_norm": 4.554258346557617, "learning_rate": 4.2836859600196e-06, "loss": 4.4973, "step": 1072600 }, { "epoch": 6.467903184959178, "grad_norm": 5.947230815887451, "learning_rate": 4.278858964401821e-06, "loss": 4.2595, "step": 1072650 }, { "epoch": 6.468204676740512, "grad_norm": 4.286692142486572, "learning_rate": 4.274034650573666e-06, "loss": 4.2418, "step": 1072700 }, { "epoch": 6.468506168521846, "grad_norm": 3.9190309047698975, "learning_rate": 4.269213018623946e-06, "loss": 4.299, "step": 1072750 }, { "epoch": 6.46880766030318, "grad_norm": 4.456496238708496, "learning_rate": 4.264394068641408e-06, "loss": 4.2963, "step": 1072800 }, { "epoch": 6.469109152084514, "grad_norm": 4.193281650543213, "learning_rate": 4.259577800714697e-06, "loss": 4.4381, "step": 1072850 }, { "epoch": 6.4694106438658485, "grad_norm": 6.190328121185303, "learning_rate": 4.254764214932477e-06, "loss": 4.3718, "step": 1072900 }, { "epoch": 6.469712135647182, "grad_norm": 7.107264041900635, "learning_rate": 4.249953311383358e-06, "loss": 4.4447, "step": 1072950 }, { "epoch": 6.470013627428516, "grad_norm": 5.1403656005859375, "learning_rate": 4.245145090155821e-06, "loss": 4.3976, "step": 1073000 }, { "epoch": 6.47031511920985, "grad_norm": 4.894639492034912, "learning_rate": 4.2403395513384116e-06, "loss": 4.5704, "step": 1073050 }, { "epoch": 6.4706166109911845, "grad_norm": 4.270461082458496, "learning_rate": 4.235536695019542e-06, "loss": 4.0871, "step": 1073100 }, { "epoch": 6.470918102772519, "grad_norm": 4.148770332336426, "learning_rate": 4.230736521287575e-06, "loss": 4.4042, "step": 1073150 }, { "epoch": 6.471219594553853, "grad_norm": 6.9405388832092285, "learning_rate": 4.22593903023089e-06, "loss": 3.7438, "step": 1073200 }, { "epoch": 6.471521086335186, "grad_norm": 4.947340965270996, "learning_rate": 4.2211442219377836e-06, "loss": 4.5759, "step": 1073250 }, { "epoch": 6.47182257811652, "grad_norm": 5.704004287719727, "learning_rate": 4.21635209649645e-06, "loss": 4.4882, "step": 1073300 }, { "epoch": 6.472124069897855, "grad_norm": 7.054965019226074, "learning_rate": 4.211562653995121e-06, "loss": 4.6271, "step": 1073350 }, { "epoch": 6.472425561679189, "grad_norm": 3.928463935852051, "learning_rate": 4.206775894521958e-06, "loss": 3.8385, "step": 1073400 }, { "epoch": 6.472727053460523, "grad_norm": 8.478699684143066, "learning_rate": 4.201991818165007e-06, "loss": 4.085, "step": 1073450 }, { "epoch": 6.473028545241856, "grad_norm": 7.055768013000488, "learning_rate": 4.197210425012331e-06, "loss": 4.0666, "step": 1073500 }, { "epoch": 6.4733300370231905, "grad_norm": 5.542347431182861, "learning_rate": 4.192431715151928e-06, "loss": 4.3779, "step": 1073550 }, { "epoch": 6.473631528804525, "grad_norm": 3.4673118591308594, "learning_rate": 4.187655688671726e-06, "loss": 4.0531, "step": 1073600 }, { "epoch": 6.473933020585859, "grad_norm": 4.084895610809326, "learning_rate": 4.182882345659655e-06, "loss": 4.2673, "step": 1073650 }, { "epoch": 6.474234512367193, "grad_norm": 2.1138112545013428, "learning_rate": 4.178111686203545e-06, "loss": 4.3861, "step": 1073700 }, { "epoch": 6.474536004148527, "grad_norm": 5.010216236114502, "learning_rate": 4.173343710391158e-06, "loss": 4.4803, "step": 1073750 }, { "epoch": 6.474837495929861, "grad_norm": 2.9163546562194824, "learning_rate": 4.1685784183103085e-06, "loss": 3.9244, "step": 1073800 }, { "epoch": 6.475138987711195, "grad_norm": 2.99417781829834, "learning_rate": 4.163815810048643e-06, "loss": 4.1759, "step": 1073850 }, { "epoch": 6.475440479492529, "grad_norm": 5.512857437133789, "learning_rate": 4.159055885693841e-06, "loss": 4.3374, "step": 1073900 }, { "epoch": 6.475741971273863, "grad_norm": 6.447080612182617, "learning_rate": 4.154298645333481e-06, "loss": 3.9621, "step": 1073950 }, { "epoch": 6.4760434630551975, "grad_norm": 2.832470417022705, "learning_rate": 4.149544089055112e-06, "loss": 4.1725, "step": 1074000 }, { "epoch": 6.4760434630551975, "eval_loss": 4.801137447357178, "eval_runtime": 39.0384, "eval_samples_per_second": 13.115, "eval_steps_per_second": 6.558, "eval_tts_loss": 8.145335528192113, "step": 1074000 }, { "epoch": 6.476344954836531, "grad_norm": 4.0015997886657715, "learning_rate": 4.144792216946263e-06, "loss": 4.4087, "step": 1074050 }, { "epoch": 6.476646446617865, "grad_norm": 3.151115894317627, "learning_rate": 4.140043029094348e-06, "loss": 4.3528, "step": 1074100 }, { "epoch": 6.476947938399199, "grad_norm": 6.0786614418029785, "learning_rate": 4.135296525586796e-06, "loss": 4.0834, "step": 1074150 }, { "epoch": 6.4772494301805335, "grad_norm": 3.331329107284546, "learning_rate": 4.13055270651097e-06, "loss": 4.2753, "step": 1074200 }, { "epoch": 6.477550921961868, "grad_norm": 2.9182355403900146, "learning_rate": 4.125811571954135e-06, "loss": 4.0749, "step": 1074250 }, { "epoch": 6.477852413743201, "grad_norm": 4.001204490661621, "learning_rate": 4.121073122003571e-06, "loss": 4.2207, "step": 1074300 }, { "epoch": 6.478153905524535, "grad_norm": 4.204019546508789, "learning_rate": 4.116337356746474e-06, "loss": 4.2283, "step": 1074350 }, { "epoch": 6.478455397305869, "grad_norm": 3.9073286056518555, "learning_rate": 4.111604276269992e-06, "loss": 4.2721, "step": 1074400 }, { "epoch": 6.478756889087204, "grad_norm": 3.880800247192383, "learning_rate": 4.106873880661238e-06, "loss": 4.2582, "step": 1074450 }, { "epoch": 6.479058380868538, "grad_norm": 4.888484001159668, "learning_rate": 4.102146170007293e-06, "loss": 4.1113, "step": 1074500 }, { "epoch": 6.479359872649871, "grad_norm": 5.2851243019104, "learning_rate": 4.097421144395118e-06, "loss": 4.1494, "step": 1074550 }, { "epoch": 6.479661364431205, "grad_norm": 3.197683811187744, "learning_rate": 4.092698803911698e-06, "loss": 4.0774, "step": 1074600 }, { "epoch": 6.4799628562125395, "grad_norm": 1.6289910078048706, "learning_rate": 4.087979148643944e-06, "loss": 3.9634, "step": 1074650 }, { "epoch": 6.480264347993874, "grad_norm": 4.454398155212402, "learning_rate": 4.083262178678703e-06, "loss": 3.9532, "step": 1074700 }, { "epoch": 6.480565839775208, "grad_norm": 5.231525421142578, "learning_rate": 4.078547894102774e-06, "loss": 4.4505, "step": 1074750 }, { "epoch": 6.480867331556542, "grad_norm": 1.8731635808944702, "learning_rate": 4.073836295002953e-06, "loss": 3.9795, "step": 1074800 }, { "epoch": 6.4811688233378755, "grad_norm": 6.386946201324463, "learning_rate": 4.0691273814659035e-06, "loss": 4.4032, "step": 1074850 }, { "epoch": 6.48147031511921, "grad_norm": 4.080874443054199, "learning_rate": 4.064421153578307e-06, "loss": 3.9452, "step": 1074900 }, { "epoch": 6.481771806900544, "grad_norm": 5.943363666534424, "learning_rate": 4.0597176114268105e-06, "loss": 3.8253, "step": 1074950 }, { "epoch": 6.482073298681878, "grad_norm": 3.9029746055603027, "learning_rate": 4.055016755097928e-06, "loss": 4.4689, "step": 1075000 }, { "epoch": 6.482374790463212, "grad_norm": 5.911691665649414, "learning_rate": 4.050318584678175e-06, "loss": 4.0231, "step": 1075050 }, { "epoch": 6.482676282244546, "grad_norm": 5.026363849639893, "learning_rate": 4.045623100254047e-06, "loss": 3.9552, "step": 1075100 }, { "epoch": 6.48297777402588, "grad_norm": 6.008035182952881, "learning_rate": 4.04093030191191e-06, "loss": 4.4819, "step": 1075150 }, { "epoch": 6.483279265807214, "grad_norm": 5.509647846221924, "learning_rate": 4.036240189738177e-06, "loss": 4.2865, "step": 1075200 }, { "epoch": 6.483580757588548, "grad_norm": 4.552309036254883, "learning_rate": 4.031552763819146e-06, "loss": 4.2534, "step": 1075250 }, { "epoch": 6.4838822493698824, "grad_norm": 6.07069730758667, "learning_rate": 4.026868024241064e-06, "loss": 4.1436, "step": 1075300 }, { "epoch": 6.484183741151217, "grad_norm": 7.074362754821777, "learning_rate": 4.0221859710901474e-06, "loss": 4.5073, "step": 1075350 }, { "epoch": 6.48448523293255, "grad_norm": 4.533592700958252, "learning_rate": 4.0175066044526086e-06, "loss": 4.0668, "step": 1075400 }, { "epoch": 6.484786724713884, "grad_norm": 5.613218307495117, "learning_rate": 4.012829924414529e-06, "loss": 4.0432, "step": 1075450 }, { "epoch": 6.485088216495218, "grad_norm": 4.475771903991699, "learning_rate": 4.0081559310619735e-06, "loss": 4.0843, "step": 1075500 }, { "epoch": 6.485389708276553, "grad_norm": 4.06483793258667, "learning_rate": 4.003484624480957e-06, "loss": 4.1517, "step": 1075550 }, { "epoch": 6.485691200057887, "grad_norm": 5.126816272735596, "learning_rate": 3.99881600475746e-06, "loss": 4.2011, "step": 1075600 }, { "epoch": 6.48599269183922, "grad_norm": 8.11823844909668, "learning_rate": 3.994150071977398e-06, "loss": 3.95, "step": 1075650 }, { "epoch": 6.486294183620554, "grad_norm": 2.879767656326294, "learning_rate": 3.989486826226668e-06, "loss": 3.9676, "step": 1075700 }, { "epoch": 6.4865956754018885, "grad_norm": 2.765259027481079, "learning_rate": 3.984826267591035e-06, "loss": 4.0654, "step": 1075750 }, { "epoch": 6.486897167183223, "grad_norm": 3.2260751724243164, "learning_rate": 3.980168396156314e-06, "loss": 4.1031, "step": 1075800 }, { "epoch": 6.487198658964557, "grad_norm": 4.4447197914123535, "learning_rate": 3.975513212008219e-06, "loss": 4.2615, "step": 1075850 }, { "epoch": 6.48750015074589, "grad_norm": 5.874541759490967, "learning_rate": 3.970860715232399e-06, "loss": 4.1068, "step": 1075900 }, { "epoch": 6.4878016425272245, "grad_norm": 4.489511013031006, "learning_rate": 3.966210905914485e-06, "loss": 4.2421, "step": 1075950 }, { "epoch": 6.488103134308559, "grad_norm": 4.886292457580566, "learning_rate": 3.961563784140076e-06, "loss": 4.1855, "step": 1076000 }, { "epoch": 6.488404626089893, "grad_norm": 5.01751184463501, "learning_rate": 3.956919349994669e-06, "loss": 4.6425, "step": 1076050 }, { "epoch": 6.488706117871227, "grad_norm": 3.2230136394500732, "learning_rate": 3.9522776035637474e-06, "loss": 3.8681, "step": 1076100 }, { "epoch": 6.48900760965256, "grad_norm": 5.005280017852783, "learning_rate": 3.947638544932741e-06, "loss": 4.3252, "step": 1076150 }, { "epoch": 6.489309101433895, "grad_norm": 8.09465217590332, "learning_rate": 3.943002174187032e-06, "loss": 4.1745, "step": 1076200 }, { "epoch": 6.489610593215229, "grad_norm": 5.818909168243408, "learning_rate": 3.938368491411903e-06, "loss": 4.1585, "step": 1076250 }, { "epoch": 6.489912084996563, "grad_norm": 3.873276472091675, "learning_rate": 3.933737496692685e-06, "loss": 3.8299, "step": 1076300 }, { "epoch": 6.490213576777897, "grad_norm": 8.665993690490723, "learning_rate": 3.92910919011456e-06, "loss": 4.1846, "step": 1076350 }, { "epoch": 6.490515068559231, "grad_norm": 5.453371524810791, "learning_rate": 3.924483571762743e-06, "loss": 4.0134, "step": 1076400 }, { "epoch": 6.490816560340565, "grad_norm": 1.2115052938461304, "learning_rate": 3.91986064172235e-06, "loss": 3.9434, "step": 1076450 }, { "epoch": 6.491118052121899, "grad_norm": 4.360414981842041, "learning_rate": 3.915240400078429e-06, "loss": 4.077, "step": 1076500 }, { "epoch": 6.491419543903233, "grad_norm": 2.3845198154449463, "learning_rate": 3.910622846916045e-06, "loss": 4.0469, "step": 1076550 }, { "epoch": 6.491721035684567, "grad_norm": 6.531676292419434, "learning_rate": 3.906007982320181e-06, "loss": 4.0866, "step": 1076600 }, { "epoch": 6.492022527465902, "grad_norm": 6.68721342086792, "learning_rate": 3.901395806375751e-06, "loss": 3.9886, "step": 1076650 }, { "epoch": 6.492324019247235, "grad_norm": 4.224345684051514, "learning_rate": 3.896786319167622e-06, "loss": 4.1675, "step": 1076700 }, { "epoch": 6.492625511028569, "grad_norm": 5.625993251800537, "learning_rate": 3.892179520780642e-06, "loss": 4.1761, "step": 1076750 }, { "epoch": 6.492927002809903, "grad_norm": 5.756751537322998, "learning_rate": 3.887575411299593e-06, "loss": 4.1927, "step": 1076800 }, { "epoch": 6.4932284945912375, "grad_norm": 5.555429935455322, "learning_rate": 3.882973990809208e-06, "loss": 4.2011, "step": 1076850 }, { "epoch": 6.493529986372572, "grad_norm": 5.400635719299316, "learning_rate": 3.878375259394168e-06, "loss": 4.1028, "step": 1076900 }, { "epoch": 6.493831478153906, "grad_norm": 1.5962047576904297, "learning_rate": 3.873779217139089e-06, "loss": 4.4194, "step": 1076950 }, { "epoch": 6.494132969935239, "grad_norm": 5.0473151206970215, "learning_rate": 3.869185864128605e-06, "loss": 4.2181, "step": 1077000 }, { "epoch": 6.494132969935239, "eval_loss": 4.80047607421875, "eval_runtime": 39.1023, "eval_samples_per_second": 13.094, "eval_steps_per_second": 6.547, "eval_tts_loss": 8.162526840361409, "step": 1077000 }, { "epoch": 6.4944344617165735, "grad_norm": 4.237756729125977, "learning_rate": 3.864595200447196e-06, "loss": 4.1446, "step": 1077050 }, { "epoch": 6.494735953497908, "grad_norm": 3.9284849166870117, "learning_rate": 3.860007226179363e-06, "loss": 4.1304, "step": 1077100 }, { "epoch": 6.495037445279242, "grad_norm": 4.836784362792969, "learning_rate": 3.8554219414095364e-06, "loss": 4.0456, "step": 1077150 }, { "epoch": 6.495338937060576, "grad_norm": 3.57572078704834, "learning_rate": 3.850839346222117e-06, "loss": 3.7537, "step": 1077200 }, { "epoch": 6.495640428841909, "grad_norm": 5.606709003448486, "learning_rate": 3.846259440701438e-06, "loss": 4.3018, "step": 1077250 }, { "epoch": 6.495941920623244, "grad_norm": 5.903209209442139, "learning_rate": 3.841682224931763e-06, "loss": 4.2178, "step": 1077300 }, { "epoch": 6.496243412404578, "grad_norm": 6.745561122894287, "learning_rate": 3.837107698997377e-06, "loss": 4.4227, "step": 1077350 }, { "epoch": 6.496544904185912, "grad_norm": 6.00337553024292, "learning_rate": 3.832535862982411e-06, "loss": 4.1075, "step": 1077400 }, { "epoch": 6.496846395967246, "grad_norm": 4.847118854522705, "learning_rate": 3.827966716971048e-06, "loss": 4.1917, "step": 1077450 }, { "epoch": 6.49714788774858, "grad_norm": 4.9955949783325195, "learning_rate": 3.823400261047355e-06, "loss": 3.9926, "step": 1077500 }, { "epoch": 6.497449379529914, "grad_norm": 5.6605682373046875, "learning_rate": 3.8188364952953796e-06, "loss": 4.3353, "step": 1077550 }, { "epoch": 6.497750871311248, "grad_norm": 3.6104180812835693, "learning_rate": 3.81427541979909e-06, "loss": 4.6273, "step": 1077600 }, { "epoch": 6.498052363092582, "grad_norm": 7.0595855712890625, "learning_rate": 3.809717034642451e-06, "loss": 4.1727, "step": 1077650 }, { "epoch": 6.498353854873916, "grad_norm": 7.6106462478637695, "learning_rate": 3.8051613399093628e-06, "loss": 4.6871, "step": 1077700 }, { "epoch": 6.498655346655251, "grad_norm": 5.155742168426514, "learning_rate": 3.800608335683625e-06, "loss": 4.5928, "step": 1077750 }, { "epoch": 6.498956838436584, "grad_norm": 4.581457614898682, "learning_rate": 3.79605802204907e-06, "loss": 3.8817, "step": 1077800 }, { "epoch": 6.499258330217918, "grad_norm": 7.074932098388672, "learning_rate": 3.791510399089431e-06, "loss": 4.1737, "step": 1077850 }, { "epoch": 6.499559821999252, "grad_norm": 6.902014255523682, "learning_rate": 3.7869654668883586e-06, "loss": 4.3044, "step": 1077900 }, { "epoch": 6.4998613137805865, "grad_norm": 5.512104511260986, "learning_rate": 3.7824232255295517e-06, "loss": 4.3217, "step": 1077950 }, { "epoch": 6.500162805561921, "grad_norm": 5.557478427886963, "learning_rate": 3.777883675096577e-06, "loss": 4.082, "step": 1078000 }, { "epoch": 6.500464297343254, "grad_norm": 4.824726581573486, "learning_rate": 3.7733468156729842e-06, "loss": 4.3383, "step": 1078050 }, { "epoch": 6.500765789124588, "grad_norm": 1.802963137626648, "learning_rate": 3.7688126473422564e-06, "loss": 3.7743, "step": 1078100 }, { "epoch": 6.5010672809059225, "grad_norm": 6.448294639587402, "learning_rate": 3.7642811701878607e-06, "loss": 4.4122, "step": 1078150 }, { "epoch": 6.501368772687257, "grad_norm": 8.43116283416748, "learning_rate": 3.7597523842931963e-06, "loss": 4.4661, "step": 1078200 }, { "epoch": 6.501670264468591, "grad_norm": 6.873931407928467, "learning_rate": 3.7552262897415642e-06, "loss": 4.398, "step": 1078250 }, { "epoch": 6.501971756249924, "grad_norm": 7.438912391662598, "learning_rate": 3.7507028866163135e-06, "loss": 4.2908, "step": 1078300 }, { "epoch": 6.502273248031258, "grad_norm": 1.4924557209014893, "learning_rate": 3.746182175000645e-06, "loss": 4.4504, "step": 1078350 }, { "epoch": 6.502574739812593, "grad_norm": 4.347042083740234, "learning_rate": 3.741664154977775e-06, "loss": 4.2721, "step": 1078400 }, { "epoch": 6.502876231593927, "grad_norm": 3.2682785987854004, "learning_rate": 3.73714882663087e-06, "loss": 4.0948, "step": 1078450 }, { "epoch": 6.503177723375261, "grad_norm": 5.404279708862305, "learning_rate": 3.7326361900430147e-06, "loss": 4.1553, "step": 1078500 }, { "epoch": 6.503479215156595, "grad_norm": 6.7438812255859375, "learning_rate": 3.7281262452972414e-06, "loss": 4.3565, "step": 1078550 }, { "epoch": 6.5037807069379285, "grad_norm": 4.425195217132568, "learning_rate": 3.723618992476585e-06, "loss": 4.1717, "step": 1078600 }, { "epoch": 6.504082198719263, "grad_norm": 5.712733268737793, "learning_rate": 3.719114431663961e-06, "loss": 4.0759, "step": 1078650 }, { "epoch": 6.504383690500597, "grad_norm": 3.962728500366211, "learning_rate": 3.714612562942271e-06, "loss": 3.7338, "step": 1078700 }, { "epoch": 6.504685182281931, "grad_norm": 2.961639165878296, "learning_rate": 3.7101133863943975e-06, "loss": 4.2581, "step": 1078750 }, { "epoch": 6.504986674063265, "grad_norm": 8.169790267944336, "learning_rate": 3.705616902103109e-06, "loss": 4.1935, "step": 1078800 }, { "epoch": 6.505288165844599, "grad_norm": 4.534443378448486, "learning_rate": 3.7011231101511554e-06, "loss": 4.3499, "step": 1078850 }, { "epoch": 6.505589657625933, "grad_norm": 4.738056182861328, "learning_rate": 3.69663201062127e-06, "loss": 4.4799, "step": 1078900 }, { "epoch": 6.505891149407267, "grad_norm": 4.027085781097412, "learning_rate": 3.6921436035960872e-06, "loss": 4.4724, "step": 1078950 }, { "epoch": 6.506192641188601, "grad_norm": 5.470463752746582, "learning_rate": 3.6876578891582074e-06, "loss": 4.223, "step": 1079000 }, { "epoch": 6.5064941329699355, "grad_norm": 3.739736795425415, "learning_rate": 3.6831748673901816e-06, "loss": 4.0195, "step": 1079050 }, { "epoch": 6.50679562475127, "grad_norm": 4.42885684967041, "learning_rate": 3.6786945383745104e-06, "loss": 3.8147, "step": 1079100 }, { "epoch": 6.507097116532603, "grad_norm": 5.431604385375977, "learning_rate": 3.6742169021936606e-06, "loss": 4.019, "step": 1079150 }, { "epoch": 6.507398608313937, "grad_norm": 2.666184902191162, "learning_rate": 3.6697419589300334e-06, "loss": 4.0542, "step": 1079200 }, { "epoch": 6.5077001000952714, "grad_norm": 7.6052069664001465, "learning_rate": 3.6652697086659798e-06, "loss": 3.916, "step": 1079250 }, { "epoch": 6.508001591876606, "grad_norm": 4.411352157592773, "learning_rate": 3.6608001514837837e-06, "loss": 4.1606, "step": 1079300 }, { "epoch": 6.50830308365794, "grad_norm": 4.5268330574035645, "learning_rate": 3.6563332874657624e-06, "loss": 4.2251, "step": 1079350 }, { "epoch": 6.508604575439273, "grad_norm": 3.8093228340148926, "learning_rate": 3.651869116694067e-06, "loss": 4.1188, "step": 1079400 }, { "epoch": 6.508906067220607, "grad_norm": 4.296595573425293, "learning_rate": 3.6474076392508656e-06, "loss": 4.4009, "step": 1079450 }, { "epoch": 6.509207559001942, "grad_norm": 5.871171951293945, "learning_rate": 3.642948855218275e-06, "loss": 4.0268, "step": 1079500 }, { "epoch": 6.509509050783276, "grad_norm": 3.364609479904175, "learning_rate": 3.6384927646783467e-06, "loss": 4.2648, "step": 1079550 }, { "epoch": 6.50981054256461, "grad_norm": 1.084891438484192, "learning_rate": 3.634039367713082e-06, "loss": 4.0024, "step": 1079600 }, { "epoch": 6.510112034345944, "grad_norm": 7.383243083953857, "learning_rate": 3.629588664404465e-06, "loss": 4.5435, "step": 1079650 }, { "epoch": 6.5104135261272775, "grad_norm": 5.958620548248291, "learning_rate": 3.625140654834363e-06, "loss": 4.5693, "step": 1079700 }, { "epoch": 6.510715017908612, "grad_norm": 4.425209999084473, "learning_rate": 3.6206953390846784e-06, "loss": 4.2004, "step": 1079750 }, { "epoch": 6.511016509689946, "grad_norm": 4.365593910217285, "learning_rate": 3.6162527172371945e-06, "loss": 4.1662, "step": 1079800 }, { "epoch": 6.51131800147128, "grad_norm": 4.151853084564209, "learning_rate": 3.611812789373697e-06, "loss": 4.1706, "step": 1079850 }, { "epoch": 6.5116194932526135, "grad_norm": 5.143742084503174, "learning_rate": 3.607375555575853e-06, "loss": 3.9773, "step": 1079900 }, { "epoch": 6.511920985033948, "grad_norm": 6.707606315612793, "learning_rate": 3.6029410159253645e-06, "loss": 3.9211, "step": 1079950 }, { "epoch": 6.512222476815282, "grad_norm": 2.166513204574585, "learning_rate": 3.5985091705037995e-06, "loss": 4.1112, "step": 1080000 }, { "epoch": 6.512222476815282, "eval_loss": 4.798316955566406, "eval_runtime": 38.9959, "eval_samples_per_second": 13.13, "eval_steps_per_second": 6.565, "eval_tts_loss": 8.162523889316649, "step": 1080000 }, { "epoch": 6.512523968596616, "grad_norm": 4.582483768463135, "learning_rate": 3.5940800193927587e-06, "loss": 4.2748, "step": 1080050 }, { "epoch": 6.51282546037795, "grad_norm": 6.47785758972168, "learning_rate": 3.589653562673761e-06, "loss": 4.1043, "step": 1080100 }, { "epoch": 6.5131269521592845, "grad_norm": 3.559680938720703, "learning_rate": 3.585229800428241e-06, "loss": 4.1838, "step": 1080150 }, { "epoch": 6.513428443940618, "grad_norm": 5.882527828216553, "learning_rate": 3.5808087327376336e-06, "loss": 3.9364, "step": 1080200 }, { "epoch": 6.513729935721952, "grad_norm": 6.410920143127441, "learning_rate": 3.576390359683273e-06, "loss": 4.3623, "step": 1080250 }, { "epoch": 6.514031427503286, "grad_norm": 5.881185531616211, "learning_rate": 3.5719746813464956e-06, "loss": 3.9113, "step": 1080300 }, { "epoch": 6.51433291928462, "grad_norm": 5.1782546043396, "learning_rate": 3.567561697808552e-06, "loss": 4.6907, "step": 1080350 }, { "epoch": 6.514634411065955, "grad_norm": 3.0000851154327393, "learning_rate": 3.5631514091506773e-06, "loss": 3.9471, "step": 1080400 }, { "epoch": 6.514935902847288, "grad_norm": 4.229923725128174, "learning_rate": 3.5587438154540073e-06, "loss": 3.9505, "step": 1080450 }, { "epoch": 6.515237394628622, "grad_norm": 6.100955486297607, "learning_rate": 3.554338916799676e-06, "loss": 3.9544, "step": 1080500 }, { "epoch": 6.515538886409956, "grad_norm": 5.382606029510498, "learning_rate": 3.5499367132687697e-06, "loss": 4.6171, "step": 1080550 }, { "epoch": 6.515840378191291, "grad_norm": 4.8215837478637695, "learning_rate": 3.5455372049422725e-06, "loss": 4.6916, "step": 1080600 }, { "epoch": 6.516141869972625, "grad_norm": 4.768974781036377, "learning_rate": 3.5411403919011536e-06, "loss": 4.1181, "step": 1080650 }, { "epoch": 6.516443361753959, "grad_norm": 2.213726043701172, "learning_rate": 3.5367462742263487e-06, "loss": 4.0573, "step": 1080700 }, { "epoch": 6.516744853535292, "grad_norm": 5.087025165557861, "learning_rate": 3.5323548519987087e-06, "loss": 4.1827, "step": 1080750 }, { "epoch": 6.5170463453166265, "grad_norm": 6.526362419128418, "learning_rate": 3.5279661252990532e-06, "loss": 4.3191, "step": 1080800 }, { "epoch": 6.517347837097961, "grad_norm": 4.546803951263428, "learning_rate": 3.5235800942081503e-06, "loss": 3.9253, "step": 1080850 }, { "epoch": 6.517649328879295, "grad_norm": 4.114415168762207, "learning_rate": 3.519196758806736e-06, "loss": 4.0994, "step": 1080900 }, { "epoch": 6.517950820660629, "grad_norm": 5.208202362060547, "learning_rate": 3.514816119175462e-06, "loss": 3.9035, "step": 1080950 }, { "epoch": 6.5182523124419625, "grad_norm": 4.377071857452393, "learning_rate": 3.510438175394964e-06, "loss": 4.1333, "step": 1081000 }, { "epoch": 6.518553804223297, "grad_norm": 7.516661167144775, "learning_rate": 3.5060629275457942e-06, "loss": 4.1837, "step": 1081050 }, { "epoch": 6.518855296004631, "grad_norm": 7.127838611602783, "learning_rate": 3.5016903757084548e-06, "loss": 4.1632, "step": 1081100 }, { "epoch": 6.519156787785965, "grad_norm": 5.265097618103027, "learning_rate": 3.497320519963448e-06, "loss": 4.2315, "step": 1081150 }, { "epoch": 6.519458279567299, "grad_norm": 5.102887153625488, "learning_rate": 3.49295336039121e-06, "loss": 4.1506, "step": 1081200 }, { "epoch": 6.5197597713486335, "grad_norm": 4.586232662200928, "learning_rate": 3.488588897072059e-06, "loss": 4.398, "step": 1081250 }, { "epoch": 6.520061263129967, "grad_norm": 4.443787097930908, "learning_rate": 3.4842271300863478e-06, "loss": 4.025, "step": 1081300 }, { "epoch": 6.520362754911301, "grad_norm": 9.459039688110352, "learning_rate": 3.479868059514379e-06, "loss": 4.4123, "step": 1081350 }, { "epoch": 6.520664246692635, "grad_norm": 8.525763511657715, "learning_rate": 3.475511685436322e-06, "loss": 4.0812, "step": 1081400 }, { "epoch": 6.520965738473969, "grad_norm": 4.28407621383667, "learning_rate": 3.471158007932362e-06, "loss": 4.1872, "step": 1081450 }, { "epoch": 6.521267230255303, "grad_norm": 4.7147135734558105, "learning_rate": 3.466807027082652e-06, "loss": 4.0324, "step": 1081500 }, { "epoch": 6.521568722036637, "grad_norm": 6.0381364822387695, "learning_rate": 3.462458742967228e-06, "loss": 4.1421, "step": 1081550 }, { "epoch": 6.521870213817971, "grad_norm": 4.443257808685303, "learning_rate": 3.4581131556661257e-06, "loss": 4.4291, "step": 1081600 }, { "epoch": 6.522171705599305, "grad_norm": 4.067380428314209, "learning_rate": 3.453770265259348e-06, "loss": 3.8369, "step": 1081650 }, { "epoch": 6.52247319738064, "grad_norm": 4.546622276306152, "learning_rate": 3.4494300718267807e-06, "loss": 4.0667, "step": 1081700 }, { "epoch": 6.522774689161974, "grad_norm": 4.584053039550781, "learning_rate": 3.4450925754483105e-06, "loss": 4.4482, "step": 1081750 }, { "epoch": 6.523076180943307, "grad_norm": 7.784656047821045, "learning_rate": 3.4407577762037897e-06, "loss": 4.2297, "step": 1081800 }, { "epoch": 6.523377672724641, "grad_norm": 4.543907642364502, "learning_rate": 3.436425674172938e-06, "loss": 4.1811, "step": 1081850 }, { "epoch": 6.5236791645059755, "grad_norm": 5.449003219604492, "learning_rate": 3.432096269435525e-06, "loss": 3.9421, "step": 1081900 }, { "epoch": 6.52398065628731, "grad_norm": 3.9499990940093994, "learning_rate": 3.4277695620712363e-06, "loss": 4.6162, "step": 1081950 }, { "epoch": 6.524282148068644, "grad_norm": 5.115171432495117, "learning_rate": 3.423445552159659e-06, "loss": 4.6926, "step": 1082000 }, { "epoch": 6.524583639849977, "grad_norm": 6.621134281158447, "learning_rate": 3.41912423978038e-06, "loss": 4.1578, "step": 1082050 }, { "epoch": 6.5248851316313115, "grad_norm": 8.697261810302734, "learning_rate": 3.4148056250129675e-06, "loss": 4.1461, "step": 1082100 }, { "epoch": 6.525186623412646, "grad_norm": 3.9901487827301025, "learning_rate": 3.410489707936842e-06, "loss": 4.3345, "step": 1082150 }, { "epoch": 6.52548811519398, "grad_norm": 1.6401569843292236, "learning_rate": 3.4061764886314736e-06, "loss": 4.0846, "step": 1082200 }, { "epoch": 6.525789606975314, "grad_norm": 5.581540107727051, "learning_rate": 3.4018659671762314e-06, "loss": 4.1581, "step": 1082250 }, { "epoch": 6.526091098756648, "grad_norm": 5.584101676940918, "learning_rate": 3.3975581436504194e-06, "loss": 4.2204, "step": 1082300 }, { "epoch": 6.526392590537982, "grad_norm": 4.95651388168335, "learning_rate": 3.3932530181333406e-06, "loss": 3.8387, "step": 1082350 }, { "epoch": 6.526694082319316, "grad_norm": 3.1465044021606445, "learning_rate": 3.388950590704248e-06, "loss": 4.1217, "step": 1082400 }, { "epoch": 6.52699557410065, "grad_norm": 4.359903812408447, "learning_rate": 3.3846508614422616e-06, "loss": 4.1536, "step": 1082450 }, { "epoch": 6.527297065881984, "grad_norm": 5.652475357055664, "learning_rate": 3.380353830426552e-06, "loss": 4.331, "step": 1082500 }, { "epoch": 6.527598557663318, "grad_norm": 5.133270740509033, "learning_rate": 3.3760594977362056e-06, "loss": 4.2991, "step": 1082550 }, { "epoch": 6.527900049444652, "grad_norm": 8.699448585510254, "learning_rate": 3.371767863450242e-06, "loss": 3.7127, "step": 1082600 }, { "epoch": 6.528201541225986, "grad_norm": 4.680994510650635, "learning_rate": 3.3674789276476323e-06, "loss": 4.254, "step": 1082650 }, { "epoch": 6.52850303300732, "grad_norm": 4.372464179992676, "learning_rate": 3.363192690407329e-06, "loss": 4.1628, "step": 1082700 }, { "epoch": 6.528804524788654, "grad_norm": 3.7676305770874023, "learning_rate": 3.3589091518081867e-06, "loss": 4.3295, "step": 1082750 }, { "epoch": 6.529106016569989, "grad_norm": 6.139403343200684, "learning_rate": 3.3546283119290418e-06, "loss": 4.2188, "step": 1082800 }, { "epoch": 6.529407508351323, "grad_norm": 6.891448497772217, "learning_rate": 3.3503501708487147e-06, "loss": 3.7127, "step": 1082850 }, { "epoch": 6.529709000132656, "grad_norm": 4.831855297088623, "learning_rate": 3.346074728645892e-06, "loss": 3.9847, "step": 1082900 }, { "epoch": 6.53001049191399, "grad_norm": 4.343402862548828, "learning_rate": 3.341801985399295e-06, "loss": 4.3727, "step": 1082950 }, { "epoch": 6.5303119836953245, "grad_norm": 4.127737998962402, "learning_rate": 3.3375319411875267e-06, "loss": 4.3101, "step": 1083000 }, { "epoch": 6.5303119836953245, "eval_loss": 4.796136856079102, "eval_runtime": 38.8734, "eval_samples_per_second": 13.171, "eval_steps_per_second": 6.585, "eval_tts_loss": 8.160615513765881, "step": 1083000 }, { "epoch": 6.530613475476659, "grad_norm": 4.355660438537598, "learning_rate": 3.333264596089208e-06, "loss": 4.3937, "step": 1083050 }, { "epoch": 6.530914967257993, "grad_norm": 3.742063045501709, "learning_rate": 3.3289999501828255e-06, "loss": 4.4187, "step": 1083100 }, { "epoch": 6.531216459039326, "grad_norm": 6.915212154388428, "learning_rate": 3.324738003546901e-06, "loss": 4.215, "step": 1083150 }, { "epoch": 6.5315179508206604, "grad_norm": 3.143228769302368, "learning_rate": 3.3204787562598534e-06, "loss": 3.7742, "step": 1083200 }, { "epoch": 6.531819442601995, "grad_norm": 4.9143877029418945, "learning_rate": 3.3162222084000543e-06, "loss": 4.2637, "step": 1083250 }, { "epoch": 6.532120934383329, "grad_norm": 5.726261138916016, "learning_rate": 3.3119683600458747e-06, "loss": 3.5944, "step": 1083300 }, { "epoch": 6.532422426164663, "grad_norm": 5.616898059844971, "learning_rate": 3.3077172112755847e-06, "loss": 4.2539, "step": 1083350 }, { "epoch": 6.532723917945997, "grad_norm": 4.116008281707764, "learning_rate": 3.303468762167405e-06, "loss": 4.162, "step": 1083400 }, { "epoch": 6.533025409727331, "grad_norm": 4.610593795776367, "learning_rate": 3.299223012799557e-06, "loss": 4.2219, "step": 1083450 }, { "epoch": 6.533326901508665, "grad_norm": 4.577077865600586, "learning_rate": 3.2949799632501437e-06, "loss": 4.3947, "step": 1083500 }, { "epoch": 6.533628393289999, "grad_norm": 3.868016242980957, "learning_rate": 3.290739613597254e-06, "loss": 4.6178, "step": 1083550 }, { "epoch": 6.533929885071333, "grad_norm": 7.450252532958984, "learning_rate": 3.2865019639189416e-06, "loss": 4.3438, "step": 1083600 }, { "epoch": 6.5342313768526665, "grad_norm": 5.240962982177734, "learning_rate": 3.282267014293194e-06, "loss": 4.3124, "step": 1083650 }, { "epoch": 6.534532868634001, "grad_norm": 5.816193580627441, "learning_rate": 3.2780347647979322e-06, "loss": 3.8868, "step": 1083700 }, { "epoch": 6.534834360415335, "grad_norm": 4.6147780418396, "learning_rate": 3.273805215511077e-06, "loss": 4.3282, "step": 1083750 }, { "epoch": 6.535135852196669, "grad_norm": 5.429851531982422, "learning_rate": 3.26957836651045e-06, "loss": 4.2541, "step": 1083800 }, { "epoch": 6.535437343978003, "grad_norm": 6.805815696716309, "learning_rate": 3.2653542178738224e-06, "loss": 4.1677, "step": 1083850 }, { "epoch": 6.535738835759338, "grad_norm": 4.23463773727417, "learning_rate": 3.261132769678948e-06, "loss": 4.4377, "step": 1083900 }, { "epoch": 6.536040327540671, "grad_norm": 2.0236966609954834, "learning_rate": 3.2569140220035316e-06, "loss": 4.1422, "step": 1083950 }, { "epoch": 6.536341819322005, "grad_norm": 5.386285781860352, "learning_rate": 3.252697974925178e-06, "loss": 4.2242, "step": 1084000 }, { "epoch": 6.536643311103339, "grad_norm": 5.238350868225098, "learning_rate": 3.248484628521508e-06, "loss": 4.3907, "step": 1084050 }, { "epoch": 6.5369448028846735, "grad_norm": 4.737112045288086, "learning_rate": 3.2442739828700602e-06, "loss": 4.3103, "step": 1084100 }, { "epoch": 6.537246294666008, "grad_norm": 4.23315954208374, "learning_rate": 3.240066038048306e-06, "loss": 4.0798, "step": 1084150 }, { "epoch": 6.537547786447341, "grad_norm": 2.8566389083862305, "learning_rate": 3.2358607941337157e-06, "loss": 4.2578, "step": 1084200 }, { "epoch": 6.537849278228675, "grad_norm": 4.898245334625244, "learning_rate": 3.2316582512036616e-06, "loss": 4.229, "step": 1084250 }, { "epoch": 6.538150770010009, "grad_norm": 5.175380229949951, "learning_rate": 3.227458409335465e-06, "loss": 4.1926, "step": 1084300 }, { "epoch": 6.538452261791344, "grad_norm": 4.536269664764404, "learning_rate": 3.223261268606464e-06, "loss": 4.3494, "step": 1084350 }, { "epoch": 6.538753753572678, "grad_norm": 5.276246547698975, "learning_rate": 3.2190668290938636e-06, "loss": 4.2058, "step": 1084400 }, { "epoch": 6.539055245354012, "grad_norm": 4.209084510803223, "learning_rate": 3.214875090874869e-06, "loss": 4.2746, "step": 1084450 }, { "epoch": 6.539356737135345, "grad_norm": 6.094232559204102, "learning_rate": 3.2106860540266346e-06, "loss": 4.0184, "step": 1084500 }, { "epoch": 6.53965822891668, "grad_norm": 5.655417442321777, "learning_rate": 3.2064997186262485e-06, "loss": 4.6556, "step": 1084550 }, { "epoch": 6.539959720698014, "grad_norm": 4.280646800994873, "learning_rate": 3.20231608475075e-06, "loss": 3.9251, "step": 1084600 }, { "epoch": 6.540261212479348, "grad_norm": 6.736729145050049, "learning_rate": 3.1981351524771103e-06, "loss": 4.227, "step": 1084650 }, { "epoch": 6.540562704260682, "grad_norm": 4.007170677185059, "learning_rate": 3.193956921882318e-06, "loss": 4.4811, "step": 1084700 }, { "epoch": 6.5408641960420155, "grad_norm": 5.3045268058776855, "learning_rate": 3.189781393043228e-06, "loss": 4.5731, "step": 1084750 }, { "epoch": 6.54116568782335, "grad_norm": 6.651361465454102, "learning_rate": 3.1856085660367125e-06, "loss": 3.9811, "step": 1084800 }, { "epoch": 6.541467179604684, "grad_norm": 4.595302581787109, "learning_rate": 3.18143844093956e-06, "loss": 4.4939, "step": 1084850 }, { "epoch": 6.541768671386018, "grad_norm": 5.679404258728027, "learning_rate": 3.1772710178285086e-06, "loss": 4.6228, "step": 1084900 }, { "epoch": 6.542070163167352, "grad_norm": 4.22878885269165, "learning_rate": 3.173106296780248e-06, "loss": 3.9237, "step": 1084950 }, { "epoch": 6.5423716549486866, "grad_norm": 4.4521894454956055, "learning_rate": 3.168944277871449e-06, "loss": 4.3481, "step": 1085000 }, { "epoch": 6.54267314673002, "grad_norm": 3.174548387527466, "learning_rate": 3.1647849611786847e-06, "loss": 4.0144, "step": 1085050 }, { "epoch": 6.542974638511354, "grad_norm": 4.381572246551514, "learning_rate": 3.160628346778493e-06, "loss": 4.4324, "step": 1085100 }, { "epoch": 6.543276130292688, "grad_norm": 4.81899356842041, "learning_rate": 3.1564744347474134e-06, "loss": 4.395, "step": 1085150 }, { "epoch": 6.5435776220740225, "grad_norm": 5.081292152404785, "learning_rate": 3.1523232251618347e-06, "loss": 3.9552, "step": 1085200 }, { "epoch": 6.543879113855356, "grad_norm": 4.83864688873291, "learning_rate": 3.1481747180981954e-06, "loss": 4.1139, "step": 1085250 }, { "epoch": 6.54418060563669, "grad_norm": 4.421170711517334, "learning_rate": 3.1440289136328345e-06, "loss": 4.1643, "step": 1085300 }, { "epoch": 6.544482097418024, "grad_norm": 5.932126045227051, "learning_rate": 3.139885811842041e-06, "loss": 4.2832, "step": 1085350 }, { "epoch": 6.544783589199358, "grad_norm": 7.280465602874756, "learning_rate": 3.1357454128020876e-06, "loss": 4.3444, "step": 1085400 }, { "epoch": 6.545085080980693, "grad_norm": 5.1137919425964355, "learning_rate": 3.1316077165891464e-06, "loss": 4.0741, "step": 1085450 }, { "epoch": 6.545386572762027, "grad_norm": 4.3407464027404785, "learning_rate": 3.127472723279356e-06, "loss": 4.0192, "step": 1085500 }, { "epoch": 6.54568806454336, "grad_norm": 4.232221603393555, "learning_rate": 3.12334043294884e-06, "loss": 4.0458, "step": 1085550 }, { "epoch": 6.545989556324694, "grad_norm": 5.189199924468994, "learning_rate": 3.119210845673653e-06, "loss": 4.2886, "step": 1085600 }, { "epoch": 6.546291048106029, "grad_norm": 3.132375717163086, "learning_rate": 3.115083961529752e-06, "loss": 4.2315, "step": 1085650 }, { "epoch": 6.546592539887363, "grad_norm": 3.964106798171997, "learning_rate": 3.1109597805931252e-06, "loss": 4.3348, "step": 1085700 }, { "epoch": 6.546894031668697, "grad_norm": 4.306720733642578, "learning_rate": 3.106838302939679e-06, "loss": 3.9435, "step": 1085750 }, { "epoch": 6.54719552345003, "grad_norm": 4.84709358215332, "learning_rate": 3.1027195286452367e-06, "loss": 4.1274, "step": 1085800 }, { "epoch": 6.5474970152313645, "grad_norm": 3.8845417499542236, "learning_rate": 3.098603457785603e-06, "loss": 4.1554, "step": 1085850 }, { "epoch": 6.547798507012699, "grad_norm": 4.066973686218262, "learning_rate": 3.0944900904365523e-06, "loss": 4.2401, "step": 1085900 }, { "epoch": 6.548099998794033, "grad_norm": 5.927350997924805, "learning_rate": 3.090379426673739e-06, "loss": 4.3933, "step": 1085950 }, { "epoch": 6.548401490575367, "grad_norm": 5.957882404327393, "learning_rate": 3.086271466572854e-06, "loss": 4.015, "step": 1086000 }, { "epoch": 6.548401490575367, "eval_loss": 4.795315742492676, "eval_runtime": 39.2506, "eval_samples_per_second": 13.044, "eval_steps_per_second": 6.522, "eval_tts_loss": 8.163740305802653, "step": 1086000 }, { "epoch": 6.548702982356701, "grad_norm": 3.401337146759033, "learning_rate": 3.0821662102094856e-06, "loss": 4.1013, "step": 1086050 }, { "epoch": 6.549004474138035, "grad_norm": 3.818377733230591, "learning_rate": 3.0780636576591746e-06, "loss": 4.0259, "step": 1086100 }, { "epoch": 6.549305965919369, "grad_norm": 6.956812381744385, "learning_rate": 3.073963808997443e-06, "loss": 4.4403, "step": 1086150 }, { "epoch": 6.549607457700703, "grad_norm": 5.1440043449401855, "learning_rate": 3.069866664299747e-06, "loss": 3.9951, "step": 1086200 }, { "epoch": 6.549908949482037, "grad_norm": 4.485376834869385, "learning_rate": 3.0657722236414437e-06, "loss": 4.0911, "step": 1086250 }, { "epoch": 6.5502104412633715, "grad_norm": 4.505449295043945, "learning_rate": 3.061680487097923e-06, "loss": 3.8983, "step": 1086300 }, { "epoch": 6.550511933044705, "grad_norm": 4.086272716522217, "learning_rate": 3.0575914547445067e-06, "loss": 4.2143, "step": 1086350 }, { "epoch": 6.550813424826039, "grad_norm": 4.343625068664551, "learning_rate": 3.0535051266564025e-06, "loss": 4.2989, "step": 1086400 }, { "epoch": 6.551114916607373, "grad_norm": 2.653369665145874, "learning_rate": 3.0494215029088166e-06, "loss": 4.1139, "step": 1086450 }, { "epoch": 6.551416408388707, "grad_norm": 5.735227584838867, "learning_rate": 3.0453405835769385e-06, "loss": 4.0754, "step": 1086500 }, { "epoch": 6.551717900170042, "grad_norm": 4.380020618438721, "learning_rate": 3.0412623687358415e-06, "loss": 4.3442, "step": 1086550 }, { "epoch": 6.552019391951376, "grad_norm": 1.3293428421020508, "learning_rate": 3.0371868584606164e-06, "loss": 4.0956, "step": 1086600 }, { "epoch": 6.552320883732709, "grad_norm": 2.8650028705596924, "learning_rate": 3.0331140528262188e-06, "loss": 4.1763, "step": 1086650 }, { "epoch": 6.552622375514043, "grad_norm": 7.336196422576904, "learning_rate": 3.0290439519076225e-06, "loss": 4.4474, "step": 1086700 }, { "epoch": 6.552923867295378, "grad_norm": 4.766290664672852, "learning_rate": 3.0249765557797345e-06, "loss": 4.2153, "step": 1086750 }, { "epoch": 6.553225359076712, "grad_norm": 5.182925701141357, "learning_rate": 3.0209118645174113e-06, "loss": 4.2096, "step": 1086800 }, { "epoch": 6.553526850858046, "grad_norm": 5.070083141326904, "learning_rate": 3.016849878195443e-06, "loss": 4.2411, "step": 1086850 }, { "epoch": 6.553828342639379, "grad_norm": 2.418729066848755, "learning_rate": 3.0127905968886036e-06, "loss": 3.9657, "step": 1086900 }, { "epoch": 6.5541298344207135, "grad_norm": 5.1527814865112305, "learning_rate": 3.008734020671599e-06, "loss": 4.2593, "step": 1086950 }, { "epoch": 6.554431326202048, "grad_norm": 5.100046157836914, "learning_rate": 3.0046801496190876e-06, "loss": 4.1843, "step": 1087000 }, { "epoch": 6.554732817983382, "grad_norm": 4.720084190368652, "learning_rate": 3.0006289838056417e-06, "loss": 4.4547, "step": 1087050 }, { "epoch": 6.555034309764716, "grad_norm": 5.662113189697266, "learning_rate": 2.9965805233058526e-06, "loss": 4.1507, "step": 1087100 }, { "epoch": 6.55533580154605, "grad_norm": 5.5052080154418945, "learning_rate": 2.9925347681941934e-06, "loss": 4.0413, "step": 1087150 }, { "epoch": 6.555637293327384, "grad_norm": 4.524762153625488, "learning_rate": 2.988491718545155e-06, "loss": 4.26, "step": 1087200 }, { "epoch": 6.555938785108718, "grad_norm": 5.653930187225342, "learning_rate": 2.9844513744331442e-06, "loss": 4.0826, "step": 1087250 }, { "epoch": 6.556240276890052, "grad_norm": 4.450830459594727, "learning_rate": 2.9804137359324853e-06, "loss": 3.9682, "step": 1087300 }, { "epoch": 6.556541768671386, "grad_norm": 3.6607964038848877, "learning_rate": 2.9763788031175185e-06, "loss": 4.1355, "step": 1087350 }, { "epoch": 6.55684326045272, "grad_norm": 8.614477157592773, "learning_rate": 2.9723465760624677e-06, "loss": 4.0237, "step": 1087400 }, { "epoch": 6.557144752234054, "grad_norm": 4.042533874511719, "learning_rate": 2.9683170548415737e-06, "loss": 4.3265, "step": 1087450 }, { "epoch": 6.557446244015388, "grad_norm": 6.106268882751465, "learning_rate": 2.9642902395289602e-06, "loss": 4.0669, "step": 1087500 }, { "epoch": 6.557747735796722, "grad_norm": 6.68553352355957, "learning_rate": 2.9602661301987516e-06, "loss": 4.4145, "step": 1087550 }, { "epoch": 6.558049227578056, "grad_norm": 6.621220588684082, "learning_rate": 2.9562447269250388e-06, "loss": 4.4671, "step": 1087600 }, { "epoch": 6.558350719359391, "grad_norm": 5.268040657043457, "learning_rate": 2.952226029781779e-06, "loss": 4.1826, "step": 1087650 }, { "epoch": 6.558652211140724, "grad_norm": 5.881492614746094, "learning_rate": 2.948210038842946e-06, "loss": 4.5553, "step": 1087700 }, { "epoch": 6.558953702922058, "grad_norm": 3.8749654293060303, "learning_rate": 2.9441967541824653e-06, "loss": 4.0447, "step": 1087750 }, { "epoch": 6.559255194703392, "grad_norm": 5.149021148681641, "learning_rate": 2.9401861758741763e-06, "loss": 3.9812, "step": 1087800 }, { "epoch": 6.559556686484727, "grad_norm": 1.9939379692077637, "learning_rate": 2.936178303991904e-06, "loss": 4.0108, "step": 1087850 }, { "epoch": 6.559858178266061, "grad_norm": 3.1760482788085938, "learning_rate": 2.932173138609389e-06, "loss": 3.7035, "step": 1087900 }, { "epoch": 6.560159670047394, "grad_norm": 6.628845691680908, "learning_rate": 2.928170679800357e-06, "loss": 3.98, "step": 1087950 }, { "epoch": 6.560461161828728, "grad_norm": 6.410694122314453, "learning_rate": 2.9241709276384473e-06, "loss": 4.0386, "step": 1088000 }, { "epoch": 6.5607626536100625, "grad_norm": 1.7321922779083252, "learning_rate": 2.920173882197302e-06, "loss": 4.3894, "step": 1088050 }, { "epoch": 6.561064145391397, "grad_norm": 3.6287031173706055, "learning_rate": 2.9161795435504454e-06, "loss": 4.1237, "step": 1088100 }, { "epoch": 6.561365637172731, "grad_norm": 4.5597243309021, "learning_rate": 2.9121879117714184e-06, "loss": 4.5084, "step": 1088150 }, { "epoch": 6.561667128954065, "grad_norm": 5.452258586883545, "learning_rate": 2.908198986933663e-06, "loss": 4.373, "step": 1088200 }, { "epoch": 6.561968620735398, "grad_norm": 5.300853729248047, "learning_rate": 2.9042127691105866e-06, "loss": 4.4247, "step": 1088250 }, { "epoch": 6.562270112516733, "grad_norm": 5.143680095672607, "learning_rate": 2.9002292583755636e-06, "loss": 4.1244, "step": 1088300 }, { "epoch": 6.562571604298067, "grad_norm": 1.3366378545761108, "learning_rate": 2.8962484548019027e-06, "loss": 3.871, "step": 1088350 }, { "epoch": 6.562873096079401, "grad_norm": 5.209975719451904, "learning_rate": 2.8922703584628447e-06, "loss": 3.9915, "step": 1088400 }, { "epoch": 6.563174587860735, "grad_norm": 3.3379580974578857, "learning_rate": 2.8882949694316317e-06, "loss": 4.2186, "step": 1088450 }, { "epoch": 6.563476079642069, "grad_norm": 4.43959379196167, "learning_rate": 2.884322287781404e-06, "loss": 3.7523, "step": 1088500 }, { "epoch": 6.563777571423403, "grad_norm": 5.608637809753418, "learning_rate": 2.8803523135852715e-06, "loss": 4.2884, "step": 1088550 }, { "epoch": 6.564079063204737, "grad_norm": 7.412516117095947, "learning_rate": 2.876385046916324e-06, "loss": 3.9976, "step": 1088600 }, { "epoch": 6.564380554986071, "grad_norm": 4.128163814544678, "learning_rate": 2.8724204878475544e-06, "loss": 4.0545, "step": 1088650 }, { "epoch": 6.564682046767405, "grad_norm": 4.138880252838135, "learning_rate": 2.8684586364519035e-06, "loss": 4.2489, "step": 1088700 }, { "epoch": 6.56498353854874, "grad_norm": 2.2640161514282227, "learning_rate": 2.864499492802297e-06, "loss": 4.2151, "step": 1088750 }, { "epoch": 6.565285030330073, "grad_norm": 5.877399444580078, "learning_rate": 2.860543056971626e-06, "loss": 4.3887, "step": 1088800 }, { "epoch": 6.565586522111407, "grad_norm": 6.301597595214844, "learning_rate": 2.856589329032666e-06, "loss": 4.3671, "step": 1088850 }, { "epoch": 6.565888013892741, "grad_norm": 7.2813239097595215, "learning_rate": 2.8526383090582085e-06, "loss": 4.4976, "step": 1088900 }, { "epoch": 6.5661895056740756, "grad_norm": 10.38479995727539, "learning_rate": 2.8486899971209453e-06, "loss": 4.1531, "step": 1088950 }, { "epoch": 6.566490997455409, "grad_norm": 2.1672701835632324, "learning_rate": 2.844744393293552e-06, "loss": 4.0514, "step": 1089000 }, { "epoch": 6.566490997455409, "eval_loss": 4.79559326171875, "eval_runtime": 38.9537, "eval_samples_per_second": 13.144, "eval_steps_per_second": 6.572, "eval_tts_loss": 8.155340595660773, "step": 1089000 }, { "epoch": 6.566792489236743, "grad_norm": 2.8678884506225586, "learning_rate": 2.8408014976486204e-06, "loss": 4.165, "step": 1089050 }, { "epoch": 6.567093981018077, "grad_norm": 1.5543054342269897, "learning_rate": 2.8368613102587588e-06, "loss": 4.0828, "step": 1089100 }, { "epoch": 6.5673954727994115, "grad_norm": 3.8912978172302246, "learning_rate": 2.8329238311964263e-06, "loss": 4.1168, "step": 1089150 }, { "epoch": 6.567696964580746, "grad_norm": 4.540226459503174, "learning_rate": 2.828989060534115e-06, "loss": 3.7531, "step": 1089200 }, { "epoch": 6.56799845636208, "grad_norm": 5.265042304992676, "learning_rate": 2.825056998344233e-06, "loss": 4.608, "step": 1089250 }, { "epoch": 6.568299948143413, "grad_norm": 5.989539623260498, "learning_rate": 2.821127644699156e-06, "loss": 4.2386, "step": 1089300 }, { "epoch": 6.568601439924747, "grad_norm": 6.574896812438965, "learning_rate": 2.817200999671193e-06, "loss": 4.4273, "step": 1089350 }, { "epoch": 6.568902931706082, "grad_norm": 5.227540969848633, "learning_rate": 2.813277063332603e-06, "loss": 4.3346, "step": 1089400 }, { "epoch": 6.569204423487416, "grad_norm": 6.084560871124268, "learning_rate": 2.8093558357555946e-06, "loss": 4.398, "step": 1089450 }, { "epoch": 6.56950591526875, "grad_norm": 4.351879596710205, "learning_rate": 2.805437317012327e-06, "loss": 3.7356, "step": 1089500 }, { "epoch": 6.569807407050083, "grad_norm": 7.184326171875, "learning_rate": 2.8015215071749584e-06, "loss": 4.347, "step": 1089550 }, { "epoch": 6.570108898831418, "grad_norm": 7.344510555267334, "learning_rate": 2.797608406315499e-06, "loss": 4.0845, "step": 1089600 }, { "epoch": 6.570410390612752, "grad_norm": 5.884844779968262, "learning_rate": 2.7936980145059906e-06, "loss": 4.1703, "step": 1089650 }, { "epoch": 6.570711882394086, "grad_norm": 6.499780654907227, "learning_rate": 2.789790331818409e-06, "loss": 4.3377, "step": 1089700 }, { "epoch": 6.57101337417542, "grad_norm": 7.596011161804199, "learning_rate": 2.7858853583246465e-06, "loss": 4.0359, "step": 1089750 }, { "epoch": 6.571314865956754, "grad_norm": 4.250834941864014, "learning_rate": 2.781983094096579e-06, "loss": 4.0364, "step": 1089800 }, { "epoch": 6.571616357738088, "grad_norm": 5.094759941101074, "learning_rate": 2.778083539206033e-06, "loss": 4.13, "step": 1089850 }, { "epoch": 6.571917849519422, "grad_norm": 2.633791446685791, "learning_rate": 2.77418669372475e-06, "loss": 4.0467, "step": 1089900 }, { "epoch": 6.572219341300756, "grad_norm": 5.0501837730407715, "learning_rate": 2.770292557724457e-06, "loss": 4.0756, "step": 1089950 }, { "epoch": 6.57252083308209, "grad_norm": 1.9781173467636108, "learning_rate": 2.7664011312768297e-06, "loss": 3.9984, "step": 1090000 }, { "epoch": 6.5728223248634245, "grad_norm": 4.30168342590332, "learning_rate": 2.7625124144534605e-06, "loss": 4.4204, "step": 1090050 }, { "epoch": 6.573123816644758, "grad_norm": 4.72907018661499, "learning_rate": 2.7586264073259424e-06, "loss": 4.0559, "step": 1090100 }, { "epoch": 6.573425308426092, "grad_norm": 4.929039001464844, "learning_rate": 2.7547431099657845e-06, "loss": 4.1471, "step": 1090150 }, { "epoch": 6.573726800207426, "grad_norm": 5.259273052215576, "learning_rate": 2.750862522444447e-06, "loss": 4.2039, "step": 1090200 }, { "epoch": 6.5740282919887605, "grad_norm": 6.6173014640808105, "learning_rate": 2.746984644833339e-06, "loss": 4.4364, "step": 1090250 }, { "epoch": 6.574329783770095, "grad_norm": 4.493743419647217, "learning_rate": 2.743109477203853e-06, "loss": 4.5055, "step": 1090300 }, { "epoch": 6.574631275551429, "grad_norm": 3.916578769683838, "learning_rate": 2.7392370196272826e-06, "loss": 3.9651, "step": 1090350 }, { "epoch": 6.574932767332762, "grad_norm": 5.590425968170166, "learning_rate": 2.7353672721748877e-06, "loss": 3.9132, "step": 1090400 }, { "epoch": 6.575234259114096, "grad_norm": 4.221266269683838, "learning_rate": 2.7315002349179273e-06, "loss": 4.5416, "step": 1090450 }, { "epoch": 6.575535750895431, "grad_norm": 5.093245029449463, "learning_rate": 2.727635907927511e-06, "loss": 4.3066, "step": 1090500 }, { "epoch": 6.575837242676765, "grad_norm": 4.301716327667236, "learning_rate": 2.723774291274816e-06, "loss": 4.055, "step": 1090550 }, { "epoch": 6.576138734458099, "grad_norm": 8.12581729888916, "learning_rate": 2.719915385030885e-06, "loss": 4.2472, "step": 1090600 }, { "epoch": 6.576440226239432, "grad_norm": 7.476332187652588, "learning_rate": 2.716059189266695e-06, "loss": 4.1107, "step": 1090650 }, { "epoch": 6.576741718020767, "grad_norm": 3.2690155506134033, "learning_rate": 2.712205704053272e-06, "loss": 4.3511, "step": 1090700 }, { "epoch": 6.577043209802101, "grad_norm": 5.455105781555176, "learning_rate": 2.7083549294615094e-06, "loss": 4.3745, "step": 1090750 }, { "epoch": 6.577344701583435, "grad_norm": 6.418973922729492, "learning_rate": 2.7045068655622837e-06, "loss": 4.0858, "step": 1090800 }, { "epoch": 6.577646193364769, "grad_norm": 5.084366798400879, "learning_rate": 2.7006615124263887e-06, "loss": 3.9076, "step": 1090850 }, { "epoch": 6.577947685146103, "grad_norm": 4.552306652069092, "learning_rate": 2.6968188701246174e-06, "loss": 4.5546, "step": 1090900 }, { "epoch": 6.578249176927437, "grad_norm": 4.228947162628174, "learning_rate": 2.692978938727697e-06, "loss": 4.3353, "step": 1090950 }, { "epoch": 6.578550668708771, "grad_norm": 5.928131580352783, "learning_rate": 2.689141718306287e-06, "loss": 4.5066, "step": 1091000 }, { "epoch": 6.578852160490105, "grad_norm": 6.6604204177856445, "learning_rate": 2.6853072089309814e-06, "loss": 4.1497, "step": 1091050 }, { "epoch": 6.579153652271439, "grad_norm": 5.4374680519104, "learning_rate": 2.6814754106723735e-06, "loss": 4.2164, "step": 1091100 }, { "epoch": 6.579455144052773, "grad_norm": 4.987297058105469, "learning_rate": 2.6776463236009737e-06, "loss": 4.0934, "step": 1091150 }, { "epoch": 6.579756635834107, "grad_norm": 5.35908317565918, "learning_rate": 2.6738199477872423e-06, "loss": 4.2382, "step": 1091200 }, { "epoch": 6.580058127615441, "grad_norm": 5.0301737785339355, "learning_rate": 2.66999628330164e-06, "loss": 4.3044, "step": 1091250 }, { "epoch": 6.580359619396775, "grad_norm": 6.088274955749512, "learning_rate": 2.666175330214476e-06, "loss": 4.4146, "step": 1091300 }, { "epoch": 6.5806611111781095, "grad_norm": 4.660872459411621, "learning_rate": 2.662357088596112e-06, "loss": 3.8687, "step": 1091350 }, { "epoch": 6.580962602959444, "grad_norm": 3.7075140476226807, "learning_rate": 2.658541558516808e-06, "loss": 3.9254, "step": 1091400 }, { "epoch": 6.581264094740777, "grad_norm": 7.762622833251953, "learning_rate": 2.6547287400467577e-06, "loss": 4.1367, "step": 1091450 }, { "epoch": 6.581565586522111, "grad_norm": 4.4359660148620605, "learning_rate": 2.6509186332561716e-06, "loss": 4.0633, "step": 1091500 }, { "epoch": 6.581867078303445, "grad_norm": 6.735803604125977, "learning_rate": 2.6471112382151438e-06, "loss": 3.9319, "step": 1091550 }, { "epoch": 6.58216857008478, "grad_norm": 1.9664541482925415, "learning_rate": 2.643306554993735e-06, "loss": 4.2447, "step": 1091600 }, { "epoch": 6.582470061866114, "grad_norm": 4.886983394622803, "learning_rate": 2.639504583661989e-06, "loss": 4.1697, "step": 1091650 }, { "epoch": 6.582771553647447, "grad_norm": 4.521001815795898, "learning_rate": 2.635705324289883e-06, "loss": 4.1994, "step": 1091700 }, { "epoch": 6.583073045428781, "grad_norm": 5.735030651092529, "learning_rate": 2.6319087769473113e-06, "loss": 4.2323, "step": 1091750 }, { "epoch": 6.583374537210116, "grad_norm": 4.432435035705566, "learning_rate": 2.6281149417041514e-06, "loss": 4.2908, "step": 1091800 }, { "epoch": 6.58367602899145, "grad_norm": 7.144222736358643, "learning_rate": 2.624323818630231e-06, "loss": 4.3423, "step": 1091850 }, { "epoch": 6.583977520772784, "grad_norm": 7.556519985198975, "learning_rate": 2.620535407795293e-06, "loss": 4.3391, "step": 1091900 }, { "epoch": 6.584279012554118, "grad_norm": 6.4009528160095215, "learning_rate": 2.616749709269084e-06, "loss": 4.0924, "step": 1091950 }, { "epoch": 6.5845805043354515, "grad_norm": 6.912029266357422, "learning_rate": 2.612966723121296e-06, "loss": 3.9361, "step": 1092000 }, { "epoch": 6.5845805043354515, "eval_loss": 4.793539047241211, "eval_runtime": 39.0989, "eval_samples_per_second": 13.095, "eval_steps_per_second": 6.548, "eval_tts_loss": 8.158407265771775, "step": 1092000 }, { "epoch": 6.584881996116786, "grad_norm": 2.3415863513946533, "learning_rate": 2.609186449421491e-06, "loss": 3.9081, "step": 1092050 }, { "epoch": 6.58518348789812, "grad_norm": 3.923076629638672, "learning_rate": 2.6054088882392797e-06, "loss": 4.3807, "step": 1092100 }, { "epoch": 6.585484979679454, "grad_norm": 1.4953746795654297, "learning_rate": 2.6016340396441903e-06, "loss": 4.0279, "step": 1092150 }, { "epoch": 6.585786471460788, "grad_norm": 5.537300109863281, "learning_rate": 2.5978619037056836e-06, "loss": 4.459, "step": 1092200 }, { "epoch": 6.586087963242122, "grad_norm": 5.714343070983887, "learning_rate": 2.594092480493154e-06, "loss": 4.2064, "step": 1092250 }, { "epoch": 6.586389455023456, "grad_norm": 3.8174705505371094, "learning_rate": 2.590325770075996e-06, "loss": 4.205, "step": 1092300 }, { "epoch": 6.58669094680479, "grad_norm": 4.671273708343506, "learning_rate": 2.586561772523521e-06, "loss": 4.0049, "step": 1092350 }, { "epoch": 6.586992438586124, "grad_norm": 4.297212600708008, "learning_rate": 2.5828004879050067e-06, "loss": 3.8046, "step": 1092400 }, { "epoch": 6.5872939303674585, "grad_norm": 6.340620517730713, "learning_rate": 2.579041916289698e-06, "loss": 4.1361, "step": 1092450 }, { "epoch": 6.587595422148793, "grad_norm": 3.5747087001800537, "learning_rate": 2.575286057746706e-06, "loss": 4.2394, "step": 1092500 }, { "epoch": 6.587896913930126, "grad_norm": 7.709863662719727, "learning_rate": 2.5715329123452255e-06, "loss": 4.2468, "step": 1092550 }, { "epoch": 6.58819840571146, "grad_norm": 4.1273016929626465, "learning_rate": 2.5677824801542678e-06, "loss": 4.1994, "step": 1092600 }, { "epoch": 6.588499897492794, "grad_norm": 4.896385669708252, "learning_rate": 2.564034761242878e-06, "loss": 4.2005, "step": 1092650 }, { "epoch": 6.588801389274129, "grad_norm": 8.533053398132324, "learning_rate": 2.560289755680017e-06, "loss": 4.4487, "step": 1092700 }, { "epoch": 6.589102881055462, "grad_norm": 8.732133865356445, "learning_rate": 2.556547463534647e-06, "loss": 3.9717, "step": 1092750 }, { "epoch": 6.589404372836796, "grad_norm": 5.237840175628662, "learning_rate": 2.5528078848755794e-06, "loss": 4.1332, "step": 1092800 }, { "epoch": 6.58970586461813, "grad_norm": 3.710275173187256, "learning_rate": 2.549071019771659e-06, "loss": 4.1758, "step": 1092850 }, { "epoch": 6.5900073563994646, "grad_norm": 3.3855252265930176, "learning_rate": 2.545336868291681e-06, "loss": 4.0807, "step": 1092900 }, { "epoch": 6.590308848180799, "grad_norm": 5.076731204986572, "learning_rate": 2.5416054305043564e-06, "loss": 4.544, "step": 1092950 }, { "epoch": 6.590610339962133, "grad_norm": 5.651259899139404, "learning_rate": 2.5378767064783312e-06, "loss": 4.1333, "step": 1093000 }, { "epoch": 6.590911831743466, "grad_norm": 7.601364612579346, "learning_rate": 2.53415069628225e-06, "loss": 3.9505, "step": 1093050 }, { "epoch": 6.5912133235248005, "grad_norm": 4.108497142791748, "learning_rate": 2.530427399984675e-06, "loss": 3.8589, "step": 1093100 }, { "epoch": 6.591514815306135, "grad_norm": 3.877113103866577, "learning_rate": 2.526706817654134e-06, "loss": 4.0833, "step": 1093150 }, { "epoch": 6.591816307087469, "grad_norm": 8.251444816589355, "learning_rate": 2.522988949359106e-06, "loss": 4.5395, "step": 1093200 }, { "epoch": 6.592117798868803, "grad_norm": 4.325554370880127, "learning_rate": 2.519273795168003e-06, "loss": 4.2168, "step": 1093250 }, { "epoch": 6.592419290650136, "grad_norm": 5.926553249359131, "learning_rate": 2.515561355149204e-06, "loss": 4.2289, "step": 1093300 }, { "epoch": 6.592720782431471, "grad_norm": 4.207643032073975, "learning_rate": 2.51185162937102e-06, "loss": 3.954, "step": 1093350 }, { "epoch": 6.593022274212805, "grad_norm": 1.9283170700073242, "learning_rate": 2.508144617901747e-06, "loss": 4.0678, "step": 1093400 }, { "epoch": 6.593323765994139, "grad_norm": 3.0470244884490967, "learning_rate": 2.5044403208095643e-06, "loss": 4.0765, "step": 1093450 }, { "epoch": 6.593625257775473, "grad_norm": 2.8545982837677, "learning_rate": 2.500738738162683e-06, "loss": 4.4911, "step": 1093500 }, { "epoch": 6.5939267495568075, "grad_norm": 4.454176902770996, "learning_rate": 2.4970398700291993e-06, "loss": 4.2673, "step": 1093550 }, { "epoch": 6.594228241338141, "grad_norm": 4.28707218170166, "learning_rate": 2.4933437164772087e-06, "loss": 4.5333, "step": 1093600 }, { "epoch": 6.594529733119475, "grad_norm": 4.176891803741455, "learning_rate": 2.4896502775747394e-06, "loss": 4.5611, "step": 1093650 }, { "epoch": 6.594831224900809, "grad_norm": 4.466330051422119, "learning_rate": 2.4859595533897213e-06, "loss": 4.1285, "step": 1093700 }, { "epoch": 6.595132716682143, "grad_norm": 4.555290222167969, "learning_rate": 2.4822715439901164e-06, "loss": 4.0676, "step": 1093750 }, { "epoch": 6.595434208463478, "grad_norm": 2.579577684402466, "learning_rate": 2.4785862494438036e-06, "loss": 4.3147, "step": 1093800 }, { "epoch": 6.595735700244811, "grad_norm": 4.185835361480713, "learning_rate": 2.474903669818562e-06, "loss": 3.9259, "step": 1093850 }, { "epoch": 6.596037192026145, "grad_norm": 3.8365814685821533, "learning_rate": 2.4712238051821875e-06, "loss": 4.2716, "step": 1093900 }, { "epoch": 6.596338683807479, "grad_norm": 6.7337164878845215, "learning_rate": 2.4675466556024094e-06, "loss": 4.4218, "step": 1093950 }, { "epoch": 6.5966401755888135, "grad_norm": 8.709383010864258, "learning_rate": 2.4638722211468897e-06, "loss": 4.2369, "step": 1094000 }, { "epoch": 6.596941667370148, "grad_norm": 4.45875358581543, "learning_rate": 2.460200501883258e-06, "loss": 4.5756, "step": 1094050 }, { "epoch": 6.597243159151482, "grad_norm": 6.11183500289917, "learning_rate": 2.456531497879094e-06, "loss": 4.18, "step": 1094100 }, { "epoch": 6.597544650932815, "grad_norm": 6.281890869140625, "learning_rate": 2.4528652092019095e-06, "loss": 3.985, "step": 1094150 }, { "epoch": 6.5978461427141495, "grad_norm": 3.9925806522369385, "learning_rate": 2.449201635919185e-06, "loss": 4.1349, "step": 1094200 }, { "epoch": 6.598147634495484, "grad_norm": 5.93960428237915, "learning_rate": 2.4455407780983317e-06, "loss": 4.0697, "step": 1094250 }, { "epoch": 6.598449126276818, "grad_norm": 4.939383029937744, "learning_rate": 2.44188263580673e-06, "loss": 4.4967, "step": 1094300 }, { "epoch": 6.598750618058152, "grad_norm": 4.399597644805908, "learning_rate": 2.438227209111693e-06, "loss": 4.4014, "step": 1094350 }, { "epoch": 6.599052109839485, "grad_norm": 6.700409412384033, "learning_rate": 2.4345744980804994e-06, "loss": 3.8385, "step": 1094400 }, { "epoch": 6.59935360162082, "grad_norm": 5.054574966430664, "learning_rate": 2.430924502780396e-06, "loss": 4.0229, "step": 1094450 }, { "epoch": 6.599655093402154, "grad_norm": 6.670895576477051, "learning_rate": 2.427277223278512e-06, "loss": 4.4993, "step": 1094500 }, { "epoch": 6.599956585183488, "grad_norm": 4.154286861419678, "learning_rate": 2.4236326596420107e-06, "loss": 3.9762, "step": 1094550 }, { "epoch": 6.600258076964822, "grad_norm": 5.208902835845947, "learning_rate": 2.419990811937955e-06, "loss": 4.4293, "step": 1094600 }, { "epoch": 6.6005595687461565, "grad_norm": 4.659842014312744, "learning_rate": 2.4163516802333414e-06, "loss": 3.8528, "step": 1094650 }, { "epoch": 6.60086106052749, "grad_norm": 3.30326771736145, "learning_rate": 2.4127152645951665e-06, "loss": 4.0042, "step": 1094700 }, { "epoch": 6.601162552308824, "grad_norm": 4.966635704040527, "learning_rate": 2.4090815650903592e-06, "loss": 3.9674, "step": 1094750 }, { "epoch": 6.601464044090158, "grad_norm": 2.334949254989624, "learning_rate": 2.4054505817857673e-06, "loss": 3.9568, "step": 1094800 }, { "epoch": 6.601765535871492, "grad_norm": 3.986088991165161, "learning_rate": 2.4018223147482196e-06, "loss": 3.6509, "step": 1094850 }, { "epoch": 6.602067027652826, "grad_norm": 4.138854026794434, "learning_rate": 2.3981967640445297e-06, "loss": 4.3055, "step": 1094900 }, { "epoch": 6.60236851943416, "grad_norm": 6.61519193649292, "learning_rate": 2.394573929741378e-06, "loss": 4.0934, "step": 1094950 }, { "epoch": 6.602670011215494, "grad_norm": 5.17799711227417, "learning_rate": 2.3909538119054273e-06, "loss": 4.0537, "step": 1095000 }, { "epoch": 6.602670011215494, "eval_loss": 4.792328834533691, "eval_runtime": 39.1577, "eval_samples_per_second": 13.075, "eval_steps_per_second": 6.538, "eval_tts_loss": 8.169718609927335, "step": 1095000 }, { "epoch": 6.602971502996828, "grad_norm": 4.041853427886963, "learning_rate": 2.387336410603358e-06, "loss": 4.1298, "step": 1095050 }, { "epoch": 6.6032729947781625, "grad_norm": 5.5088629722595215, "learning_rate": 2.3837217259016827e-06, "loss": 4.6227, "step": 1095100 }, { "epoch": 6.603574486559497, "grad_norm": 5.235195636749268, "learning_rate": 2.3801097578669492e-06, "loss": 4.2349, "step": 1095150 }, { "epoch": 6.60387597834083, "grad_norm": 8.014405250549316, "learning_rate": 2.376500506565637e-06, "loss": 4.1704, "step": 1095200 }, { "epoch": 6.604177470122164, "grad_norm": 4.78262996673584, "learning_rate": 2.3728939720641605e-06, "loss": 4.2936, "step": 1095250 }, { "epoch": 6.6044789619034985, "grad_norm": 3.736828088760376, "learning_rate": 2.3692901544288824e-06, "loss": 4.2719, "step": 1095300 }, { "epoch": 6.604780453684833, "grad_norm": 8.458244323730469, "learning_rate": 2.3656890537261665e-06, "loss": 4.353, "step": 1095350 }, { "epoch": 6.605081945466167, "grad_norm": 4.782772064208984, "learning_rate": 2.362090670022243e-06, "loss": 4.3103, "step": 1095400 }, { "epoch": 6.6053834372475, "grad_norm": 4.640180587768555, "learning_rate": 2.358495003383359e-06, "loss": 4.2303, "step": 1095450 }, { "epoch": 6.605684929028834, "grad_norm": 6.120372772216797, "learning_rate": 2.354902053875679e-06, "loss": 4.1433, "step": 1095500 }, { "epoch": 6.605986420810169, "grad_norm": 4.913968086242676, "learning_rate": 2.3513118215652995e-06, "loss": 4.348, "step": 1095550 }, { "epoch": 6.606287912591503, "grad_norm": 5.945783615112305, "learning_rate": 2.3477243065183336e-06, "loss": 4.4141, "step": 1095600 }, { "epoch": 6.606589404372837, "grad_norm": 3.8844199180603027, "learning_rate": 2.344139508800813e-06, "loss": 4.175, "step": 1095650 }, { "epoch": 6.606890896154171, "grad_norm": 4.279473304748535, "learning_rate": 2.340557428478668e-06, "loss": 4.2402, "step": 1095700 }, { "epoch": 6.607192387935505, "grad_norm": 5.348222732543945, "learning_rate": 2.336978065617845e-06, "loss": 4.216, "step": 1095750 }, { "epoch": 6.607493879716839, "grad_norm": 2.1728997230529785, "learning_rate": 2.3334014202842253e-06, "loss": 3.7299, "step": 1095800 }, { "epoch": 6.607795371498173, "grad_norm": 6.709888935089111, "learning_rate": 2.3298274925436066e-06, "loss": 4.1833, "step": 1095850 }, { "epoch": 6.608096863279507, "grad_norm": 7.562806129455566, "learning_rate": 2.326256282461769e-06, "loss": 4.3549, "step": 1095900 }, { "epoch": 6.608398355060841, "grad_norm": 6.0655999183654785, "learning_rate": 2.3226877901044605e-06, "loss": 4.1465, "step": 1095950 }, { "epoch": 6.608699846842175, "grad_norm": 5.122233867645264, "learning_rate": 2.319122015537328e-06, "loss": 4.2815, "step": 1096000 }, { "epoch": 6.609001338623509, "grad_norm": 3.8405649662017822, "learning_rate": 2.3155589588260027e-06, "loss": 4.1482, "step": 1096050 }, { "epoch": 6.609302830404843, "grad_norm": 4.8060221672058105, "learning_rate": 2.311998620036065e-06, "loss": 4.2684, "step": 1096100 }, { "epoch": 6.609604322186177, "grad_norm": 1.2372928857803345, "learning_rate": 2.3084409992330297e-06, "loss": 4.2911, "step": 1096150 }, { "epoch": 6.6099058139675115, "grad_norm": 3.0135786533355713, "learning_rate": 2.3048860964823612e-06, "loss": 3.9766, "step": 1096200 }, { "epoch": 6.610207305748846, "grad_norm": 5.656800270080566, "learning_rate": 2.3013339118495065e-06, "loss": 4.1647, "step": 1096250 }, { "epoch": 6.610508797530179, "grad_norm": 4.987633228302002, "learning_rate": 2.2977844453997975e-06, "loss": 4.0856, "step": 1096300 }, { "epoch": 6.610810289311513, "grad_norm": 4.40760612487793, "learning_rate": 2.2942376971985977e-06, "loss": 4.2262, "step": 1096350 }, { "epoch": 6.6111117810928475, "grad_norm": 4.547379016876221, "learning_rate": 2.290693667311172e-06, "loss": 4.2203, "step": 1096400 }, { "epoch": 6.611413272874182, "grad_norm": 4.559544086456299, "learning_rate": 2.2871523558027348e-06, "loss": 4.3731, "step": 1096450 }, { "epoch": 6.611714764655515, "grad_norm": 7.4566473960876465, "learning_rate": 2.283613762738451e-06, "loss": 4.5441, "step": 1096500 }, { "epoch": 6.612016256436849, "grad_norm": 4.5233964920043945, "learning_rate": 2.280077888183468e-06, "loss": 4.0336, "step": 1096550 }, { "epoch": 6.612317748218183, "grad_norm": 5.484185695648193, "learning_rate": 2.2765447322028506e-06, "loss": 3.9297, "step": 1096600 }, { "epoch": 6.612619239999518, "grad_norm": 5.1305131912231445, "learning_rate": 2.273014294861597e-06, "loss": 4.3495, "step": 1096650 }, { "epoch": 6.612920731780852, "grad_norm": 4.734964847564697, "learning_rate": 2.269486576224705e-06, "loss": 4.1623, "step": 1096700 }, { "epoch": 6.613222223562186, "grad_norm": 4.625046253204346, "learning_rate": 2.2659615763570726e-06, "loss": 4.368, "step": 1096750 }, { "epoch": 6.613523715343519, "grad_norm": 4.182204246520996, "learning_rate": 2.262439295323598e-06, "loss": 4.0899, "step": 1096800 }, { "epoch": 6.6138252071248536, "grad_norm": 5.274919033050537, "learning_rate": 2.258919733189113e-06, "loss": 4.069, "step": 1096850 }, { "epoch": 6.614126698906188, "grad_norm": 2.686321973800659, "learning_rate": 2.2554028900183486e-06, "loss": 3.9257, "step": 1096900 }, { "epoch": 6.614428190687522, "grad_norm": 7.371222972869873, "learning_rate": 2.25188876587607e-06, "loss": 4.1012, "step": 1096950 }, { "epoch": 6.614729682468856, "grad_norm": 2.358110189437866, "learning_rate": 2.2483773608269417e-06, "loss": 4.2386, "step": 1097000 }, { "epoch": 6.6150311742501895, "grad_norm": 5.426910877227783, "learning_rate": 2.244868674935546e-06, "loss": 4.1299, "step": 1097050 }, { "epoch": 6.615332666031524, "grad_norm": 4.173310279846191, "learning_rate": 2.2413627082664974e-06, "loss": 4.2075, "step": 1097100 }, { "epoch": 6.615634157812858, "grad_norm": 5.128707408905029, "learning_rate": 2.2378594608843114e-06, "loss": 4.3075, "step": 1097150 }, { "epoch": 6.615935649594192, "grad_norm": 6.911615371704102, "learning_rate": 2.2343589328534364e-06, "loss": 4.3633, "step": 1097200 }, { "epoch": 6.616237141375526, "grad_norm": 5.107937812805176, "learning_rate": 2.23086112423832e-06, "loss": 4.1006, "step": 1097250 }, { "epoch": 6.6165386331568605, "grad_norm": 4.780616283416748, "learning_rate": 2.2273660351033285e-06, "loss": 4.3362, "step": 1097300 }, { "epoch": 6.616840124938194, "grad_norm": 5.777379989624023, "learning_rate": 2.223873665512793e-06, "loss": 4.2171, "step": 1097350 }, { "epoch": 6.617141616719528, "grad_norm": 2.127368211746216, "learning_rate": 2.2203840155309627e-06, "loss": 3.906, "step": 1097400 }, { "epoch": 6.617443108500862, "grad_norm": 4.405691623687744, "learning_rate": 2.216897085222069e-06, "loss": 4.4884, "step": 1097450 }, { "epoch": 6.6177446002821965, "grad_norm": 6.22355318069458, "learning_rate": 2.2134128746502945e-06, "loss": 4.4266, "step": 1097500 }, { "epoch": 6.618046092063531, "grad_norm": 5.386502742767334, "learning_rate": 2.209931383879737e-06, "loss": 4.1277, "step": 1097550 }, { "epoch": 6.618347583844864, "grad_norm": 5.9432830810546875, "learning_rate": 2.206452612974496e-06, "loss": 4.0714, "step": 1097600 }, { "epoch": 6.618649075626198, "grad_norm": 5.363640308380127, "learning_rate": 2.2029765619985874e-06, "loss": 4.167, "step": 1097650 }, { "epoch": 6.618950567407532, "grad_norm": 5.297013759613037, "learning_rate": 2.1995032310159753e-06, "loss": 4.0294, "step": 1097700 }, { "epoch": 6.619252059188867, "grad_norm": 4.375049114227295, "learning_rate": 2.196032620090593e-06, "loss": 4.1989, "step": 1097750 }, { "epoch": 6.619553550970201, "grad_norm": 5.109136581420898, "learning_rate": 2.192564729286289e-06, "loss": 4.1085, "step": 1097800 }, { "epoch": 6.619855042751535, "grad_norm": 4.355531692504883, "learning_rate": 2.189099558666896e-06, "loss": 3.885, "step": 1097850 }, { "epoch": 6.620156534532868, "grad_norm": 5.0076518058776855, "learning_rate": 2.1856371082961954e-06, "loss": 4.5037, "step": 1097900 }, { "epoch": 6.6204580263142025, "grad_norm": 10.690961837768555, "learning_rate": 2.1821773782379035e-06, "loss": 4.072, "step": 1097950 }, { "epoch": 6.620759518095537, "grad_norm": 4.871262073516846, "learning_rate": 2.17872036855567e-06, "loss": 4.2981, "step": 1098000 }, { "epoch": 6.620759518095537, "eval_loss": 4.791132926940918, "eval_runtime": 39.0701, "eval_samples_per_second": 13.105, "eval_steps_per_second": 6.552, "eval_tts_loss": 8.158146515645418, "step": 1098000 }, { "epoch": 6.621061009876871, "grad_norm": 4.123307704925537, "learning_rate": 2.175266079313159e-06, "loss": 4.2461, "step": 1098050 }, { "epoch": 6.621362501658205, "grad_norm": 5.30792236328125, "learning_rate": 2.171814510573905e-06, "loss": 4.0887, "step": 1098100 }, { "epoch": 6.6216639934395385, "grad_norm": 4.1224260330200195, "learning_rate": 2.1683656624014555e-06, "loss": 4.0592, "step": 1098150 }, { "epoch": 6.621965485220873, "grad_norm": 4.277826309204102, "learning_rate": 2.164919534859261e-06, "loss": 4.4158, "step": 1098200 }, { "epoch": 6.622266977002207, "grad_norm": 7.076958656311035, "learning_rate": 2.1614761280107694e-06, "loss": 4.4951, "step": 1098250 }, { "epoch": 6.622568468783541, "grad_norm": 4.314687728881836, "learning_rate": 2.158035441919298e-06, "loss": 4.2275, "step": 1098300 }, { "epoch": 6.622869960564875, "grad_norm": 4.103309631347656, "learning_rate": 2.154597476648212e-06, "loss": 4.0138, "step": 1098350 }, { "epoch": 6.6231714523462095, "grad_norm": 7.082942008972168, "learning_rate": 2.1511622322607947e-06, "loss": 4.2242, "step": 1098400 }, { "epoch": 6.623472944127543, "grad_norm": 4.285478591918945, "learning_rate": 2.147729708820228e-06, "loss": 4.1656, "step": 1098450 }, { "epoch": 6.623774435908877, "grad_norm": 4.471106052398682, "learning_rate": 2.144299906389696e-06, "loss": 4.2421, "step": 1098500 }, { "epoch": 6.624075927690211, "grad_norm": 3.668836832046509, "learning_rate": 2.140872825032347e-06, "loss": 4.0042, "step": 1098550 }, { "epoch": 6.6243774194715455, "grad_norm": 6.272404670715332, "learning_rate": 2.1374484648111976e-06, "loss": 3.9947, "step": 1098600 }, { "epoch": 6.624678911252879, "grad_norm": 4.4339280128479, "learning_rate": 2.1340268257893145e-06, "loss": 4.3968, "step": 1098650 }, { "epoch": 6.624980403034213, "grad_norm": 6.290050506591797, "learning_rate": 2.130607908029647e-06, "loss": 3.8959, "step": 1098700 }, { "epoch": 6.625281894815547, "grad_norm": 4.9315032958984375, "learning_rate": 2.1271917115951108e-06, "loss": 4.0489, "step": 1098750 }, { "epoch": 6.625583386596881, "grad_norm": 7.176652908325195, "learning_rate": 2.12377823654859e-06, "loss": 4.2724, "step": 1098800 }, { "epoch": 6.625884878378216, "grad_norm": 6.319835662841797, "learning_rate": 2.120367482952917e-06, "loss": 4.0378, "step": 1098850 }, { "epoch": 6.62618637015955, "grad_norm": 4.136597633361816, "learning_rate": 2.1169594508708256e-06, "loss": 4.3625, "step": 1098900 }, { "epoch": 6.626487861940883, "grad_norm": 4.052674770355225, "learning_rate": 2.1135541403650645e-06, "loss": 4.2243, "step": 1098950 }, { "epoch": 6.626789353722217, "grad_norm": 7.173666000366211, "learning_rate": 2.1101515514983013e-06, "loss": 4.2889, "step": 1099000 }, { "epoch": 6.6270908455035515, "grad_norm": 6.129220485687256, "learning_rate": 2.1067516843331355e-06, "loss": 4.2295, "step": 1099050 }, { "epoch": 6.627392337284886, "grad_norm": 7.574476718902588, "learning_rate": 2.1033545389321505e-06, "loss": 4.0252, "step": 1099100 }, { "epoch": 6.62769382906622, "grad_norm": 3.5396947860717773, "learning_rate": 2.0999601153578626e-06, "loss": 4.4046, "step": 1099150 }, { "epoch": 6.627995320847553, "grad_norm": 3.903926134109497, "learning_rate": 2.0965684136727555e-06, "loss": 3.9067, "step": 1099200 }, { "epoch": 6.6282968126288875, "grad_norm": 6.322651386260986, "learning_rate": 2.0931794339392125e-06, "loss": 4.2122, "step": 1099250 }, { "epoch": 6.628598304410222, "grad_norm": 4.680136680603027, "learning_rate": 2.0897931762196675e-06, "loss": 4.3121, "step": 1099300 }, { "epoch": 6.628899796191556, "grad_norm": 7.500360012054443, "learning_rate": 2.0864096405763698e-06, "loss": 4.2282, "step": 1099350 }, { "epoch": 6.62920128797289, "grad_norm": 4.407736301422119, "learning_rate": 2.083028827071637e-06, "loss": 4.2659, "step": 1099400 }, { "epoch": 6.629502779754224, "grad_norm": 4.139984130859375, "learning_rate": 2.079650735767652e-06, "loss": 3.9702, "step": 1099450 }, { "epoch": 6.629804271535558, "grad_norm": 5.201292037963867, "learning_rate": 2.076275366726615e-06, "loss": 4.2723, "step": 1099500 }, { "epoch": 6.630105763316892, "grad_norm": 5.418088436126709, "learning_rate": 2.0729027200106106e-06, "loss": 4.4616, "step": 1099550 }, { "epoch": 6.630407255098226, "grad_norm": 4.195554256439209, "learning_rate": 2.0695327956817555e-06, "loss": 4.1353, "step": 1099600 }, { "epoch": 6.63070874687956, "grad_norm": 5.705583095550537, "learning_rate": 2.066165593802033e-06, "loss": 3.9932, "step": 1099650 }, { "epoch": 6.6310102386608945, "grad_norm": 5.959312915802002, "learning_rate": 2.06280111443341e-06, "loss": 4.1018, "step": 1099700 }, { "epoch": 6.631311730442228, "grad_norm": 5.035968780517578, "learning_rate": 2.059439357637821e-06, "loss": 4.0273, "step": 1099750 }, { "epoch": 6.631613222223562, "grad_norm": 6.86531400680542, "learning_rate": 2.056080323477133e-06, "loss": 4.5649, "step": 1099800 }, { "epoch": 6.631914714004896, "grad_norm": 7.092698097229004, "learning_rate": 2.052724012013146e-06, "loss": 4.2087, "step": 1099850 }, { "epoch": 6.63221620578623, "grad_norm": 5.551528453826904, "learning_rate": 2.049370423307661e-06, "loss": 4.1786, "step": 1099900 }, { "epoch": 6.632517697567565, "grad_norm": 5.275874614715576, "learning_rate": 2.0460195574223625e-06, "loss": 4.3209, "step": 1099950 }, { "epoch": 6.632819189348899, "grad_norm": 5.8669047355651855, "learning_rate": 2.042671414418934e-06, "loss": 4.0526, "step": 1100000 }, { "epoch": 6.633120681130232, "grad_norm": 6.957724094390869, "learning_rate": 2.0393259943589922e-06, "loss": 4.0979, "step": 1100050 }, { "epoch": 6.633422172911566, "grad_norm": 5.233582019805908, "learning_rate": 2.0359832973041055e-06, "loss": 4.3606, "step": 1100100 }, { "epoch": 6.6337236646929005, "grad_norm": 6.4209513664245605, "learning_rate": 2.032643323315791e-06, "loss": 4.5145, "step": 1100150 }, { "epoch": 6.634025156474235, "grad_norm": 4.971188545227051, "learning_rate": 2.0293060724555333e-06, "loss": 4.3759, "step": 1100200 }, { "epoch": 6.634326648255568, "grad_norm": 8.993685722351074, "learning_rate": 2.0259715447846992e-06, "loss": 4.0048, "step": 1100250 }, { "epoch": 6.634628140036902, "grad_norm": 5.229689598083496, "learning_rate": 2.0226397403647066e-06, "loss": 4.1803, "step": 1100300 }, { "epoch": 6.6349296318182365, "grad_norm": 3.172816038131714, "learning_rate": 2.01931065925684e-06, "loss": 4.104, "step": 1100350 }, { "epoch": 6.635231123599571, "grad_norm": 5.379385948181152, "learning_rate": 2.0159843015224e-06, "loss": 4.2776, "step": 1100400 }, { "epoch": 6.635532615380905, "grad_norm": 4.215533256530762, "learning_rate": 2.012660667222571e-06, "loss": 4.2763, "step": 1100450 }, { "epoch": 6.635834107162239, "grad_norm": 4.848653793334961, "learning_rate": 2.009339756418538e-06, "loss": 4.2045, "step": 1100500 }, { "epoch": 6.636135598943572, "grad_norm": 3.205756187438965, "learning_rate": 2.0060215691714176e-06, "loss": 4.1813, "step": 1100550 }, { "epoch": 6.636437090724907, "grad_norm": 6.086948394775391, "learning_rate": 2.0027061055422624e-06, "loss": 4.2672, "step": 1100600 }, { "epoch": 6.636738582506241, "grad_norm": 2.308504343032837, "learning_rate": 1.999393365592106e-06, "loss": 4.3369, "step": 1100650 }, { "epoch": 6.637040074287575, "grad_norm": 4.9747700691223145, "learning_rate": 1.9960833493818995e-06, "loss": 3.6274, "step": 1100700 }, { "epoch": 6.637341566068909, "grad_norm": 3.737058401107788, "learning_rate": 1.9927760569725613e-06, "loss": 3.8738, "step": 1100750 }, { "epoch": 6.637643057850243, "grad_norm": 5.620214462280273, "learning_rate": 1.9894714884249595e-06, "loss": 4.4573, "step": 1100800 }, { "epoch": 6.637944549631577, "grad_norm": 5.478801727294922, "learning_rate": 1.9861696437999284e-06, "loss": 4.07, "step": 1100850 }, { "epoch": 6.638246041412911, "grad_norm": 5.599678039550781, "learning_rate": 1.98287052315822e-06, "loss": 4.1225, "step": 1100900 }, { "epoch": 6.638547533194245, "grad_norm": 6.7436723709106445, "learning_rate": 1.9795741265605347e-06, "loss": 4.298, "step": 1100950 }, { "epoch": 6.638849024975579, "grad_norm": 3.7589221000671387, "learning_rate": 1.9762804540675747e-06, "loss": 4.0888, "step": 1101000 }, { "epoch": 6.638849024975579, "eval_loss": 4.790594100952148, "eval_runtime": 39.0592, "eval_samples_per_second": 13.108, "eval_steps_per_second": 6.554, "eval_tts_loss": 8.157228215917899, "step": 1101000 }, { "epoch": 6.639150516756914, "grad_norm": 5.554348945617676, "learning_rate": 1.972989505739908e-06, "loss": 4.3872, "step": 1101050 }, { "epoch": 6.639452008538247, "grad_norm": 5.302098751068115, "learning_rate": 1.9697012816381197e-06, "loss": 4.0846, "step": 1101100 }, { "epoch": 6.639753500319581, "grad_norm": 6.000420570373535, "learning_rate": 1.9664157818227446e-06, "loss": 3.8161, "step": 1101150 }, { "epoch": 6.640054992100915, "grad_norm": 7.967617511749268, "learning_rate": 1.9631330063542173e-06, "loss": 4.2647, "step": 1101200 }, { "epoch": 6.6403564838822495, "grad_norm": 6.26718807220459, "learning_rate": 1.9598529552929732e-06, "loss": 4.4118, "step": 1101250 }, { "epoch": 6.640657975663584, "grad_norm": 2.0659029483795166, "learning_rate": 1.9565756286993804e-06, "loss": 4.3331, "step": 1101300 }, { "epoch": 6.640959467444917, "grad_norm": 6.784183979034424, "learning_rate": 1.9533010266337413e-06, "loss": 4.5445, "step": 1101350 }, { "epoch": 6.641260959226251, "grad_norm": 4.2803874015808105, "learning_rate": 1.950029149156307e-06, "loss": 4.1015, "step": 1101400 }, { "epoch": 6.6415624510075855, "grad_norm": 4.739297866821289, "learning_rate": 1.9467599963273294e-06, "loss": 4.1188, "step": 1101450 }, { "epoch": 6.64186394278892, "grad_norm": 4.557848930358887, "learning_rate": 1.9434935682069277e-06, "loss": 4.4392, "step": 1101500 }, { "epoch": 6.642165434570254, "grad_norm": 4.20066499710083, "learning_rate": 1.940229864855253e-06, "loss": 4.1628, "step": 1101550 }, { "epoch": 6.642466926351588, "grad_norm": 3.215616464614868, "learning_rate": 1.9369688863323574e-06, "loss": 3.8415, "step": 1101600 }, { "epoch": 6.642768418132921, "grad_norm": 5.948782920837402, "learning_rate": 1.9337106326982434e-06, "loss": 4.1595, "step": 1101650 }, { "epoch": 6.643069909914256, "grad_norm": 6.017777919769287, "learning_rate": 1.930455104012896e-06, "loss": 4.3581, "step": 1101700 }, { "epoch": 6.64337140169559, "grad_norm": 4.401530742645264, "learning_rate": 1.927202300336217e-06, "loss": 4.4499, "step": 1101750 }, { "epoch": 6.643672893476924, "grad_norm": 3.7520554065704346, "learning_rate": 1.923952221728042e-06, "loss": 4.2267, "step": 1101800 }, { "epoch": 6.643974385258258, "grad_norm": 6.346116542816162, "learning_rate": 1.920704868248224e-06, "loss": 3.8469, "step": 1101850 }, { "epoch": 6.6442758770395915, "grad_norm": 2.806222915649414, "learning_rate": 1.9174602399565307e-06, "loss": 4.5486, "step": 1101900 }, { "epoch": 6.644577368820926, "grad_norm": 4.962444305419922, "learning_rate": 1.9142183369126317e-06, "loss": 4.466, "step": 1101950 }, { "epoch": 6.64487886060226, "grad_norm": 4.972400665283203, "learning_rate": 1.9109791591762124e-06, "loss": 4.0974, "step": 1102000 }, { "epoch": 6.645180352383594, "grad_norm": 6.185055255889893, "learning_rate": 1.9077427068069082e-06, "loss": 4.1594, "step": 1102050 }, { "epoch": 6.645481844164928, "grad_norm": 5.495204925537109, "learning_rate": 1.9045089798642555e-06, "loss": 4.3235, "step": 1102100 }, { "epoch": 6.645783335946263, "grad_norm": 5.5151495933532715, "learning_rate": 1.9012779784077725e-06, "loss": 4.3975, "step": 1102150 }, { "epoch": 6.646084827727596, "grad_norm": 7.7602972984313965, "learning_rate": 1.898049702496912e-06, "loss": 4.5412, "step": 1102200 }, { "epoch": 6.64638631950893, "grad_norm": 2.9885103702545166, "learning_rate": 1.8948241521910934e-06, "loss": 3.8516, "step": 1102250 }, { "epoch": 6.646687811290264, "grad_norm": 5.091947078704834, "learning_rate": 1.8916013275496855e-06, "loss": 4.1874, "step": 1102300 }, { "epoch": 6.6469893030715985, "grad_norm": 6.613893985748291, "learning_rate": 1.888381228631991e-06, "loss": 4.455, "step": 1102350 }, { "epoch": 6.647290794852932, "grad_norm": 6.3494873046875, "learning_rate": 1.8851638554972626e-06, "loss": 3.9636, "step": 1102400 }, { "epoch": 6.647592286634266, "grad_norm": 6.343989849090576, "learning_rate": 1.881949208204736e-06, "loss": 4.1876, "step": 1102450 }, { "epoch": 6.6478937784156, "grad_norm": 5.755059242248535, "learning_rate": 1.8787372868135642e-06, "loss": 4.2454, "step": 1102500 }, { "epoch": 6.6481952701969345, "grad_norm": 5.6874871253967285, "learning_rate": 1.875528091382833e-06, "loss": 4.5481, "step": 1102550 }, { "epoch": 6.648496761978269, "grad_norm": 10.005180358886719, "learning_rate": 1.8723216219716286e-06, "loss": 3.9744, "step": 1102600 }, { "epoch": 6.648798253759603, "grad_norm": 6.0677995681762695, "learning_rate": 1.8691178786389704e-06, "loss": 4.3149, "step": 1102650 }, { "epoch": 6.649099745540936, "grad_norm": 5.073037147521973, "learning_rate": 1.8659168614437947e-06, "loss": 3.4708, "step": 1102700 }, { "epoch": 6.64940123732227, "grad_norm": 4.749525547027588, "learning_rate": 1.8627185704450042e-06, "loss": 3.798, "step": 1102750 }, { "epoch": 6.649702729103605, "grad_norm": 5.973727703094482, "learning_rate": 1.859523005701502e-06, "loss": 4.1095, "step": 1102800 }, { "epoch": 6.650004220884939, "grad_norm": 4.251709938049316, "learning_rate": 1.856330167272041e-06, "loss": 4.3011, "step": 1102850 }, { "epoch": 6.650305712666273, "grad_norm": 6.109402656555176, "learning_rate": 1.8531400552154407e-06, "loss": 3.8717, "step": 1102900 }, { "epoch": 6.650607204447606, "grad_norm": 5.179881572723389, "learning_rate": 1.849952669590371e-06, "loss": 4.3237, "step": 1102950 }, { "epoch": 6.6509086962289405, "grad_norm": 4.113331317901611, "learning_rate": 1.846768010455485e-06, "loss": 3.5678, "step": 1103000 }, { "epoch": 6.651210188010275, "grad_norm": 1.3345441818237305, "learning_rate": 1.843586077869419e-06, "loss": 4.3427, "step": 1103050 }, { "epoch": 6.651511679791609, "grad_norm": 5.164570331573486, "learning_rate": 1.8404068718907261e-06, "loss": 4.1254, "step": 1103100 }, { "epoch": 6.651813171572943, "grad_norm": 7.364490985870361, "learning_rate": 1.8372303925779097e-06, "loss": 4.2616, "step": 1103150 }, { "epoch": 6.652114663354277, "grad_norm": 6.005956172943115, "learning_rate": 1.8340566399894064e-06, "loss": 3.7899, "step": 1103200 }, { "epoch": 6.652416155135611, "grad_norm": 4.093573093414307, "learning_rate": 1.8308856141836692e-06, "loss": 3.9702, "step": 1103250 }, { "epoch": 6.652717646916945, "grad_norm": 6.441170692443848, "learning_rate": 1.827717315219035e-06, "loss": 4.427, "step": 1103300 }, { "epoch": 6.653019138698279, "grad_norm": 5.744950294494629, "learning_rate": 1.824551743153807e-06, "loss": 4.4758, "step": 1103350 }, { "epoch": 6.653320630479613, "grad_norm": 3.748976945877075, "learning_rate": 1.8213888980462387e-06, "loss": 3.9583, "step": 1103400 }, { "epoch": 6.6536221222609475, "grad_norm": 7.212212085723877, "learning_rate": 1.8182287799545503e-06, "loss": 3.9936, "step": 1103450 }, { "epoch": 6.653923614042281, "grad_norm": 4.604905128479004, "learning_rate": 1.815071388936895e-06, "loss": 4.2869, "step": 1103500 }, { "epoch": 6.654225105823615, "grad_norm": 5.185705661773682, "learning_rate": 1.8119167250513765e-06, "loss": 3.9836, "step": 1103550 }, { "epoch": 6.654526597604949, "grad_norm": 5.883387088775635, "learning_rate": 1.8087647883560652e-06, "loss": 3.7853, "step": 1103600 }, { "epoch": 6.6548280893862835, "grad_norm": 9.924818992614746, "learning_rate": 1.805615578908931e-06, "loss": 4.2102, "step": 1103650 }, { "epoch": 6.655129581167618, "grad_norm": 7.070631980895996, "learning_rate": 1.8024690967679944e-06, "loss": 4.463, "step": 1103700 }, { "epoch": 6.655431072948952, "grad_norm": 4.569149017333984, "learning_rate": 1.799325341991109e-06, "loss": 4.3072, "step": 1103750 }, { "epoch": 6.655732564730285, "grad_norm": 4.836187839508057, "learning_rate": 1.7961843146361289e-06, "loss": 4.517, "step": 1103800 }, { "epoch": 6.656034056511619, "grad_norm": 10.231403350830078, "learning_rate": 1.7930460147609071e-06, "loss": 3.8112, "step": 1103850 }, { "epoch": 6.656335548292954, "grad_norm": 5.57838249206543, "learning_rate": 1.789910442423148e-06, "loss": 4.136, "step": 1103900 }, { "epoch": 6.656637040074288, "grad_norm": 5.978504657745361, "learning_rate": 1.7867775976805888e-06, "loss": 4.0939, "step": 1103950 }, { "epoch": 6.656938531855621, "grad_norm": 4.085102558135986, "learning_rate": 1.783647480590883e-06, "loss": 4.0856, "step": 1104000 }, { "epoch": 6.656938531855621, "eval_loss": 4.789715766906738, "eval_runtime": 39.1049, "eval_samples_per_second": 13.093, "eval_steps_per_second": 6.547, "eval_tts_loss": 8.159197739761566, "step": 1104000 }, { "epoch": 6.657240023636955, "grad_norm": 3.7532835006713867, "learning_rate": 1.780520091211618e-06, "loss": 4.4759, "step": 1104050 }, { "epoch": 6.6575415154182895, "grad_norm": 3.464975118637085, "learning_rate": 1.7773954296003644e-06, "loss": 4.2549, "step": 1104100 }, { "epoch": 6.657843007199624, "grad_norm": 5.798441410064697, "learning_rate": 1.7742734958146265e-06, "loss": 4.3094, "step": 1104150 }, { "epoch": 6.658144498980958, "grad_norm": 4.438235759735107, "learning_rate": 1.771154289911858e-06, "loss": 4.4391, "step": 1104200 }, { "epoch": 6.658445990762292, "grad_norm": 4.711209774017334, "learning_rate": 1.7680378119494465e-06, "loss": 4.296, "step": 1104250 }, { "epoch": 6.6587474825436255, "grad_norm": 5.53789758682251, "learning_rate": 1.7649240619847626e-06, "loss": 3.9965, "step": 1104300 }, { "epoch": 6.65904897432496, "grad_norm": 6.815347194671631, "learning_rate": 1.7618130400751273e-06, "loss": 4.167, "step": 1104350 }, { "epoch": 6.659350466106294, "grad_norm": 4.9884161949157715, "learning_rate": 1.7587047462777614e-06, "loss": 3.9931, "step": 1104400 }, { "epoch": 6.659651957887628, "grad_norm": 4.819449424743652, "learning_rate": 1.755599180649886e-06, "loss": 4.2215, "step": 1104450 }, { "epoch": 6.659953449668962, "grad_norm": 4.396287441253662, "learning_rate": 1.752496343248655e-06, "loss": 4.4184, "step": 1104500 }, { "epoch": 6.660254941450296, "grad_norm": 3.9964184761047363, "learning_rate": 1.7493962341311728e-06, "loss": 4.1052, "step": 1104550 }, { "epoch": 6.66055643323163, "grad_norm": 8.322312355041504, "learning_rate": 1.746298853354461e-06, "loss": 4.5151, "step": 1104600 }, { "epoch": 6.660857925012964, "grad_norm": 5.316822528839111, "learning_rate": 1.743204200975573e-06, "loss": 4.2493, "step": 1104650 }, { "epoch": 6.661159416794298, "grad_norm": 4.985698223114014, "learning_rate": 1.7401122770514308e-06, "loss": 4.3866, "step": 1104700 }, { "epoch": 6.6614609085756324, "grad_norm": 6.872057914733887, "learning_rate": 1.7370230816389386e-06, "loss": 4.4298, "step": 1104750 }, { "epoch": 6.661762400356967, "grad_norm": 5.16295051574707, "learning_rate": 1.7339366147949674e-06, "loss": 4.0157, "step": 1104800 }, { "epoch": 6.6620638921383, "grad_norm": 4.472875595092773, "learning_rate": 1.7308528765763053e-06, "loss": 3.9982, "step": 1104850 }, { "epoch": 6.662365383919634, "grad_norm": 4.359952449798584, "learning_rate": 1.7277718670397067e-06, "loss": 4.2804, "step": 1104900 }, { "epoch": 6.662666875700968, "grad_norm": 3.9883008003234863, "learning_rate": 1.7246935862418599e-06, "loss": 4.4922, "step": 1104950 }, { "epoch": 6.662968367482303, "grad_norm": 4.50124454498291, "learning_rate": 1.7216180342394358e-06, "loss": 3.9141, "step": 1105000 }, { "epoch": 6.663269859263637, "grad_norm": 5.998719692230225, "learning_rate": 1.7185452110890064e-06, "loss": 4.3143, "step": 1105050 }, { "epoch": 6.66357135104497, "grad_norm": 5.641767978668213, "learning_rate": 1.715475116847176e-06, "loss": 4.3983, "step": 1105100 }, { "epoch": 6.663872842826304, "grad_norm": 7.388015270233154, "learning_rate": 1.7124077515703826e-06, "loss": 4.4813, "step": 1105150 }, { "epoch": 6.6641743346076385, "grad_norm": 5.15199613571167, "learning_rate": 1.709343115315115e-06, "loss": 4.042, "step": 1105200 }, { "epoch": 6.664475826388973, "grad_norm": 4.492806911468506, "learning_rate": 1.7062812081377774e-06, "loss": 4.1546, "step": 1105250 }, { "epoch": 6.664777318170307, "grad_norm": 3.8489530086517334, "learning_rate": 1.7032220300947086e-06, "loss": 3.8103, "step": 1105300 }, { "epoch": 6.665078809951641, "grad_norm": 7.698802471160889, "learning_rate": 1.7001655812421966e-06, "loss": 4.0643, "step": 1105350 }, { "epoch": 6.6653803017329745, "grad_norm": 5.762256145477295, "learning_rate": 1.6971118616365132e-06, "loss": 4.1938, "step": 1105400 }, { "epoch": 6.665681793514309, "grad_norm": 3.2086446285247803, "learning_rate": 1.69406087133383e-06, "loss": 4.232, "step": 1105450 }, { "epoch": 6.665983285295643, "grad_norm": 3.236704111099243, "learning_rate": 1.6910126103903187e-06, "loss": 4.339, "step": 1105500 }, { "epoch": 6.666284777076977, "grad_norm": 3.8549015522003174, "learning_rate": 1.687967078862068e-06, "loss": 4.0534, "step": 1105550 }, { "epoch": 6.666586268858311, "grad_norm": 3.1488122940063477, "learning_rate": 1.684924276805133e-06, "loss": 4.0233, "step": 1105600 }, { "epoch": 6.666887760639645, "grad_norm": 5.675065517425537, "learning_rate": 1.6818842042755187e-06, "loss": 4.4446, "step": 1105650 }, { "epoch": 6.667189252420979, "grad_norm": 5.5588860511779785, "learning_rate": 1.678846861329164e-06, "loss": 4.208, "step": 1105700 }, { "epoch": 6.667490744202313, "grad_norm": 4.2749857902526855, "learning_rate": 1.675812248021957e-06, "loss": 4.0993, "step": 1105750 }, { "epoch": 6.667792235983647, "grad_norm": 4.6535773277282715, "learning_rate": 1.6727803644097704e-06, "loss": 4.2924, "step": 1105800 }, { "epoch": 6.668093727764981, "grad_norm": 5.967102527618408, "learning_rate": 1.6697512105483756e-06, "loss": 4.4933, "step": 1105850 }, { "epoch": 6.668395219546316, "grad_norm": 5.5305304527282715, "learning_rate": 1.6667247864935284e-06, "loss": 4.2348, "step": 1105900 }, { "epoch": 6.668696711327649, "grad_norm": 5.618762969970703, "learning_rate": 1.6637010923009343e-06, "loss": 3.8274, "step": 1105950 }, { "epoch": 6.668998203108983, "grad_norm": 7.148965358734131, "learning_rate": 1.6606801280262317e-06, "loss": 4.3084, "step": 1106000 }, { "epoch": 6.669299694890317, "grad_norm": 5.856303691864014, "learning_rate": 1.6576618937250098e-06, "loss": 4.1851, "step": 1106050 }, { "epoch": 6.669601186671652, "grad_norm": 3.9617934226989746, "learning_rate": 1.6546463894528406e-06, "loss": 4.2483, "step": 1106100 }, { "epoch": 6.669902678452985, "grad_norm": 6.60610294342041, "learning_rate": 1.6516336152652133e-06, "loss": 3.921, "step": 1106150 }, { "epoch": 6.670204170234319, "grad_norm": 5.712179660797119, "learning_rate": 1.6486235712175499e-06, "loss": 4.0182, "step": 1106200 }, { "epoch": 6.670505662015653, "grad_norm": 5.59325647354126, "learning_rate": 1.6456162573652565e-06, "loss": 3.9318, "step": 1106250 }, { "epoch": 6.6708071537969875, "grad_norm": 5.1982622146606445, "learning_rate": 1.6426116737636886e-06, "loss": 4.4142, "step": 1106300 }, { "epoch": 6.671108645578322, "grad_norm": 6.625081539154053, "learning_rate": 1.6396098204681352e-06, "loss": 4.2681, "step": 1106350 }, { "epoch": 6.671410137359656, "grad_norm": 8.37369155883789, "learning_rate": 1.6366106975338356e-06, "loss": 4.0925, "step": 1106400 }, { "epoch": 6.671711629140989, "grad_norm": 3.4025700092315674, "learning_rate": 1.6336143050159956e-06, "loss": 3.9749, "step": 1106450 }, { "epoch": 6.6720131209223235, "grad_norm": 5.359844207763672, "learning_rate": 1.6306206429697543e-06, "loss": 4.1861, "step": 1106500 }, { "epoch": 6.672314612703658, "grad_norm": 5.355064868927002, "learning_rate": 1.627629711450218e-06, "loss": 4.3491, "step": 1106550 }, { "epoch": 6.672616104484992, "grad_norm": 3.4009478092193604, "learning_rate": 1.6246415105124089e-06, "loss": 3.8642, "step": 1106600 }, { "epoch": 6.672917596266326, "grad_norm": 3.5330536365509033, "learning_rate": 1.6216560402113166e-06, "loss": 4.1911, "step": 1106650 }, { "epoch": 6.673219088047659, "grad_norm": 5.941165924072266, "learning_rate": 1.6186733006019137e-06, "loss": 4.136, "step": 1106700 }, { "epoch": 6.673520579828994, "grad_norm": 5.5703840255737305, "learning_rate": 1.615693291739073e-06, "loss": 4.4859, "step": 1106750 }, { "epoch": 6.673822071610328, "grad_norm": 3.2846591472625732, "learning_rate": 1.612716013677634e-06, "loss": 3.9494, "step": 1106800 }, { "epoch": 6.674123563391662, "grad_norm": 4.334377288818359, "learning_rate": 1.6097414664724029e-06, "loss": 4.1536, "step": 1106850 }, { "epoch": 6.674425055172996, "grad_norm": 3.6216814517974854, "learning_rate": 1.606769650178119e-06, "loss": 4.1221, "step": 1106900 }, { "epoch": 6.67472654695433, "grad_norm": 3.858445882797241, "learning_rate": 1.6038005648494723e-06, "loss": 4.1221, "step": 1106950 }, { "epoch": 6.675028038735664, "grad_norm": 4.398157596588135, "learning_rate": 1.6008342105410855e-06, "loss": 4.2086, "step": 1107000 }, { "epoch": 6.675028038735664, "eval_loss": 4.790506362915039, "eval_runtime": 39.096, "eval_samples_per_second": 13.096, "eval_steps_per_second": 6.548, "eval_tts_loss": 8.17524794495587, "step": 1107000 }, { "epoch": 6.675329530516998, "grad_norm": 4.6863813400268555, "learning_rate": 1.597870587307598e-06, "loss": 3.8616, "step": 1107050 }, { "epoch": 6.675631022298332, "grad_norm": 5.871870994567871, "learning_rate": 1.5949096952035002e-06, "loss": 4.2487, "step": 1107100 }, { "epoch": 6.675932514079666, "grad_norm": 6.09662389755249, "learning_rate": 1.5919515342832977e-06, "loss": 4.0226, "step": 1107150 }, { "epoch": 6.676234005861001, "grad_norm": 4.269097328186035, "learning_rate": 1.5889961046014643e-06, "loss": 4.3427, "step": 1107200 }, { "epoch": 6.676535497642334, "grad_norm": 5.558149337768555, "learning_rate": 1.5860434062123396e-06, "loss": 3.9678, "step": 1107250 }, { "epoch": 6.676836989423668, "grad_norm": 3.930340051651001, "learning_rate": 1.5830934391702966e-06, "loss": 3.9363, "step": 1107300 }, { "epoch": 6.677138481205002, "grad_norm": 3.218353271484375, "learning_rate": 1.5801462035296253e-06, "loss": 4.1683, "step": 1107350 }, { "epoch": 6.6774399729863365, "grad_norm": 3.9938652515411377, "learning_rate": 1.5772016993445324e-06, "loss": 4.4014, "step": 1107400 }, { "epoch": 6.677741464767671, "grad_norm": 5.289726734161377, "learning_rate": 1.574259926669258e-06, "loss": 4.3042, "step": 1107450 }, { "epoch": 6.678042956549005, "grad_norm": 5.92061185836792, "learning_rate": 1.5713208855579086e-06, "loss": 3.9361, "step": 1107500 }, { "epoch": 6.678344448330338, "grad_norm": 5.72264289855957, "learning_rate": 1.5683845760645575e-06, "loss": 3.9458, "step": 1107550 }, { "epoch": 6.6786459401116725, "grad_norm": 7.024046421051025, "learning_rate": 1.5654509982432784e-06, "loss": 4.1547, "step": 1107600 }, { "epoch": 6.678947431893007, "grad_norm": 5.612002372741699, "learning_rate": 1.5625201521480445e-06, "loss": 3.93, "step": 1107650 }, { "epoch": 6.679248923674341, "grad_norm": 10.69078254699707, "learning_rate": 1.559592037832813e-06, "loss": 4.0311, "step": 1107700 }, { "epoch": 6.679550415455674, "grad_norm": 5.516969203948975, "learning_rate": 1.5566666553514407e-06, "loss": 3.9795, "step": 1107750 }, { "epoch": 6.679851907237008, "grad_norm": 5.019056797027588, "learning_rate": 1.5537440047577843e-06, "loss": 4.3737, "step": 1107800 }, { "epoch": 6.680153399018343, "grad_norm": 4.255172252655029, "learning_rate": 1.5508240861056176e-06, "loss": 4.2781, "step": 1107850 }, { "epoch": 6.680454890799677, "grad_norm": 6.434134006500244, "learning_rate": 1.547906899448681e-06, "loss": 4.1734, "step": 1107900 }, { "epoch": 6.680756382581011, "grad_norm": 7.720279693603516, "learning_rate": 1.5449924448406647e-06, "loss": 4.2969, "step": 1107950 }, { "epoch": 6.681057874362345, "grad_norm": 5.8385138511657715, "learning_rate": 1.542080722335226e-06, "loss": 4.3807, "step": 1108000 }, { "epoch": 6.6813593661436785, "grad_norm": 9.654234886169434, "learning_rate": 1.5391717319859053e-06, "loss": 3.9724, "step": 1108050 }, { "epoch": 6.681660857925013, "grad_norm": 4.40015983581543, "learning_rate": 1.536265473846293e-06, "loss": 4.0473, "step": 1108100 }, { "epoch": 6.681962349706347, "grad_norm": 6.748693466186523, "learning_rate": 1.53336194796983e-06, "loss": 4.6959, "step": 1108150 }, { "epoch": 6.682263841487681, "grad_norm": 6.26387882232666, "learning_rate": 1.5304611544099564e-06, "loss": 4.275, "step": 1108200 }, { "epoch": 6.682565333269015, "grad_norm": 6.687446594238281, "learning_rate": 1.52756309322008e-06, "loss": 4.2867, "step": 1108250 }, { "epoch": 6.682866825050349, "grad_norm": 7.100801944732666, "learning_rate": 1.5246677644535243e-06, "loss": 3.9592, "step": 1108300 }, { "epoch": 6.683168316831683, "grad_norm": 7.499209880828857, "learning_rate": 1.5217751681635637e-06, "loss": 4.4491, "step": 1108350 }, { "epoch": 6.683469808613017, "grad_norm": 5.056241035461426, "learning_rate": 1.5188853044034554e-06, "loss": 4.1524, "step": 1108400 }, { "epoch": 6.683771300394351, "grad_norm": 1.3650054931640625, "learning_rate": 1.515998173226357e-06, "loss": 3.9236, "step": 1108450 }, { "epoch": 6.6840727921756855, "grad_norm": 4.545261383056641, "learning_rate": 1.513113774685426e-06, "loss": 3.848, "step": 1108500 }, { "epoch": 6.68437428395702, "grad_norm": 3.919753313064575, "learning_rate": 1.5102321088337199e-06, "loss": 4.1627, "step": 1108550 }, { "epoch": 6.684675775738353, "grad_norm": 5.3151702880859375, "learning_rate": 1.5073531757242962e-06, "loss": 3.975, "step": 1108600 }, { "epoch": 6.684977267519687, "grad_norm": 5.586663246154785, "learning_rate": 1.5044769754101128e-06, "loss": 4.0397, "step": 1108650 }, { "epoch": 6.6852787593010214, "grad_norm": 6.6194939613342285, "learning_rate": 1.501603507944127e-06, "loss": 4.2364, "step": 1108700 }, { "epoch": 6.685580251082356, "grad_norm": 4.379240989685059, "learning_rate": 1.4987327733792132e-06, "loss": 4.0775, "step": 1108750 }, { "epoch": 6.68588174286369, "grad_norm": 5.556980133056641, "learning_rate": 1.495864771768196e-06, "loss": 4.1818, "step": 1108800 }, { "epoch": 6.686183234645023, "grad_norm": 5.306129455566406, "learning_rate": 1.4929995031638498e-06, "loss": 4.4786, "step": 1108850 }, { "epoch": 6.686484726426357, "grad_norm": 3.2152626514434814, "learning_rate": 1.4901369676189324e-06, "loss": 4.0408, "step": 1108900 }, { "epoch": 6.686786218207692, "grad_norm": 4.009822845458984, "learning_rate": 1.4872771651861181e-06, "loss": 4.3414, "step": 1108950 }, { "epoch": 6.687087709989026, "grad_norm": 6.728217124938965, "learning_rate": 1.4844200959180152e-06, "loss": 4.1112, "step": 1109000 }, { "epoch": 6.68738920177036, "grad_norm": 5.62774658203125, "learning_rate": 1.481565759867215e-06, "loss": 4.2533, "step": 1109050 }, { "epoch": 6.687690693551694, "grad_norm": 2.521099328994751, "learning_rate": 1.4787141570862414e-06, "loss": 4.4301, "step": 1109100 }, { "epoch": 6.6879921853330275, "grad_norm": 4.641260623931885, "learning_rate": 1.4758652876275867e-06, "loss": 4.3028, "step": 1109150 }, { "epoch": 6.688293677114362, "grad_norm": 4.989048004150391, "learning_rate": 1.4730191515436918e-06, "loss": 3.9578, "step": 1109200 }, { "epoch": 6.688595168895696, "grad_norm": 4.838942050933838, "learning_rate": 1.4701757488869149e-06, "loss": 4.1999, "step": 1109250 }, { "epoch": 6.68889666067703, "grad_norm": 4.346371650695801, "learning_rate": 1.4673350797095807e-06, "loss": 3.9599, "step": 1109300 }, { "epoch": 6.689198152458364, "grad_norm": 7.4481940269470215, "learning_rate": 1.4644971440639975e-06, "loss": 4.3558, "step": 1109350 }, { "epoch": 6.689499644239698, "grad_norm": 3.3473029136657715, "learning_rate": 1.46166194200234e-06, "loss": 4.1922, "step": 1109400 }, { "epoch": 6.689801136021032, "grad_norm": 5.640780925750732, "learning_rate": 1.4588294735768334e-06, "loss": 4.14, "step": 1109450 }, { "epoch": 6.690102627802366, "grad_norm": 7.681389808654785, "learning_rate": 1.4559997388396028e-06, "loss": 4.164, "step": 1109500 }, { "epoch": 6.6904041195837, "grad_norm": 1.917159080505371, "learning_rate": 1.453172737842706e-06, "loss": 3.945, "step": 1109550 }, { "epoch": 6.6907056113650345, "grad_norm": 5.997549057006836, "learning_rate": 1.4503484706381686e-06, "loss": 4.2769, "step": 1109600 }, { "epoch": 6.691007103146369, "grad_norm": 5.8065266609191895, "learning_rate": 1.447526937277982e-06, "loss": 4.4782, "step": 1109650 }, { "epoch": 6.691308594927702, "grad_norm": 4.118165016174316, "learning_rate": 1.4447081378140712e-06, "loss": 4.0881, "step": 1109700 }, { "epoch": 6.691610086709036, "grad_norm": 4.4250593185424805, "learning_rate": 1.441892072298295e-06, "loss": 3.9195, "step": 1109750 }, { "epoch": 6.69191157849037, "grad_norm": 6.458625793457031, "learning_rate": 1.4390787407825122e-06, "loss": 4.3783, "step": 1109800 }, { "epoch": 6.692213070271705, "grad_norm": 6.385942459106445, "learning_rate": 1.4362681433184641e-06, "loss": 4.1795, "step": 1109850 }, { "epoch": 6.692514562053038, "grad_norm": 6.151184558868408, "learning_rate": 1.4334602799578765e-06, "loss": 4.5081, "step": 1109900 }, { "epoch": 6.692816053834372, "grad_norm": 4.038865566253662, "learning_rate": 1.4306551507524578e-06, "loss": 4.0812, "step": 1109950 }, { "epoch": 6.693117545615706, "grad_norm": 4.566601276397705, "learning_rate": 1.4278527557538e-06, "loss": 4.2758, "step": 1110000 }, { "epoch": 6.693117545615706, "eval_loss": 4.787927150726318, "eval_runtime": 39.2229, "eval_samples_per_second": 13.054, "eval_steps_per_second": 6.527, "eval_tts_loss": 8.170783714180551, "step": 1110000 }, { "epoch": 6.693419037397041, "grad_norm": 5.054520130157471, "learning_rate": 1.4250530950134785e-06, "loss": 4.4296, "step": 1110050 }, { "epoch": 6.693720529178375, "grad_norm": 1.335096836090088, "learning_rate": 1.4222561685830525e-06, "loss": 4.231, "step": 1110100 }, { "epoch": 6.694022020959709, "grad_norm": 4.129940509796143, "learning_rate": 1.4194619765139636e-06, "loss": 4.3518, "step": 1110150 }, { "epoch": 6.694323512741042, "grad_norm": 5.260523796081543, "learning_rate": 1.416670518857621e-06, "loss": 4.1329, "step": 1110200 }, { "epoch": 6.6946250045223765, "grad_norm": 3.6059062480926514, "learning_rate": 1.4138817956654503e-06, "loss": 4.0936, "step": 1110250 }, { "epoch": 6.694926496303711, "grad_norm": 1.9498378038406372, "learning_rate": 1.4110958069887102e-06, "loss": 3.9854, "step": 1110300 }, { "epoch": 6.695227988085045, "grad_norm": 3.2778701782226562, "learning_rate": 1.4083125528787098e-06, "loss": 4.0053, "step": 1110350 }, { "epoch": 6.695529479866379, "grad_norm": 6.754920959472656, "learning_rate": 1.405532033386675e-06, "loss": 4.5124, "step": 1110400 }, { "epoch": 6.6958309716477125, "grad_norm": 5.207382678985596, "learning_rate": 1.4027542485637644e-06, "loss": 4.4506, "step": 1110450 }, { "epoch": 6.696132463429047, "grad_norm": 4.011944770812988, "learning_rate": 1.3999791984611043e-06, "loss": 4.2734, "step": 1110500 }, { "epoch": 6.696433955210381, "grad_norm": 5.504586696624756, "learning_rate": 1.3972068831297701e-06, "loss": 4.299, "step": 1110550 }, { "epoch": 6.696735446991715, "grad_norm": 5.133404731750488, "learning_rate": 1.3944373026207544e-06, "loss": 4.2538, "step": 1110600 }, { "epoch": 6.697036938773049, "grad_norm": 5.356471538543701, "learning_rate": 1.3916704569850666e-06, "loss": 4.2268, "step": 1110650 }, { "epoch": 6.6973384305543835, "grad_norm": 5.517430782318115, "learning_rate": 1.3889063462735994e-06, "loss": 4.5246, "step": 1110700 }, { "epoch": 6.697639922335717, "grad_norm": 4.367376804351807, "learning_rate": 1.3861449705372286e-06, "loss": 4.0353, "step": 1110750 }, { "epoch": 6.697941414117051, "grad_norm": 1.8200300931930542, "learning_rate": 1.38338632982678e-06, "loss": 3.9271, "step": 1110800 }, { "epoch": 6.698242905898385, "grad_norm": 4.425048351287842, "learning_rate": 1.38063042419303e-06, "loss": 4.5143, "step": 1110850 }, { "epoch": 6.698544397679719, "grad_norm": 4.901700973510742, "learning_rate": 1.3778772536866711e-06, "loss": 3.7938, "step": 1110900 }, { "epoch": 6.698845889461054, "grad_norm": 3.69688081741333, "learning_rate": 1.3751268183583964e-06, "loss": 4.0801, "step": 1110950 }, { "epoch": 6.699147381242387, "grad_norm": 8.168766021728516, "learning_rate": 1.372379118258815e-06, "loss": 3.9461, "step": 1111000 }, { "epoch": 6.699448873023721, "grad_norm": 4.943164825439453, "learning_rate": 1.3696341534384868e-06, "loss": 3.9952, "step": 1111050 }, { "epoch": 6.699750364805055, "grad_norm": 5.865137100219727, "learning_rate": 1.3668919239479214e-06, "loss": 4.1979, "step": 1111100 }, { "epoch": 6.70005185658639, "grad_norm": 4.018784046173096, "learning_rate": 1.3641524298376117e-06, "loss": 4.275, "step": 1111150 }, { "epoch": 6.700353348367724, "grad_norm": 1.3743456602096558, "learning_rate": 1.3614156711579672e-06, "loss": 3.5485, "step": 1111200 }, { "epoch": 6.700654840149058, "grad_norm": 5.254673480987549, "learning_rate": 1.3586816479593476e-06, "loss": 4.1628, "step": 1111250 }, { "epoch": 6.700956331930391, "grad_norm": 2.373396635055542, "learning_rate": 1.355950360292063e-06, "loss": 4.2445, "step": 1111300 }, { "epoch": 6.7012578237117255, "grad_norm": 2.8974974155426025, "learning_rate": 1.3532218082064062e-06, "loss": 3.7848, "step": 1111350 }, { "epoch": 6.70155931549306, "grad_norm": 4.655543327331543, "learning_rate": 1.3504959917525538e-06, "loss": 3.9725, "step": 1111400 }, { "epoch": 6.701860807274394, "grad_norm": 6.413947582244873, "learning_rate": 1.3477729109806823e-06, "loss": 4.0944, "step": 1111450 }, { "epoch": 6.702162299055727, "grad_norm": 5.378730297088623, "learning_rate": 1.3450525659409183e-06, "loss": 4.2007, "step": 1111500 }, { "epoch": 6.7024637908370615, "grad_norm": 3.5271503925323486, "learning_rate": 1.3423349566833218e-06, "loss": 4.0766, "step": 1111550 }, { "epoch": 6.702765282618396, "grad_norm": 5.8858771324157715, "learning_rate": 1.3396200832579029e-06, "loss": 4.121, "step": 1111600 }, { "epoch": 6.70306677439973, "grad_norm": 1.5927045345306396, "learning_rate": 1.3369079457146214e-06, "loss": 4.1116, "step": 1111650 }, { "epoch": 6.703368266181064, "grad_norm": 3.916633129119873, "learning_rate": 1.3341985441034042e-06, "loss": 4.4758, "step": 1111700 }, { "epoch": 6.703669757962398, "grad_norm": 4.543774127960205, "learning_rate": 1.3314918784741113e-06, "loss": 4.5435, "step": 1111750 }, { "epoch": 6.703971249743732, "grad_norm": 6.488616466522217, "learning_rate": 1.3287879488765364e-06, "loss": 4.1355, "step": 1111800 }, { "epoch": 6.704272741525066, "grad_norm": 6.2227630615234375, "learning_rate": 1.326086755360456e-06, "loss": 4.1701, "step": 1111850 }, { "epoch": 6.7045742333064, "grad_norm": 5.6447601318359375, "learning_rate": 1.3233882979755805e-06, "loss": 4.3705, "step": 1111900 }, { "epoch": 6.704875725087734, "grad_norm": 2.3772921562194824, "learning_rate": 1.3206925767715704e-06, "loss": 4.2642, "step": 1111950 }, { "epoch": 6.705177216869068, "grad_norm": 8.665020942687988, "learning_rate": 1.3179995917980357e-06, "loss": 4.3251, "step": 1112000 }, { "epoch": 6.705478708650402, "grad_norm": 8.82210636138916, "learning_rate": 1.3153093431045536e-06, "loss": 4.2815, "step": 1112050 }, { "epoch": 6.705780200431736, "grad_norm": 2.262622594833374, "learning_rate": 1.312621830740601e-06, "loss": 4.4182, "step": 1112100 }, { "epoch": 6.70608169221307, "grad_norm": 4.922812461853027, "learning_rate": 1.3099370547556553e-06, "loss": 4.2731, "step": 1112150 }, { "epoch": 6.706383183994404, "grad_norm": 8.680999755859375, "learning_rate": 1.3072550151991268e-06, "loss": 4.415, "step": 1112200 }, { "epoch": 6.706684675775739, "grad_norm": 4.487733840942383, "learning_rate": 1.3045757121203926e-06, "loss": 4.4688, "step": 1112250 }, { "epoch": 6.706986167557073, "grad_norm": 2.665029287338257, "learning_rate": 1.3018991455687133e-06, "loss": 4.0684, "step": 1112300 }, { "epoch": 6.707287659338406, "grad_norm": 6.319240093231201, "learning_rate": 1.2992253155933995e-06, "loss": 4.189, "step": 1112350 }, { "epoch": 6.70758915111974, "grad_norm": 6.693734169006348, "learning_rate": 1.2965542222436287e-06, "loss": 4.1276, "step": 1112400 }, { "epoch": 6.7078906429010745, "grad_norm": 4.7116546630859375, "learning_rate": 1.2938858655685614e-06, "loss": 3.8584, "step": 1112450 }, { "epoch": 6.708192134682409, "grad_norm": 7.062678337097168, "learning_rate": 1.2912202456173082e-06, "loss": 4.5427, "step": 1112500 }, { "epoch": 6.708493626463743, "grad_norm": 5.501864433288574, "learning_rate": 1.2885573624389466e-06, "loss": 4.0361, "step": 1112550 }, { "epoch": 6.708795118245076, "grad_norm": 5.061567306518555, "learning_rate": 1.2858972160824376e-06, "loss": 4.3124, "step": 1112600 }, { "epoch": 6.7090966100264104, "grad_norm": 5.632868766784668, "learning_rate": 1.2832398065967752e-06, "loss": 4.2296, "step": 1112650 }, { "epoch": 6.709398101807745, "grad_norm": 3.2987821102142334, "learning_rate": 1.2805851340308705e-06, "loss": 4.2129, "step": 1112700 }, { "epoch": 6.709699593589079, "grad_norm": 4.051407337188721, "learning_rate": 1.277933198433534e-06, "loss": 4.1784, "step": 1112750 }, { "epoch": 6.710001085370413, "grad_norm": 4.117868900299072, "learning_rate": 1.2752839998536101e-06, "loss": 4.2644, "step": 1112800 }, { "epoch": 6.710302577151747, "grad_norm": 5.296352863311768, "learning_rate": 1.2726375383398602e-06, "loss": 4.0864, "step": 1112850 }, { "epoch": 6.710604068933081, "grad_norm": 8.687728881835938, "learning_rate": 1.269993813940945e-06, "loss": 3.9209, "step": 1112900 }, { "epoch": 6.710905560714415, "grad_norm": 5.354580402374268, "learning_rate": 1.2673528267055589e-06, "loss": 4.2916, "step": 1112950 }, { "epoch": 6.711207052495749, "grad_norm": 4.530486583709717, "learning_rate": 1.2647145766822963e-06, "loss": 4.2589, "step": 1113000 }, { "epoch": 6.711207052495749, "eval_loss": 4.78757905960083, "eval_runtime": 38.9777, "eval_samples_per_second": 13.136, "eval_steps_per_second": 6.568, "eval_tts_loss": 8.171921085043468, "step": 1113000 }, { "epoch": 6.711508544277083, "grad_norm": 4.178707122802734, "learning_rate": 1.2620790639196854e-06, "loss": 4.3588, "step": 1113050 }, { "epoch": 6.711810036058417, "grad_norm": 8.279898643493652, "learning_rate": 1.2594462884662537e-06, "loss": 4.3152, "step": 1113100 }, { "epoch": 6.712111527839751, "grad_norm": 6.364372253417969, "learning_rate": 1.2568162503704626e-06, "loss": 4.2247, "step": 1113150 }, { "epoch": 6.712413019621085, "grad_norm": 2.7821879386901855, "learning_rate": 1.2541889496806901e-06, "loss": 4.25, "step": 1113200 }, { "epoch": 6.712714511402419, "grad_norm": 4.776930332183838, "learning_rate": 1.2515643864452973e-06, "loss": 4.3226, "step": 1113250 }, { "epoch": 6.713016003183753, "grad_norm": 2.60374116897583, "learning_rate": 1.2489425607125957e-06, "loss": 4.1139, "step": 1113300 }, { "epoch": 6.713317494965088, "grad_norm": 6.691949844360352, "learning_rate": 1.2463234725308303e-06, "loss": 4.3429, "step": 1113350 }, { "epoch": 6.713618986746422, "grad_norm": 5.164791584014893, "learning_rate": 1.2437071219481953e-06, "loss": 4.1683, "step": 1113400 }, { "epoch": 6.713920478527755, "grad_norm": 7.894080638885498, "learning_rate": 1.241093509012836e-06, "loss": 3.937, "step": 1113450 }, { "epoch": 6.714221970309089, "grad_norm": 6.075294494628906, "learning_rate": 1.2384826337728637e-06, "loss": 4.3402, "step": 1113500 }, { "epoch": 6.7145234620904235, "grad_norm": 4.349011421203613, "learning_rate": 1.2358744962763234e-06, "loss": 4.2559, "step": 1113550 }, { "epoch": 6.714824953871758, "grad_norm": 4.494654178619385, "learning_rate": 1.23326909657121e-06, "loss": 4.1237, "step": 1113600 }, { "epoch": 6.715126445653091, "grad_norm": 6.364657402038574, "learning_rate": 1.2306664347054852e-06, "loss": 4.301, "step": 1113650 }, { "epoch": 6.715427937434425, "grad_norm": 4.909545421600342, "learning_rate": 1.2280665107270437e-06, "loss": 4.2702, "step": 1113700 }, { "epoch": 6.715729429215759, "grad_norm": 10.313638687133789, "learning_rate": 1.225469324683731e-06, "loss": 4.414, "step": 1113750 }, { "epoch": 6.716030920997094, "grad_norm": 7.201240062713623, "learning_rate": 1.2228748766233253e-06, "loss": 4.0552, "step": 1113800 }, { "epoch": 6.716332412778428, "grad_norm": 9.600903511047363, "learning_rate": 1.2202831665935885e-06, "loss": 4.2184, "step": 1113850 }, { "epoch": 6.716633904559762, "grad_norm": 2.861912250518799, "learning_rate": 1.217694194642249e-06, "loss": 4.4253, "step": 1113900 }, { "epoch": 6.716935396341095, "grad_norm": 5.188373565673828, "learning_rate": 1.2151079608169024e-06, "loss": 4.7231, "step": 1113950 }, { "epoch": 6.71723688812243, "grad_norm": 3.2459819316864014, "learning_rate": 1.21252446516516e-06, "loss": 4.1729, "step": 1114000 }, { "epoch": 6.717538379903764, "grad_norm": 7.311131954193115, "learning_rate": 1.2099437077346009e-06, "loss": 3.7938, "step": 1114050 }, { "epoch": 6.717839871685098, "grad_norm": 4.411768913269043, "learning_rate": 1.2073656885726702e-06, "loss": 3.9223, "step": 1114100 }, { "epoch": 6.718141363466432, "grad_norm": 4.189707279205322, "learning_rate": 1.2047904077268467e-06, "loss": 4.2335, "step": 1114150 }, { "epoch": 6.7184428552477655, "grad_norm": 4.738330841064453, "learning_rate": 1.2022178652445092e-06, "loss": 4.1391, "step": 1114200 }, { "epoch": 6.7187443470291, "grad_norm": 9.451645851135254, "learning_rate": 1.199648061173003e-06, "loss": 4.1151, "step": 1114250 }, { "epoch": 6.719045838810434, "grad_norm": 1.926942229270935, "learning_rate": 1.1970809955596238e-06, "loss": 4.1177, "step": 1114300 }, { "epoch": 6.719347330591768, "grad_norm": 4.98463249206543, "learning_rate": 1.1945166684516173e-06, "loss": 4.3258, "step": 1114350 }, { "epoch": 6.719648822373102, "grad_norm": 3.0772738456726074, "learning_rate": 1.1919550798961786e-06, "loss": 4.1489, "step": 1114400 }, { "epoch": 6.7199503141544366, "grad_norm": 5.533058166503906, "learning_rate": 1.189396229940437e-06, "loss": 4.2191, "step": 1114450 }, { "epoch": 6.72025180593577, "grad_norm": 5.457443714141846, "learning_rate": 1.1868401186315046e-06, "loss": 3.9888, "step": 1114500 }, { "epoch": 6.720553297717104, "grad_norm": 4.529794692993164, "learning_rate": 1.1842867460164274e-06, "loss": 4.1352, "step": 1114550 }, { "epoch": 6.720854789498438, "grad_norm": 7.038906574249268, "learning_rate": 1.181736112142151e-06, "loss": 4.3765, "step": 1114600 }, { "epoch": 6.7211562812797725, "grad_norm": 5.25425910949707, "learning_rate": 1.1791882170556543e-06, "loss": 4.2462, "step": 1114650 }, { "epoch": 6.721457773061107, "grad_norm": 6.0697431564331055, "learning_rate": 1.1766430608038334e-06, "loss": 4.2399, "step": 1114700 }, { "epoch": 6.72175926484244, "grad_norm": 4.694170951843262, "learning_rate": 1.1741006434335009e-06, "loss": 4.4609, "step": 1114750 }, { "epoch": 6.722060756623774, "grad_norm": 4.599701881408691, "learning_rate": 1.1715609649914525e-06, "loss": 4.1888, "step": 1114800 }, { "epoch": 6.722362248405108, "grad_norm": 5.972931385040283, "learning_rate": 1.169024025524451e-06, "loss": 4.3396, "step": 1114850 }, { "epoch": 6.722663740186443, "grad_norm": 5.5666375160217285, "learning_rate": 1.1664898250791754e-06, "loss": 4.5205, "step": 1114900 }, { "epoch": 6.722965231967777, "grad_norm": 4.1901469230651855, "learning_rate": 1.1639583637022386e-06, "loss": 4.3786, "step": 1114950 }, { "epoch": 6.723266723749111, "grad_norm": 4.4242706298828125, "learning_rate": 1.1614296414402535e-06, "loss": 4.2911, "step": 1115000 }, { "epoch": 6.723568215530444, "grad_norm": 2.7537078857421875, "learning_rate": 1.1589036583397493e-06, "loss": 4.3552, "step": 1115050 }, { "epoch": 6.723869707311779, "grad_norm": 5.827381610870361, "learning_rate": 1.1563804144472221e-06, "loss": 4.2097, "step": 1115100 }, { "epoch": 6.724171199093113, "grad_norm": 4.466580390930176, "learning_rate": 1.1538599098091018e-06, "loss": 4.2868, "step": 1115150 }, { "epoch": 6.724472690874447, "grad_norm": 6.6354756355285645, "learning_rate": 1.151342144471784e-06, "loss": 4.1442, "step": 1115200 }, { "epoch": 6.72477418265578, "grad_norm": 4.777061462402344, "learning_rate": 1.1488271184815823e-06, "loss": 4.3675, "step": 1115250 }, { "epoch": 6.7250756744371145, "grad_norm": 5.925576686859131, "learning_rate": 1.1463148318848092e-06, "loss": 4.2295, "step": 1115300 }, { "epoch": 6.725377166218449, "grad_norm": 5.76917028427124, "learning_rate": 1.1438052847276613e-06, "loss": 4.194, "step": 1115350 }, { "epoch": 6.725678657999783, "grad_norm": 2.4850761890411377, "learning_rate": 1.1412984770563682e-06, "loss": 4.105, "step": 1115400 }, { "epoch": 6.725980149781117, "grad_norm": 5.408670425415039, "learning_rate": 1.138794408917043e-06, "loss": 3.9294, "step": 1115450 }, { "epoch": 6.726281641562451, "grad_norm": 4.215576648712158, "learning_rate": 1.136293080355749e-06, "loss": 4.4285, "step": 1115500 }, { "epoch": 6.7265831333437855, "grad_norm": 6.720272064208984, "learning_rate": 1.1337944914185659e-06, "loss": 4.6196, "step": 1115550 }, { "epoch": 6.726884625125119, "grad_norm": 4.732380390167236, "learning_rate": 1.1312986421514402e-06, "loss": 4.2723, "step": 1115600 }, { "epoch": 6.727186116906453, "grad_norm": 4.291492462158203, "learning_rate": 1.1288055326003188e-06, "loss": 4.1884, "step": 1115650 }, { "epoch": 6.727487608687787, "grad_norm": 4.34268045425415, "learning_rate": 1.1263151628110811e-06, "loss": 4.055, "step": 1115700 }, { "epoch": 6.7277891004691215, "grad_norm": 6.806481838226318, "learning_rate": 1.1238275328295577e-06, "loss": 4.1064, "step": 1115750 }, { "epoch": 6.728090592250455, "grad_norm": 5.569710731506348, "learning_rate": 1.1213426427015282e-06, "loss": 4.1859, "step": 1115800 }, { "epoch": 6.728392084031789, "grad_norm": 4.974707126617432, "learning_rate": 1.118860492472723e-06, "loss": 4.1981, "step": 1115850 }, { "epoch": 6.728693575813123, "grad_norm": 9.332709312438965, "learning_rate": 1.1163810821888386e-06, "loss": 4.3547, "step": 1115900 }, { "epoch": 6.728995067594457, "grad_norm": 6.587496280670166, "learning_rate": 1.1139044118954888e-06, "loss": 4.4386, "step": 1115950 }, { "epoch": 6.729296559375792, "grad_norm": 4.612952709197998, "learning_rate": 1.1114304816382535e-06, "loss": 4.0899, "step": 1116000 }, { "epoch": 6.729296559375792, "eval_loss": 4.785619735717773, "eval_runtime": 39.0661, "eval_samples_per_second": 13.106, "eval_steps_per_second": 6.553, "eval_tts_loss": 8.168434092832355, "step": 1116000 }, { "epoch": 6.729598051157126, "grad_norm": 8.980507850646973, "learning_rate": 1.1089592914626633e-06, "loss": 4.3439, "step": 1116050 }, { "epoch": 6.729899542938459, "grad_norm": 9.700048446655273, "learning_rate": 1.106490841414198e-06, "loss": 4.4043, "step": 1116100 }, { "epoch": 6.730201034719793, "grad_norm": 3.410428047180176, "learning_rate": 1.1040251315382887e-06, "loss": 4.1694, "step": 1116150 }, { "epoch": 6.730502526501128, "grad_norm": 4.107187747955322, "learning_rate": 1.101562161880315e-06, "loss": 3.8337, "step": 1116200 }, { "epoch": 6.730804018282462, "grad_norm": 5.289811611175537, "learning_rate": 1.099101932485591e-06, "loss": 4.133, "step": 1116250 }, { "epoch": 6.731105510063796, "grad_norm": 5.158170223236084, "learning_rate": 1.096644443399397e-06, "loss": 4.1813, "step": 1116300 }, { "epoch": 6.731407001845129, "grad_norm": 6.4215989112854, "learning_rate": 1.0941896946669804e-06, "loss": 4.0933, "step": 1116350 }, { "epoch": 6.7317084936264635, "grad_norm": 4.418618679046631, "learning_rate": 1.0917376863335048e-06, "loss": 4.0685, "step": 1116400 }, { "epoch": 6.732009985407798, "grad_norm": 6.764793872833252, "learning_rate": 1.0892884184440841e-06, "loss": 4.302, "step": 1116450 }, { "epoch": 6.732311477189132, "grad_norm": 3.3543241024017334, "learning_rate": 1.0868418910437992e-06, "loss": 4.2655, "step": 1116500 }, { "epoch": 6.732612968970466, "grad_norm": 2.3182101249694824, "learning_rate": 1.0843981041776972e-06, "loss": 4.3786, "step": 1116550 }, { "epoch": 6.7329144607518, "grad_norm": 4.238580226898193, "learning_rate": 1.0819570578907089e-06, "loss": 4.4813, "step": 1116600 }, { "epoch": 6.733215952533134, "grad_norm": 4.241235733032227, "learning_rate": 1.0795187522277981e-06, "loss": 3.9564, "step": 1116650 }, { "epoch": 6.733517444314468, "grad_norm": 3.9430460929870605, "learning_rate": 1.0770831872338293e-06, "loss": 4.0571, "step": 1116700 }, { "epoch": 6.733818936095802, "grad_norm": 5.342256546020508, "learning_rate": 1.0746503629535996e-06, "loss": 4.2626, "step": 1116750 }, { "epoch": 6.734120427877136, "grad_norm": 4.257400989532471, "learning_rate": 1.0722202794319235e-06, "loss": 4.5, "step": 1116800 }, { "epoch": 6.7344219196584705, "grad_norm": 1.649375081062317, "learning_rate": 1.0697929367134983e-06, "loss": 4.1557, "step": 1116850 }, { "epoch": 6.734723411439804, "grad_norm": 4.06546688079834, "learning_rate": 1.0673683348429884e-06, "loss": 4.1858, "step": 1116900 }, { "epoch": 6.735024903221138, "grad_norm": 5.885489463806152, "learning_rate": 1.0649464738650416e-06, "loss": 4.1038, "step": 1116950 }, { "epoch": 6.735326395002472, "grad_norm": 2.4062466621398926, "learning_rate": 1.0625273538242052e-06, "loss": 4.1547, "step": 1117000 }, { "epoch": 6.735627886783806, "grad_norm": 2.7213165760040283, "learning_rate": 1.0601109747649938e-06, "loss": 4.1995, "step": 1117050 }, { "epoch": 6.735929378565141, "grad_norm": 4.4412994384765625, "learning_rate": 1.0576973367319052e-06, "loss": 4.2552, "step": 1117100 }, { "epoch": 6.736230870346475, "grad_norm": 2.340718984603882, "learning_rate": 1.0552864397693538e-06, "loss": 3.9545, "step": 1117150 }, { "epoch": 6.736532362127808, "grad_norm": 7.8115153312683105, "learning_rate": 1.0528782839216877e-06, "loss": 3.9848, "step": 1117200 }, { "epoch": 6.736833853909142, "grad_norm": 6.681029319763184, "learning_rate": 1.0504728692332376e-06, "loss": 4.1483, "step": 1117250 }, { "epoch": 6.737135345690477, "grad_norm": 4.604473114013672, "learning_rate": 1.0480701957482852e-06, "loss": 4.1076, "step": 1117300 }, { "epoch": 6.737436837471811, "grad_norm": 4.840397357940674, "learning_rate": 1.0456702635110114e-06, "loss": 4.2698, "step": 1117350 }, { "epoch": 6.737738329253144, "grad_norm": 4.8976311683654785, "learning_rate": 1.0432730725656312e-06, "loss": 4.3641, "step": 1117400 }, { "epoch": 6.738039821034478, "grad_norm": 7.207669734954834, "learning_rate": 1.0408786229562094e-06, "loss": 3.8866, "step": 1117450 }, { "epoch": 6.7383413128158125, "grad_norm": 2.8549561500549316, "learning_rate": 1.0384869147268604e-06, "loss": 3.9823, "step": 1117500 }, { "epoch": 6.738642804597147, "grad_norm": 7.180399417877197, "learning_rate": 1.0360979479215824e-06, "loss": 4.2146, "step": 1117550 }, { "epoch": 6.738944296378481, "grad_norm": 6.216005325317383, "learning_rate": 1.0337117225843239e-06, "loss": 3.8298, "step": 1117600 }, { "epoch": 6.739245788159815, "grad_norm": 4.212980270385742, "learning_rate": 1.0313282387590327e-06, "loss": 4.2166, "step": 1117650 }, { "epoch": 6.739547279941148, "grad_norm": 3.9364562034606934, "learning_rate": 1.0289474964895405e-06, "loss": 3.9442, "step": 1117700 }, { "epoch": 6.739848771722483, "grad_norm": 5.32330846786499, "learning_rate": 1.0265694958196956e-06, "loss": 4.4204, "step": 1117750 }, { "epoch": 6.740150263503817, "grad_norm": 7.582840919494629, "learning_rate": 1.02419423679323e-06, "loss": 3.8927, "step": 1117800 }, { "epoch": 6.740451755285151, "grad_norm": 7.090994834899902, "learning_rate": 1.021821719453858e-06, "loss": 4.3469, "step": 1117850 }, { "epoch": 6.740753247066485, "grad_norm": 4.22934103012085, "learning_rate": 1.0194519438452785e-06, "loss": 4.1483, "step": 1117900 }, { "epoch": 6.741054738847819, "grad_norm": 5.126598834991455, "learning_rate": 1.0170849100110568e-06, "loss": 4.262, "step": 1117950 }, { "epoch": 6.741356230629153, "grad_norm": 4.361120223999023, "learning_rate": 1.0147206179947908e-06, "loss": 4.4954, "step": 1118000 }, { "epoch": 6.741657722410487, "grad_norm": 5.876070976257324, "learning_rate": 1.0123590678399796e-06, "loss": 4.3209, "step": 1118050 }, { "epoch": 6.741959214191821, "grad_norm": 5.0943708419799805, "learning_rate": 1.0100002595900714e-06, "loss": 4.1758, "step": 1118100 }, { "epoch": 6.742260705973155, "grad_norm": 6.831508636474609, "learning_rate": 1.0076441932884982e-06, "loss": 4.406, "step": 1118150 }, { "epoch": 6.74256219775449, "grad_norm": 5.512884616851807, "learning_rate": 1.0052908689786088e-06, "loss": 4.392, "step": 1118200 }, { "epoch": 6.742863689535823, "grad_norm": 7.916874408721924, "learning_rate": 1.0029402867037018e-06, "loss": 4.5668, "step": 1118250 }, { "epoch": 6.743165181317157, "grad_norm": 7.150059223175049, "learning_rate": 1.0005924465070592e-06, "loss": 4.2979, "step": 1118300 }, { "epoch": 6.743466673098491, "grad_norm": 6.658560276031494, "learning_rate": 9.9824734843188e-07, "loss": 4.1982, "step": 1118350 }, { "epoch": 6.7437681648798256, "grad_norm": 3.613985300064087, "learning_rate": 9.959049925213293e-07, "loss": 4.3289, "step": 1118400 }, { "epoch": 6.74406965666116, "grad_norm": 6.791964530944824, "learning_rate": 9.935653788184895e-07, "loss": 3.9258, "step": 1118450 }, { "epoch": 6.744371148442493, "grad_norm": 8.57235050201416, "learning_rate": 9.912285073664594e-07, "loss": 4.2731, "step": 1118500 }, { "epoch": 6.744672640223827, "grad_norm": 4.841362953186035, "learning_rate": 9.888943782082048e-07, "loss": 3.9896, "step": 1118550 }, { "epoch": 6.7449741320051615, "grad_norm": 4.632920742034912, "learning_rate": 9.865629913866912e-07, "loss": 4.083, "step": 1118600 }, { "epoch": 6.745275623786496, "grad_norm": 4.282576084136963, "learning_rate": 9.84234346944851e-07, "loss": 4.3133, "step": 1118650 }, { "epoch": 6.74557711556783, "grad_norm": 5.814764022827148, "learning_rate": 9.819084449255e-07, "loss": 4.0451, "step": 1118700 }, { "epoch": 6.745878607349164, "grad_norm": 5.053502559661865, "learning_rate": 9.795852853714703e-07, "loss": 4.4723, "step": 1118750 }, { "epoch": 6.746180099130497, "grad_norm": 3.5821685791015625, "learning_rate": 9.772648683255279e-07, "loss": 4.1338, "step": 1118800 }, { "epoch": 6.746481590911832, "grad_norm": 4.028204441070557, "learning_rate": 9.749471938303387e-07, "loss": 4.4626, "step": 1118850 }, { "epoch": 6.746783082693166, "grad_norm": 4.9658403396606445, "learning_rate": 9.726322619285853e-07, "loss": 3.8694, "step": 1118900 }, { "epoch": 6.7470845744745, "grad_norm": 3.9696977138519287, "learning_rate": 9.703200726628668e-07, "loss": 4.1299, "step": 1118950 }, { "epoch": 6.747386066255833, "grad_norm": 4.212327003479004, "learning_rate": 9.68010626075716e-07, "loss": 4.074, "step": 1119000 }, { "epoch": 6.747386066255833, "eval_loss": 4.7859206199646, "eval_runtime": 39.2856, "eval_samples_per_second": 13.033, "eval_steps_per_second": 6.516, "eval_tts_loss": 8.171417114082367, "step": 1119000 }, { "epoch": 6.747687558037168, "grad_norm": 2.007598876953125, "learning_rate": 9.657039222096484e-07, "loss": 3.634, "step": 1119050 }, { "epoch": 6.747989049818502, "grad_norm": 4.041232585906982, "learning_rate": 9.633999611071307e-07, "loss": 3.9514, "step": 1119100 }, { "epoch": 6.748290541599836, "grad_norm": 7.200584411621094, "learning_rate": 9.610987428105454e-07, "loss": 4.5514, "step": 1119150 }, { "epoch": 6.74859203338117, "grad_norm": 4.782243728637695, "learning_rate": 9.58800267362242e-07, "loss": 3.9928, "step": 1119200 }, { "epoch": 6.748893525162504, "grad_norm": 5.049190044403076, "learning_rate": 9.565045348045365e-07, "loss": 4.2731, "step": 1119250 }, { "epoch": 6.749195016943839, "grad_norm": 3.6003384590148926, "learning_rate": 9.542115451796618e-07, "loss": 4.2779, "step": 1119300 }, { "epoch": 6.749496508725172, "grad_norm": 5.080846309661865, "learning_rate": 9.519212985298174e-07, "loss": 4.2173, "step": 1119350 }, { "epoch": 6.749798000506506, "grad_norm": 5.733604907989502, "learning_rate": 9.4963379489717e-07, "loss": 3.5632, "step": 1119400 }, { "epoch": 6.75009949228784, "grad_norm": 4.247810363769531, "learning_rate": 9.473490343238022e-07, "loss": 4.5294, "step": 1119450 }, { "epoch": 6.7504009840691745, "grad_norm": 9.487239837646484, "learning_rate": 9.450670168517638e-07, "loss": 4.0438, "step": 1119500 }, { "epoch": 6.750702475850508, "grad_norm": 5.761942386627197, "learning_rate": 9.427877425230546e-07, "loss": 3.8732, "step": 1119550 }, { "epoch": 6.751003967631842, "grad_norm": 4.902900695800781, "learning_rate": 9.405112113796242e-07, "loss": 4.0179, "step": 1119600 }, { "epoch": 6.751305459413176, "grad_norm": 4.698395729064941, "learning_rate": 9.382374234633727e-07, "loss": 4.3372, "step": 1119650 }, { "epoch": 6.7516069511945105, "grad_norm": 6.364431381225586, "learning_rate": 9.359663788161331e-07, "loss": 4.3808, "step": 1119700 }, { "epoch": 6.751908442975845, "grad_norm": 4.227256774902344, "learning_rate": 9.336980774797054e-07, "loss": 4.1205, "step": 1119750 }, { "epoch": 6.752209934757179, "grad_norm": 3.3247933387756348, "learning_rate": 9.314325194958228e-07, "loss": 4.406, "step": 1119800 }, { "epoch": 6.752511426538512, "grad_norm": 1.891334056854248, "learning_rate": 9.291697049062019e-07, "loss": 4.0594, "step": 1119850 }, { "epoch": 6.752812918319846, "grad_norm": 5.003391265869141, "learning_rate": 9.269096337524762e-07, "loss": 4.4343, "step": 1119900 }, { "epoch": 6.753114410101181, "grad_norm": 5.332914352416992, "learning_rate": 9.246523060762457e-07, "loss": 3.9684, "step": 1119950 }, { "epoch": 6.753415901882515, "grad_norm": 5.331775665283203, "learning_rate": 9.223977219190437e-07, "loss": 4.2348, "step": 1120000 }, { "epoch": 6.753717393663849, "grad_norm": 2.377657175064087, "learning_rate": 9.201458813223706e-07, "loss": 4.4688, "step": 1120050 }, { "epoch": 6.754018885445182, "grad_norm": 3.3239874839782715, "learning_rate": 9.178967843276597e-07, "loss": 4.277, "step": 1120100 }, { "epoch": 6.754320377226517, "grad_norm": 5.130308628082275, "learning_rate": 9.156504309763113e-07, "loss": 4.2029, "step": 1120150 }, { "epoch": 6.754621869007851, "grad_norm": 3.6314456462860107, "learning_rate": 9.13406821309659e-07, "loss": 4.1754, "step": 1120200 }, { "epoch": 6.754923360789185, "grad_norm": 4.072171688079834, "learning_rate": 9.111659553689865e-07, "loss": 3.9097, "step": 1120250 }, { "epoch": 6.755224852570519, "grad_norm": 7.038859844207764, "learning_rate": 9.089278331955607e-07, "loss": 4.3462, "step": 1120300 }, { "epoch": 6.755526344351853, "grad_norm": 1.2501896619796753, "learning_rate": 9.066924548305487e-07, "loss": 3.8901, "step": 1120350 }, { "epoch": 6.755827836133187, "grad_norm": 2.1439313888549805, "learning_rate": 9.044598203150843e-07, "loss": 4.1746, "step": 1120400 }, { "epoch": 6.756129327914521, "grad_norm": 1.2163856029510498, "learning_rate": 9.02229929690268e-07, "loss": 3.8913, "step": 1120450 }, { "epoch": 6.756430819695855, "grad_norm": 5.447388172149658, "learning_rate": 9.000027829971502e-07, "loss": 3.9932, "step": 1120500 }, { "epoch": 6.756732311477189, "grad_norm": 11.257342338562012, "learning_rate": 8.977783802766981e-07, "loss": 4.2697, "step": 1120550 }, { "epoch": 6.7570338032585235, "grad_norm": 2.7321035861968994, "learning_rate": 8.955567215698456e-07, "loss": 4.1406, "step": 1120600 }, { "epoch": 6.757335295039857, "grad_norm": 3.434267520904541, "learning_rate": 8.933378069174935e-07, "loss": 4.1485, "step": 1120650 }, { "epoch": 6.757636786821191, "grad_norm": 6.860990524291992, "learning_rate": 8.911216363604756e-07, "loss": 4.2263, "step": 1120700 }, { "epoch": 6.757938278602525, "grad_norm": 4.085300922393799, "learning_rate": 8.889082099395762e-07, "loss": 3.8212, "step": 1120750 }, { "epoch": 6.7582397703838595, "grad_norm": 5.024570941925049, "learning_rate": 8.866975276955124e-07, "loss": 4.1962, "step": 1120800 }, { "epoch": 6.758541262165194, "grad_norm": 4.675129413604736, "learning_rate": 8.844895896690019e-07, "loss": 4.4996, "step": 1120850 }, { "epoch": 6.758842753946528, "grad_norm": 5.849661827087402, "learning_rate": 8.82284395900662e-07, "loss": 4.1244, "step": 1120900 }, { "epoch": 6.759144245727861, "grad_norm": 4.7461628913879395, "learning_rate": 8.800819464310605e-07, "loss": 4.2667, "step": 1120950 }, { "epoch": 6.759445737509195, "grad_norm": 5.804567813873291, "learning_rate": 8.778822413007647e-07, "loss": 4.2866, "step": 1121000 }, { "epoch": 6.75974722929053, "grad_norm": 4.587632179260254, "learning_rate": 8.75685280550209e-07, "loss": 4.1037, "step": 1121050 }, { "epoch": 6.760048721071864, "grad_norm": 2.0292766094207764, "learning_rate": 8.734910642198778e-07, "loss": 4.5907, "step": 1121100 }, { "epoch": 6.760350212853197, "grad_norm": 6.118147373199463, "learning_rate": 8.71299592350122e-07, "loss": 4.0371, "step": 1121150 }, { "epoch": 6.760651704634531, "grad_norm": 4.4962944984436035, "learning_rate": 8.691108649812761e-07, "loss": 4.2091, "step": 1121200 }, { "epoch": 6.760953196415866, "grad_norm": 4.294684886932373, "learning_rate": 8.669248821536245e-07, "loss": 4.0174, "step": 1121250 }, { "epoch": 6.7612546881972, "grad_norm": 3.937161445617676, "learning_rate": 8.647416439074018e-07, "loss": 4.175, "step": 1121300 }, { "epoch": 6.761556179978534, "grad_norm": 4.568406105041504, "learning_rate": 8.625611502827756e-07, "loss": 4.3141, "step": 1121350 }, { "epoch": 6.761857671759868, "grad_norm": 3.0549933910369873, "learning_rate": 8.603834013198973e-07, "loss": 3.9344, "step": 1121400 }, { "epoch": 6.7621591635412015, "grad_norm": 4.032445430755615, "learning_rate": 8.582083970588016e-07, "loss": 4.32, "step": 1121450 }, { "epoch": 6.762460655322536, "grad_norm": 4.838669776916504, "learning_rate": 8.560361375395563e-07, "loss": 3.9182, "step": 1121500 }, { "epoch": 6.76276214710387, "grad_norm": 6.204533576965332, "learning_rate": 8.538666228021295e-07, "loss": 4.5422, "step": 1121550 }, { "epoch": 6.763063638885204, "grad_norm": 6.066364288330078, "learning_rate": 8.516998528864394e-07, "loss": 4.0799, "step": 1121600 }, { "epoch": 6.763365130666538, "grad_norm": 3.928757429122925, "learning_rate": 8.495358278323705e-07, "loss": 4.4809, "step": 1121650 }, { "epoch": 6.763666622447872, "grad_norm": 10.268730163574219, "learning_rate": 8.473745476797577e-07, "loss": 4.2969, "step": 1121700 }, { "epoch": 6.763968114229206, "grad_norm": 2.172266721725464, "learning_rate": 8.45216012468336e-07, "loss": 4.0459, "step": 1121750 }, { "epoch": 6.76426960601054, "grad_norm": 4.84132719039917, "learning_rate": 8.430602222378735e-07, "loss": 4.0028, "step": 1121800 }, { "epoch": 6.764571097791874, "grad_norm": 5.888508319854736, "learning_rate": 8.409071770280384e-07, "loss": 4.0794, "step": 1121850 }, { "epoch": 6.7648725895732085, "grad_norm": 5.167423725128174, "learning_rate": 8.387568768784325e-07, "loss": 4.4836, "step": 1121900 }, { "epoch": 6.765174081354543, "grad_norm": 5.915904521942139, "learning_rate": 8.366093218286407e-07, "loss": 4.3516, "step": 1121950 }, { "epoch": 6.765475573135876, "grad_norm": 4.185784816741943, "learning_rate": 8.34464511918198e-07, "loss": 4.4532, "step": 1122000 }, { "epoch": 6.765475573135876, "eval_loss": 4.784151554107666, "eval_runtime": 39.2699, "eval_samples_per_second": 13.038, "eval_steps_per_second": 6.519, "eval_tts_loss": 8.172100114071627, "step": 1122000 }, { "epoch": 6.76577706491721, "grad_norm": 6.925915718078613, "learning_rate": 8.323224471865564e-07, "loss": 4.5161, "step": 1122050 }, { "epoch": 6.766078556698544, "grad_norm": 4.446605682373047, "learning_rate": 8.301831276731507e-07, "loss": 4.1188, "step": 1122100 }, { "epoch": 6.766380048479879, "grad_norm": 8.8823823928833, "learning_rate": 8.280465534173497e-07, "loss": 4.1576, "step": 1122150 }, { "epoch": 6.766681540261213, "grad_norm": 5.295784950256348, "learning_rate": 8.259127244584551e-07, "loss": 4.061, "step": 1122200 }, { "epoch": 6.766983032042546, "grad_norm": 5.7792277336120605, "learning_rate": 8.237816408357689e-07, "loss": 4.3497, "step": 1122250 }, { "epoch": 6.76728452382388, "grad_norm": 6.0654616355896, "learning_rate": 8.216533025884931e-07, "loss": 4.1233, "step": 1122300 }, { "epoch": 6.7675860156052146, "grad_norm": 5.663399696350098, "learning_rate": 8.195277097557962e-07, "loss": 4.0762, "step": 1122350 }, { "epoch": 6.767887507386549, "grad_norm": 6.0994110107421875, "learning_rate": 8.174048623768137e-07, "loss": 4.4173, "step": 1122400 }, { "epoch": 6.768188999167883, "grad_norm": 4.178627967834473, "learning_rate": 8.15284760490581e-07, "loss": 4.1243, "step": 1122450 }, { "epoch": 6.768490490949217, "grad_norm": 6.047255992889404, "learning_rate": 8.131674041361502e-07, "loss": 4.3551, "step": 1122500 }, { "epoch": 6.7687919827305505, "grad_norm": 3.970923662185669, "learning_rate": 8.110527933524569e-07, "loss": 4.1282, "step": 1122550 }, { "epoch": 6.769093474511885, "grad_norm": 3.4188010692596436, "learning_rate": 8.089409281784531e-07, "loss": 4.2122, "step": 1122600 }, { "epoch": 6.769394966293219, "grad_norm": 5.386687755584717, "learning_rate": 8.068318086529745e-07, "loss": 4.0451, "step": 1122650 }, { "epoch": 6.769696458074553, "grad_norm": 5.345402717590332, "learning_rate": 8.047254348148402e-07, "loss": 3.9842, "step": 1122700 }, { "epoch": 6.769997949855886, "grad_norm": 4.4702372550964355, "learning_rate": 8.02621806702819e-07, "loss": 4.1898, "step": 1122750 }, { "epoch": 6.770299441637221, "grad_norm": 6.0677876472473145, "learning_rate": 8.005209243556298e-07, "loss": 4.3166, "step": 1122800 }, { "epoch": 6.770600933418555, "grad_norm": 4.846196174621582, "learning_rate": 7.98422787811942e-07, "loss": 4.133, "step": 1122850 }, { "epoch": 6.770902425199889, "grad_norm": 6.778266906738281, "learning_rate": 7.963273971103579e-07, "loss": 3.9054, "step": 1122900 }, { "epoch": 6.771203916981223, "grad_norm": 4.94895601272583, "learning_rate": 7.942347522894465e-07, "loss": 4.1842, "step": 1122950 }, { "epoch": 6.7715054087625575, "grad_norm": 4.598742961883545, "learning_rate": 7.921448533877106e-07, "loss": 4.1297, "step": 1123000 }, { "epoch": 6.771806900543892, "grad_norm": 9.623598098754883, "learning_rate": 7.900577004436193e-07, "loss": 4.5081, "step": 1123050 }, { "epoch": 6.772108392325225, "grad_norm": 4.125532627105713, "learning_rate": 7.879732934955752e-07, "loss": 3.7476, "step": 1123100 }, { "epoch": 6.772409884106559, "grad_norm": 4.4630513191223145, "learning_rate": 7.858916325819475e-07, "loss": 3.9024, "step": 1123150 }, { "epoch": 6.772711375887893, "grad_norm": 5.557124614715576, "learning_rate": 7.838127177410558e-07, "loss": 4.3644, "step": 1123200 }, { "epoch": 6.773012867669228, "grad_norm": 3.45969820022583, "learning_rate": 7.81736549011136e-07, "loss": 4.2085, "step": 1123250 }, { "epoch": 6.773314359450561, "grad_norm": 7.644756317138672, "learning_rate": 7.796631264304242e-07, "loss": 4.4255, "step": 1123300 }, { "epoch": 6.773615851231895, "grad_norm": 4.425117015838623, "learning_rate": 7.775924500370567e-07, "loss": 4.1113, "step": 1123350 }, { "epoch": 6.773917343013229, "grad_norm": 3.620912551879883, "learning_rate": 7.755245198691362e-07, "loss": 4.3325, "step": 1123400 }, { "epoch": 6.7742188347945635, "grad_norm": 3.214716672897339, "learning_rate": 7.734593359647323e-07, "loss": 4.509, "step": 1123450 }, { "epoch": 6.774520326575898, "grad_norm": 3.711059093475342, "learning_rate": 7.713968983618646e-07, "loss": 3.9447, "step": 1123500 }, { "epoch": 6.774821818357232, "grad_norm": 6.14052152633667, "learning_rate": 7.693372070984693e-07, "loss": 4.4943, "step": 1123550 }, { "epoch": 6.775123310138565, "grad_norm": 6.2478508949279785, "learning_rate": 7.672802622124663e-07, "loss": 3.8365, "step": 1123600 }, { "epoch": 6.7754248019198995, "grad_norm": 4.230034351348877, "learning_rate": 7.652260637417085e-07, "loss": 4.2999, "step": 1123650 }, { "epoch": 6.775726293701234, "grad_norm": 5.397642612457275, "learning_rate": 7.631746117239823e-07, "loss": 4.2853, "step": 1123700 }, { "epoch": 6.776027785482568, "grad_norm": 5.070160388946533, "learning_rate": 7.611259061970576e-07, "loss": 4.1162, "step": 1123750 }, { "epoch": 6.776329277263902, "grad_norm": 6.320396423339844, "learning_rate": 7.590799471986375e-07, "loss": 4.0319, "step": 1123800 }, { "epoch": 6.776630769045235, "grad_norm": 5.019259452819824, "learning_rate": 7.570367347663753e-07, "loss": 4.1083, "step": 1123850 }, { "epoch": 6.77693226082657, "grad_norm": 5.634626865386963, "learning_rate": 7.549962689378575e-07, "loss": 4.5399, "step": 1123900 }, { "epoch": 6.777233752607904, "grad_norm": 6.199408054351807, "learning_rate": 7.529585497506707e-07, "loss": 3.5119, "step": 1123950 }, { "epoch": 6.777535244389238, "grad_norm": 4.1196465492248535, "learning_rate": 7.509235772422683e-07, "loss": 3.7417, "step": 1124000 }, { "epoch": 6.777836736170572, "grad_norm": 4.632112979888916, "learning_rate": 7.488913514501538e-07, "loss": 4.5238, "step": 1124050 }, { "epoch": 6.7781382279519065, "grad_norm": 5.888916492462158, "learning_rate": 7.46861872411697e-07, "loss": 4.1501, "step": 1124100 }, { "epoch": 6.77843971973324, "grad_norm": 4.812871932983398, "learning_rate": 7.44835140164235e-07, "loss": 4.3359, "step": 1124150 }, { "epoch": 6.778741211514574, "grad_norm": 4.0963873863220215, "learning_rate": 7.428111547450877e-07, "loss": 4.2842, "step": 1124200 }, { "epoch": 6.779042703295908, "grad_norm": 5.210516929626465, "learning_rate": 7.407899161915088e-07, "loss": 4.0874, "step": 1124250 }, { "epoch": 6.779344195077242, "grad_norm": 6.910029888153076, "learning_rate": 7.387714245406851e-07, "loss": 4.097, "step": 1124300 }, { "epoch": 6.779645686858577, "grad_norm": 3.179661750793457, "learning_rate": 7.367556798297536e-07, "loss": 3.9961, "step": 1124350 }, { "epoch": 6.77994717863991, "grad_norm": 4.209898948669434, "learning_rate": 7.347426820958513e-07, "loss": 3.8935, "step": 1124400 }, { "epoch": 6.780248670421244, "grad_norm": 5.591691493988037, "learning_rate": 7.327324313759653e-07, "loss": 4.3837, "step": 1124450 }, { "epoch": 6.780550162202578, "grad_norm": 4.502143859863281, "learning_rate": 7.307249277071492e-07, "loss": 4.04, "step": 1124500 }, { "epoch": 6.7808516539839125, "grad_norm": 6.982682704925537, "learning_rate": 7.287201711263069e-07, "loss": 4.0338, "step": 1124550 }, { "epoch": 6.781153145765247, "grad_norm": 6.681665897369385, "learning_rate": 7.267181616703588e-07, "loss": 4.3329, "step": 1124600 }, { "epoch": 6.781454637546581, "grad_norm": 6.529983997344971, "learning_rate": 7.247188993761255e-07, "loss": 4.141, "step": 1124650 }, { "epoch": 6.781756129327914, "grad_norm": 7.275924205780029, "learning_rate": 7.227223842804109e-07, "loss": 4.024, "step": 1124700 }, { "epoch": 6.7820576211092485, "grad_norm": 9.408448219299316, "learning_rate": 7.207286164199855e-07, "loss": 4.0787, "step": 1124750 }, { "epoch": 6.782359112890583, "grad_norm": 4.494217395782471, "learning_rate": 7.187375958314867e-07, "loss": 4.2898, "step": 1124800 }, { "epoch": 6.782660604671917, "grad_norm": 5.7524261474609375, "learning_rate": 7.167493225516185e-07, "loss": 4.1819, "step": 1124850 }, { "epoch": 6.78296209645325, "grad_norm": 4.334460735321045, "learning_rate": 7.147637966169184e-07, "loss": 4.1287, "step": 1124900 }, { "epoch": 6.783263588234584, "grad_norm": 4.589419841766357, "learning_rate": 7.127810180639571e-07, "loss": 4.3207, "step": 1124950 }, { "epoch": 6.783565080015919, "grad_norm": 5.117674827575684, "learning_rate": 7.108009869292053e-07, "loss": 3.993, "step": 1125000 }, { "epoch": 6.783565080015919, "eval_loss": 4.783478736877441, "eval_runtime": 38.9926, "eval_samples_per_second": 13.131, "eval_steps_per_second": 6.565, "eval_tts_loss": 8.168579214546057, "step": 1125000 }, { "epoch": 6.783866571797253, "grad_norm": 5.051693439483643, "learning_rate": 7.08823703249134e-07, "loss": 4.0644, "step": 1125050 }, { "epoch": 6.784168063578587, "grad_norm": 2.979086399078369, "learning_rate": 7.068491670600973e-07, "loss": 4.2428, "step": 1125100 }, { "epoch": 6.784469555359921, "grad_norm": 6.684016227722168, "learning_rate": 7.048773783984496e-07, "loss": 4.36, "step": 1125150 }, { "epoch": 6.784771047141255, "grad_norm": 8.357562065124512, "learning_rate": 7.029083373004784e-07, "loss": 4.0703, "step": 1125200 }, { "epoch": 6.785072538922589, "grad_norm": 4.535447597503662, "learning_rate": 7.009420438024382e-07, "loss": 4.4554, "step": 1125250 }, { "epoch": 6.785374030703923, "grad_norm": 5.822906970977783, "learning_rate": 6.989784979404834e-07, "loss": 4.0416, "step": 1125300 }, { "epoch": 6.785675522485257, "grad_norm": 6.0470452308654785, "learning_rate": 6.970176997507682e-07, "loss": 4.2223, "step": 1125350 }, { "epoch": 6.785977014266591, "grad_norm": 9.935749053955078, "learning_rate": 6.950596492693639e-07, "loss": 4.2051, "step": 1125400 }, { "epoch": 6.786278506047925, "grad_norm": 3.650756597518921, "learning_rate": 6.931043465323249e-07, "loss": 4.1963, "step": 1125450 }, { "epoch": 6.786579997829259, "grad_norm": 5.273577690124512, "learning_rate": 6.911517915756392e-07, "loss": 4.3535, "step": 1125500 }, { "epoch": 6.786881489610593, "grad_norm": 5.275230407714844, "learning_rate": 6.892019844352115e-07, "loss": 4.2258, "step": 1125550 }, { "epoch": 6.787182981391927, "grad_norm": 6.097180366516113, "learning_rate": 6.872549251469462e-07, "loss": 4.4063, "step": 1125600 }, { "epoch": 6.7874844731732615, "grad_norm": 4.719842433929443, "learning_rate": 6.853106137466812e-07, "loss": 3.9614, "step": 1125650 }, { "epoch": 6.787785964954596, "grad_norm": 5.054831504821777, "learning_rate": 6.833690502702049e-07, "loss": 4.3177, "step": 1125700 }, { "epoch": 6.788087456735929, "grad_norm": 3.5673065185546875, "learning_rate": 6.814302347532052e-07, "loss": 4.1106, "step": 1125750 }, { "epoch": 6.788388948517263, "grad_norm": 5.453885078430176, "learning_rate": 6.794941672314202e-07, "loss": 4.3847, "step": 1125800 }, { "epoch": 6.7886904402985975, "grad_norm": 6.634341239929199, "learning_rate": 6.775608477404215e-07, "loss": 4.1211, "step": 1125850 }, { "epoch": 6.788991932079932, "grad_norm": 4.181068420410156, "learning_rate": 6.756302763158472e-07, "loss": 4.004, "step": 1125900 }, { "epoch": 6.789293423861266, "grad_norm": 3.988386631011963, "learning_rate": 6.737024529931856e-07, "loss": 4.2484, "step": 1125950 }, { "epoch": 6.789594915642599, "grad_norm": 5.357354640960693, "learning_rate": 6.717773778079416e-07, "loss": 4.0652, "step": 1126000 }, { "epoch": 6.789896407423933, "grad_norm": 3.8291101455688477, "learning_rate": 6.698550507955369e-07, "loss": 4.2682, "step": 1126050 }, { "epoch": 6.790197899205268, "grad_norm": 5.291939735412598, "learning_rate": 6.679354719913266e-07, "loss": 4.2558, "step": 1126100 }, { "epoch": 6.790499390986602, "grad_norm": 7.460807800292969, "learning_rate": 6.660186414306657e-07, "loss": 3.9963, "step": 1126150 }, { "epoch": 6.790800882767936, "grad_norm": 3.323835611343384, "learning_rate": 6.641045591488259e-07, "loss": 4.3535, "step": 1126200 }, { "epoch": 6.79110237454927, "grad_norm": 6.136021614074707, "learning_rate": 6.621932251810125e-07, "loss": 4.3709, "step": 1126250 }, { "epoch": 6.791403866330604, "grad_norm": 1.4474811553955078, "learning_rate": 6.602846395624306e-07, "loss": 3.9828, "step": 1126300 }, { "epoch": 6.791705358111938, "grad_norm": 6.027258396148682, "learning_rate": 6.583788023281856e-07, "loss": 4.6069, "step": 1126350 }, { "epoch": 6.792006849893272, "grad_norm": 4.547706127166748, "learning_rate": 6.564757135133492e-07, "loss": 4.2714, "step": 1126400 }, { "epoch": 6.792308341674606, "grad_norm": 4.574931621551514, "learning_rate": 6.545753731529601e-07, "loss": 4.3035, "step": 1126450 }, { "epoch": 6.7926098334559395, "grad_norm": 5.975889205932617, "learning_rate": 6.526777812819906e-07, "loss": 3.9916, "step": 1126500 }, { "epoch": 6.792911325237274, "grad_norm": 5.627161979675293, "learning_rate": 6.507829379353458e-07, "loss": 4.2672, "step": 1126550 }, { "epoch": 6.793212817018608, "grad_norm": 6.7906494140625, "learning_rate": 6.488908431479145e-07, "loss": 4.0349, "step": 1126600 }, { "epoch": 6.793514308799942, "grad_norm": 4.119060516357422, "learning_rate": 6.470014969545023e-07, "loss": 3.8115, "step": 1126650 }, { "epoch": 6.793815800581276, "grad_norm": 3.9345204830169678, "learning_rate": 6.45114899389898e-07, "loss": 4.3116, "step": 1126700 }, { "epoch": 6.7941172923626105, "grad_norm": 4.339435577392578, "learning_rate": 6.432310504888072e-07, "loss": 4.4724, "step": 1126750 }, { "epoch": 6.794418784143945, "grad_norm": 6.218080520629883, "learning_rate": 6.413499502859188e-07, "loss": 3.8935, "step": 1126800 }, { "epoch": 6.794720275925278, "grad_norm": 6.261852264404297, "learning_rate": 6.394715988158217e-07, "loss": 4.3417, "step": 1126850 }, { "epoch": 6.795021767706612, "grad_norm": 7.860723972320557, "learning_rate": 6.375959961131216e-07, "loss": 3.9879, "step": 1126900 }, { "epoch": 6.7953232594879465, "grad_norm": 4.440367221832275, "learning_rate": 6.35723142212291e-07, "loss": 4.6386, "step": 1126950 }, { "epoch": 6.795624751269281, "grad_norm": 4.18937873840332, "learning_rate": 6.338530371478356e-07, "loss": 4.2919, "step": 1127000 }, { "epoch": 6.795926243050614, "grad_norm": 3.9617760181427, "learning_rate": 6.319856809541612e-07, "loss": 4.3119, "step": 1127050 }, { "epoch": 6.796227734831948, "grad_norm": 4.142611026763916, "learning_rate": 6.301210736656071e-07, "loss": 3.9325, "step": 1127100 }, { "epoch": 6.796529226613282, "grad_norm": 4.059315204620361, "learning_rate": 6.282592153165456e-07, "loss": 4.2119, "step": 1127150 }, { "epoch": 6.796830718394617, "grad_norm": 3.786936044692993, "learning_rate": 6.264001059411827e-07, "loss": 4.3277, "step": 1127200 }, { "epoch": 6.797132210175951, "grad_norm": 5.080236911773682, "learning_rate": 6.245437455737745e-07, "loss": 4.5525, "step": 1127250 }, { "epoch": 6.797433701957285, "grad_norm": 7.139583110809326, "learning_rate": 6.226901342484603e-07, "loss": 4.176, "step": 1127300 }, { "epoch": 6.797735193738618, "grad_norm": 8.35720157623291, "learning_rate": 6.208392719993627e-07, "loss": 4.5496, "step": 1127350 }, { "epoch": 6.7980366855199525, "grad_norm": 7.8896660804748535, "learning_rate": 6.189911588605379e-07, "loss": 4.3716, "step": 1127400 }, { "epoch": 6.798338177301287, "grad_norm": 4.696540355682373, "learning_rate": 6.171457948660252e-07, "loss": 4.1804, "step": 1127450 }, { "epoch": 6.798639669082621, "grad_norm": 6.31701135635376, "learning_rate": 6.153031800497311e-07, "loss": 4.3736, "step": 1127500 }, { "epoch": 6.798941160863955, "grad_norm": 3.3036112785339355, "learning_rate": 6.134633144456114e-07, "loss": 4.2701, "step": 1127550 }, { "epoch": 6.7992426526452885, "grad_norm": 7.239552974700928, "learning_rate": 6.116261980875226e-07, "loss": 4.1068, "step": 1127600 }, { "epoch": 6.799544144426623, "grad_norm": 4.5721964836120605, "learning_rate": 6.097918310092709e-07, "loss": 4.4456, "step": 1127650 }, { "epoch": 6.799845636207957, "grad_norm": 5.678816318511963, "learning_rate": 6.079602132445793e-07, "loss": 4.2946, "step": 1127700 }, { "epoch": 6.800147127989291, "grad_norm": 3.3359570503234863, "learning_rate": 6.061313448272209e-07, "loss": 4.4912, "step": 1127750 }, { "epoch": 6.800448619770625, "grad_norm": 4.389930248260498, "learning_rate": 6.043052257907855e-07, "loss": 4.3139, "step": 1127800 }, { "epoch": 6.8007501115519595, "grad_norm": 5.380824565887451, "learning_rate": 6.024818561689293e-07, "loss": 4.144, "step": 1127850 }, { "epoch": 6.801051603333293, "grad_norm": 5.872720241546631, "learning_rate": 6.006612359951924e-07, "loss": 3.8031, "step": 1127900 }, { "epoch": 6.801353095114627, "grad_norm": 5.796357154846191, "learning_rate": 5.98843365303081e-07, "loss": 4.233, "step": 1127950 }, { "epoch": 6.801654586895961, "grad_norm": 7.202731609344482, "learning_rate": 5.970282441260521e-07, "loss": 4.2186, "step": 1128000 }, { "epoch": 6.801654586895961, "eval_loss": 4.783430099487305, "eval_runtime": 38.9081, "eval_samples_per_second": 13.159, "eval_steps_per_second": 6.58, "eval_tts_loss": 8.16273244900085, "step": 1128000 }, { "epoch": 6.8019560786772955, "grad_norm": 4.460678577423096, "learning_rate": 5.95215872497512e-07, "loss": 4.0687, "step": 1128050 }, { "epoch": 6.80225757045863, "grad_norm": 6.109364032745361, "learning_rate": 5.934062504508008e-07, "loss": 3.9617, "step": 1128100 }, { "epoch": 6.802559062239963, "grad_norm": 4.406163215637207, "learning_rate": 5.915993780192419e-07, "loss": 4.2558, "step": 1128150 }, { "epoch": 6.802860554021297, "grad_norm": 8.048897743225098, "learning_rate": 5.897952552360752e-07, "loss": 3.9326, "step": 1128200 }, { "epoch": 6.803162045802631, "grad_norm": 4.774560451507568, "learning_rate": 5.879938821345076e-07, "loss": 4.0902, "step": 1128250 }, { "epoch": 6.803463537583966, "grad_norm": 7.122232913970947, "learning_rate": 5.861952587476959e-07, "loss": 3.993, "step": 1128300 }, { "epoch": 6.8037650293653, "grad_norm": 5.396366119384766, "learning_rate": 5.843993851087303e-07, "loss": 4.0575, "step": 1128350 }, { "epoch": 6.804066521146634, "grad_norm": 4.047819137573242, "learning_rate": 5.826062612506676e-07, "loss": 4.2008, "step": 1128400 }, { "epoch": 6.804368012927967, "grad_norm": 5.413188457489014, "learning_rate": 5.808158872065149e-07, "loss": 3.9504, "step": 1128450 }, { "epoch": 6.8046695047093015, "grad_norm": 6.36270809173584, "learning_rate": 5.790282630091958e-07, "loss": 3.9041, "step": 1128500 }, { "epoch": 6.804970996490636, "grad_norm": 7.616918087005615, "learning_rate": 5.772433886916339e-07, "loss": 4.2768, "step": 1128550 }, { "epoch": 6.80527248827197, "grad_norm": 6.21148157119751, "learning_rate": 5.754612642866696e-07, "loss": 3.9799, "step": 1128600 }, { "epoch": 6.805573980053303, "grad_norm": 7.896574020385742, "learning_rate": 5.736818898270934e-07, "loss": 4.2339, "step": 1128650 }, { "epoch": 6.8058754718346375, "grad_norm": 4.978679656982422, "learning_rate": 5.71905265345679e-07, "loss": 4.3454, "step": 1128700 }, { "epoch": 6.806176963615972, "grad_norm": 4.574336051940918, "learning_rate": 5.701313908750837e-07, "loss": 3.9443, "step": 1128750 }, { "epoch": 6.806478455397306, "grad_norm": 4.328269004821777, "learning_rate": 5.683602664479814e-07, "loss": 4.6051, "step": 1128800 }, { "epoch": 6.80677994717864, "grad_norm": 5.1878814697265625, "learning_rate": 5.665918920969625e-07, "loss": 4.3156, "step": 1128850 }, { "epoch": 6.807081438959974, "grad_norm": 6.813241004943848, "learning_rate": 5.648262678545512e-07, "loss": 3.937, "step": 1128900 }, { "epoch": 6.807382930741308, "grad_norm": 5.244570255279541, "learning_rate": 5.630633937532714e-07, "loss": 4.5116, "step": 1128950 }, { "epoch": 6.807684422522642, "grad_norm": 6.057560920715332, "learning_rate": 5.61303269825547e-07, "loss": 4.3234, "step": 1129000 }, { "epoch": 6.807985914303976, "grad_norm": 3.3824894428253174, "learning_rate": 5.595458961037691e-07, "loss": 3.8112, "step": 1129050 }, { "epoch": 6.80828740608531, "grad_norm": 5.661864757537842, "learning_rate": 5.577912726202949e-07, "loss": 4.0585, "step": 1129100 }, { "epoch": 6.8085888978666445, "grad_norm": 6.579214572906494, "learning_rate": 5.560393994074153e-07, "loss": 3.6551, "step": 1129150 }, { "epoch": 6.808890389647978, "grad_norm": 9.181286811828613, "learning_rate": 5.542902764973378e-07, "loss": 3.9297, "step": 1129200 }, { "epoch": 6.809191881429312, "grad_norm": 4.391714096069336, "learning_rate": 5.525439039223034e-07, "loss": 4.2753, "step": 1129250 }, { "epoch": 6.809493373210646, "grad_norm": 6.166991233825684, "learning_rate": 5.508002817144197e-07, "loss": 4.4112, "step": 1129300 }, { "epoch": 6.80979486499198, "grad_norm": 6.5628790855407715, "learning_rate": 5.49059409905761e-07, "loss": 4.1896, "step": 1129350 }, { "epoch": 6.810096356773315, "grad_norm": 1.5376349687576294, "learning_rate": 5.473212885284018e-07, "loss": 4.3368, "step": 1129400 }, { "epoch": 6.810397848554649, "grad_norm": 5.452514171600342, "learning_rate": 5.455859176143162e-07, "loss": 4.3728, "step": 1129450 }, { "epoch": 6.810699340335982, "grad_norm": 6.668664455413818, "learning_rate": 5.43853297195429e-07, "loss": 4.2486, "step": 1129500 }, { "epoch": 6.811000832117316, "grad_norm": 4.636953830718994, "learning_rate": 5.421234273036479e-07, "loss": 4.4896, "step": 1129550 }, { "epoch": 6.8113023238986505, "grad_norm": 4.375360488891602, "learning_rate": 5.403963079707807e-07, "loss": 4.129, "step": 1129600 }, { "epoch": 6.811603815679985, "grad_norm": 5.671936511993408, "learning_rate": 5.386719392286521e-07, "loss": 3.9815, "step": 1129650 }, { "epoch": 6.811905307461319, "grad_norm": 5.037105560302734, "learning_rate": 5.369503211089532e-07, "loss": 3.777, "step": 1129700 }, { "epoch": 6.812206799242652, "grad_norm": 5.341862201690674, "learning_rate": 5.352314536434088e-07, "loss": 4.1067, "step": 1129750 }, { "epoch": 6.8125082910239865, "grad_norm": 5.113439083099365, "learning_rate": 5.335153368636102e-07, "loss": 4.4236, "step": 1129800 }, { "epoch": 6.812809782805321, "grad_norm": 4.433314800262451, "learning_rate": 5.318019708011656e-07, "loss": 4.5911, "step": 1129850 }, { "epoch": 6.813111274586655, "grad_norm": 5.498994827270508, "learning_rate": 5.300913554876163e-07, "loss": 4.2806, "step": 1129900 }, { "epoch": 6.813412766367989, "grad_norm": 6.010280609130859, "learning_rate": 5.283834909544205e-07, "loss": 4.2821, "step": 1129950 }, { "epoch": 6.813714258149323, "grad_norm": 7.830631732940674, "learning_rate": 5.266783772330196e-07, "loss": 3.9716, "step": 1130000 }, { "epoch": 6.814015749930657, "grad_norm": 7.234920978546143, "learning_rate": 5.249760143547887e-07, "loss": 4.1532, "step": 1130050 }, { "epoch": 6.814317241711991, "grad_norm": 7.4386491775512695, "learning_rate": 5.23276402351086e-07, "loss": 3.9965, "step": 1130100 }, { "epoch": 6.814618733493325, "grad_norm": 6.3711256980896, "learning_rate": 5.215795412531364e-07, "loss": 4.1001, "step": 1130150 }, { "epoch": 6.814920225274659, "grad_norm": 1.4948779344558716, "learning_rate": 5.198854310922152e-07, "loss": 3.8228, "step": 1130200 }, { "epoch": 6.815221717055993, "grad_norm": 6.785820960998535, "learning_rate": 5.181940718994803e-07, "loss": 4.585, "step": 1130250 }, { "epoch": 6.815523208837327, "grad_norm": 4.510908603668213, "learning_rate": 5.165054637060573e-07, "loss": 3.9921, "step": 1130300 }, { "epoch": 6.815824700618661, "grad_norm": 6.336457252502441, "learning_rate": 5.14819606543021e-07, "loss": 4.2116, "step": 1130350 }, { "epoch": 6.816126192399995, "grad_norm": 8.573690414428711, "learning_rate": 5.131365004414134e-07, "loss": 4.4027, "step": 1130400 }, { "epoch": 6.816427684181329, "grad_norm": 1.4060322046279907, "learning_rate": 5.114561454322097e-07, "loss": 3.9752, "step": 1130450 }, { "epoch": 6.816729175962664, "grad_norm": 4.619846343994141, "learning_rate": 5.097785415463018e-07, "loss": 3.9373, "step": 1130500 }, { "epoch": 6.817030667743998, "grad_norm": 4.016543865203857, "learning_rate": 5.081036888145983e-07, "loss": 4.3573, "step": 1130550 }, { "epoch": 6.817332159525331, "grad_norm": 7.320102691650391, "learning_rate": 5.064315872679081e-07, "loss": 4.4811, "step": 1130600 }, { "epoch": 6.817633651306665, "grad_norm": 6.119865894317627, "learning_rate": 5.047622369370063e-07, "loss": 4.3182, "step": 1130650 }, { "epoch": 6.8179351430879995, "grad_norm": 6.401533603668213, "learning_rate": 5.030956378526185e-07, "loss": 4.2525, "step": 1130700 }, { "epoch": 6.818236634869334, "grad_norm": 3.279844045639038, "learning_rate": 5.014317900454035e-07, "loss": 4.5906, "step": 1130750 }, { "epoch": 6.818538126650667, "grad_norm": 7.151904582977295, "learning_rate": 4.997706935459866e-07, "loss": 4.0995, "step": 1130800 }, { "epoch": 6.818839618432001, "grad_norm": 5.519954204559326, "learning_rate": 4.981123483849603e-07, "loss": 4.2719, "step": 1130850 }, { "epoch": 6.8191411102133355, "grad_norm": 7.553266525268555, "learning_rate": 4.964567545928e-07, "loss": 4.4001, "step": 1130900 }, { "epoch": 6.81944260199467, "grad_norm": 6.148341178894043, "learning_rate": 4.948039122000147e-07, "loss": 4.2085, "step": 1130950 }, { "epoch": 6.819744093776004, "grad_norm": 6.569821357727051, "learning_rate": 4.931538212370135e-07, "loss": 4.4827, "step": 1131000 }, { "epoch": 6.819744093776004, "eval_loss": 4.783308982849121, "eval_runtime": 39.0772, "eval_samples_per_second": 13.102, "eval_steps_per_second": 6.551, "eval_tts_loss": 8.164994957883748, "step": 1131000 }, { "epoch": 6.820045585557338, "grad_norm": 4.738739490509033, "learning_rate": 4.915064817341552e-07, "loss": 4.0627, "step": 1131050 }, { "epoch": 6.820347077338671, "grad_norm": 3.9707021713256836, "learning_rate": 4.898618937217492e-07, "loss": 3.9486, "step": 1131100 }, { "epoch": 6.820648569120006, "grad_norm": 4.396550178527832, "learning_rate": 4.882200572300876e-07, "loss": 4.1899, "step": 1131150 }, { "epoch": 6.82095006090134, "grad_norm": 4.554459095001221, "learning_rate": 4.865809722893632e-07, "loss": 4.0185, "step": 1131200 }, { "epoch": 6.821251552682674, "grad_norm": 5.673370838165283, "learning_rate": 4.849446389297518e-07, "loss": 4.2618, "step": 1131250 }, { "epoch": 6.821553044464008, "grad_norm": 2.3464529514312744, "learning_rate": 4.833110571813625e-07, "loss": 4.1699, "step": 1131300 }, { "epoch": 6.8218545362453415, "grad_norm": 6.153191089630127, "learning_rate": 4.816802270742548e-07, "loss": 4.6318, "step": 1131350 }, { "epoch": 6.822156028026676, "grad_norm": 4.393435001373291, "learning_rate": 4.800521486384545e-07, "loss": 4.3093, "step": 1131400 }, { "epoch": 6.82245751980801, "grad_norm": 3.8651270866394043, "learning_rate": 4.784268219039211e-07, "loss": 4.0672, "step": 1131450 }, { "epoch": 6.822759011589344, "grad_norm": 5.018906593322754, "learning_rate": 4.76804246900564e-07, "loss": 4.4096, "step": 1131500 }, { "epoch": 6.823060503370678, "grad_norm": 2.129363775253296, "learning_rate": 4.751844236582425e-07, "loss": 4.1118, "step": 1131550 }, { "epoch": 6.823361995152013, "grad_norm": 5.707961559295654, "learning_rate": 4.7356735220676625e-07, "loss": 4.3162, "step": 1131600 }, { "epoch": 6.823663486933346, "grad_norm": 5.985139846801758, "learning_rate": 4.7195303257589466e-07, "loss": 4.1086, "step": 1131650 }, { "epoch": 6.82396497871468, "grad_norm": 4.7092413902282715, "learning_rate": 4.703414647953374e-07, "loss": 4.6556, "step": 1131700 }, { "epoch": 6.824266470496014, "grad_norm": 3.907391309738159, "learning_rate": 4.687326488947707e-07, "loss": 4.2171, "step": 1131750 }, { "epoch": 6.8245679622773485, "grad_norm": 5.516447067260742, "learning_rate": 4.6712658490377084e-07, "loss": 3.7906, "step": 1131800 }, { "epoch": 6.824869454058683, "grad_norm": 3.5064051151275635, "learning_rate": 4.6552327285191425e-07, "loss": 4.3976, "step": 1131850 }, { "epoch": 6.825170945840016, "grad_norm": 7.587329864501953, "learning_rate": 4.6392271276871065e-07, "loss": 3.9705, "step": 1131900 }, { "epoch": 6.82547243762135, "grad_norm": 8.526970863342285, "learning_rate": 4.6232490468360305e-07, "loss": 4.425, "step": 1131950 }, { "epoch": 6.8257739294026845, "grad_norm": 10.809925079345703, "learning_rate": 4.60729848626018e-07, "loss": 4.17, "step": 1132000 }, { "epoch": 6.826075421184019, "grad_norm": 4.415956020355225, "learning_rate": 4.5913754462528206e-07, "loss": 3.9444, "step": 1132050 }, { "epoch": 6.826376912965353, "grad_norm": 5.871251106262207, "learning_rate": 4.575479927107217e-07, "loss": 3.9364, "step": 1132100 }, { "epoch": 6.826678404746687, "grad_norm": 5.136732578277588, "learning_rate": 4.559611929115636e-07, "loss": 4.0927, "step": 1132150 }, { "epoch": 6.82697989652802, "grad_norm": 3.7436470985412598, "learning_rate": 4.5437714525705104e-07, "loss": 4.0454, "step": 1132200 }, { "epoch": 6.827281388309355, "grad_norm": 5.8193817138671875, "learning_rate": 4.527958497763107e-07, "loss": 3.8901, "step": 1132250 }, { "epoch": 6.827582880090689, "grad_norm": 6.153421401977539, "learning_rate": 4.512173064984359e-07, "loss": 4.5556, "step": 1132300 }, { "epoch": 6.827884371872023, "grad_norm": 5.0792951583862305, "learning_rate": 4.4964151545250346e-07, "loss": 4.1727, "step": 1132350 }, { "epoch": 6.828185863653356, "grad_norm": 4.888514518737793, "learning_rate": 4.480684766674736e-07, "loss": 4.4975, "step": 1132400 }, { "epoch": 6.8284873554346905, "grad_norm": 4.429961681365967, "learning_rate": 4.464981901723563e-07, "loss": 4.0917, "step": 1132450 }, { "epoch": 6.828788847216025, "grad_norm": 4.076488018035889, "learning_rate": 4.4493065599599533e-07, "loss": 4.2715, "step": 1132500 }, { "epoch": 6.829090338997359, "grad_norm": 4.917304515838623, "learning_rate": 4.433658741672508e-07, "loss": 4.1319, "step": 1132550 }, { "epoch": 6.829391830778693, "grad_norm": 4.9406938552856445, "learning_rate": 4.4180384471494967e-07, "loss": 3.9367, "step": 1132600 }, { "epoch": 6.829693322560027, "grad_norm": 2.2072436809539795, "learning_rate": 4.4024456766780234e-07, "loss": 4.1474, "step": 1132650 }, { "epoch": 6.829994814341361, "grad_norm": 9.926545143127441, "learning_rate": 4.3868804305453586e-07, "loss": 4.5585, "step": 1132700 }, { "epoch": 6.830296306122695, "grad_norm": 2.047659397125244, "learning_rate": 4.371342709037606e-07, "loss": 4.3861, "step": 1132750 }, { "epoch": 6.830597797904029, "grad_norm": 5.997836589813232, "learning_rate": 4.3558325124410375e-07, "loss": 4.5505, "step": 1132800 }, { "epoch": 6.830899289685363, "grad_norm": 4.634766578674316, "learning_rate": 4.340349841040924e-07, "loss": 3.8359, "step": 1132850 }, { "epoch": 6.8312007814666975, "grad_norm": 5.470135688781738, "learning_rate": 4.3248946951223715e-07, "loss": 4.3822, "step": 1132900 }, { "epoch": 6.831502273248031, "grad_norm": 5.774299144744873, "learning_rate": 4.309467074969486e-07, "loss": 4.2048, "step": 1132950 }, { "epoch": 6.831803765029365, "grad_norm": 3.9429547786712646, "learning_rate": 4.29406698086654e-07, "loss": 4.0604, "step": 1133000 }, { "epoch": 6.832105256810699, "grad_norm": 4.087082386016846, "learning_rate": 4.278694413096806e-07, "loss": 4.2184, "step": 1133050 }, { "epoch": 6.8324067485920335, "grad_norm": 6.352145671844482, "learning_rate": 4.263349371943059e-07, "loss": 4.6433, "step": 1133100 }, { "epoch": 6.832708240373368, "grad_norm": 2.9729115962982178, "learning_rate": 4.2480318576879056e-07, "loss": 4.4264, "step": 1133150 }, { "epoch": 6.833009732154702, "grad_norm": 9.262404441833496, "learning_rate": 4.232741870613288e-07, "loss": 4.1283, "step": 1133200 }, { "epoch": 6.833311223936035, "grad_norm": 5.530897617340088, "learning_rate": 4.217479411000313e-07, "loss": 3.6242, "step": 1133250 }, { "epoch": 6.833612715717369, "grad_norm": 5.59500789642334, "learning_rate": 4.202244479130257e-07, "loss": 4.4288, "step": 1133300 }, { "epoch": 6.833914207498704, "grad_norm": 6.927201271057129, "learning_rate": 4.1870370752830616e-07, "loss": 4.3615, "step": 1133350 }, { "epoch": 6.834215699280038, "grad_norm": 3.468550443649292, "learning_rate": 4.1718571997388373e-07, "loss": 4.2569, "step": 1133400 }, { "epoch": 6.834517191061372, "grad_norm": 4.348085403442383, "learning_rate": 4.1567048527770264e-07, "loss": 4.03, "step": 1133450 }, { "epoch": 6.834818682842705, "grad_norm": 6.852555274963379, "learning_rate": 4.1415800346762397e-07, "loss": 4.5149, "step": 1133500 }, { "epoch": 6.8351201746240395, "grad_norm": 8.146931648254395, "learning_rate": 4.126482745715087e-07, "loss": 4.137, "step": 1133550 }, { "epoch": 6.835421666405374, "grad_norm": 6.985273361206055, "learning_rate": 4.111412986171181e-07, "loss": 4.1892, "step": 1133600 }, { "epoch": 6.835723158186708, "grad_norm": 5.075892448425293, "learning_rate": 4.0963707563221315e-07, "loss": 4.3922, "step": 1133650 }, { "epoch": 6.836024649968042, "grad_norm": 5.270501613616943, "learning_rate": 4.081356056444551e-07, "loss": 3.8716, "step": 1133700 }, { "epoch": 6.836326141749376, "grad_norm": 7.211418151855469, "learning_rate": 4.066368886814719e-07, "loss": 4.1144, "step": 1133750 }, { "epoch": 6.83662763353071, "grad_norm": 4.6214776039123535, "learning_rate": 4.051409247708581e-07, "loss": 4.0938, "step": 1133800 }, { "epoch": 6.836929125312044, "grad_norm": 6.56776762008667, "learning_rate": 4.036477139401584e-07, "loss": 4.3992, "step": 1133850 }, { "epoch": 6.837230617093378, "grad_norm": 5.822906494140625, "learning_rate": 4.021572562168174e-07, "loss": 4.0379, "step": 1133900 }, { "epoch": 6.837532108874712, "grad_norm": 4.34934663772583, "learning_rate": 4.006695516282965e-07, "loss": 4.2721, "step": 1133950 }, { "epoch": 6.837833600656046, "grad_norm": 6.037632942199707, "learning_rate": 3.991846002019572e-07, "loss": 4.496, "step": 1134000 }, { "epoch": 6.837833600656046, "eval_loss": 4.783073902130127, "eval_runtime": 39.1927, "eval_samples_per_second": 13.064, "eval_steps_per_second": 6.532, "eval_tts_loss": 8.16860453082359, "step": 1134000 }, { "epoch": 6.83813509243738, "grad_norm": 6.56540584564209, "learning_rate": 3.977024019651276e-07, "loss": 4.1547, "step": 1134050 }, { "epoch": 6.838436584218714, "grad_norm": 4.9516448974609375, "learning_rate": 3.9622295694510254e-07, "loss": 4.0426, "step": 1134100 }, { "epoch": 6.838738076000048, "grad_norm": 6.154425621032715, "learning_rate": 3.9474626516909357e-07, "loss": 4.3442, "step": 1134150 }, { "epoch": 6.8390395677813824, "grad_norm": 2.789173126220703, "learning_rate": 3.9327232666427897e-07, "loss": 3.9762, "step": 1134200 }, { "epoch": 6.839341059562717, "grad_norm": 6.2992072105407715, "learning_rate": 3.9180114145778707e-07, "loss": 4.2707, "step": 1134250 }, { "epoch": 6.839642551344051, "grad_norm": 4.555030345916748, "learning_rate": 3.9033270957669614e-07, "loss": 4.5268, "step": 1134300 }, { "epoch": 6.839944043125384, "grad_norm": 4.837050914764404, "learning_rate": 3.8886703104803465e-07, "loss": 4.409, "step": 1134350 }, { "epoch": 6.840245534906718, "grad_norm": 6.730375289916992, "learning_rate": 3.8740410589876427e-07, "loss": 3.9724, "step": 1134400 }, { "epoch": 6.840547026688053, "grad_norm": 4.875951766967773, "learning_rate": 3.8594393415579684e-07, "loss": 3.7789, "step": 1134450 }, { "epoch": 6.840848518469387, "grad_norm": 3.5384106636047363, "learning_rate": 3.844865158460442e-07, "loss": 4.1071, "step": 1134500 }, { "epoch": 6.84115001025072, "grad_norm": 6.776313781738281, "learning_rate": 3.8303185099628486e-07, "loss": 4.4177, "step": 1134550 }, { "epoch": 6.841451502032054, "grad_norm": 6.204708099365234, "learning_rate": 3.8157993963331414e-07, "loss": 4.2, "step": 1134600 }, { "epoch": 6.8417529938133885, "grad_norm": 3.89072585105896, "learning_rate": 3.8013078178386057e-07, "loss": 3.9545, "step": 1134650 }, { "epoch": 6.842054485594723, "grad_norm": 4.633174419403076, "learning_rate": 3.786843774745696e-07, "loss": 4.2551, "step": 1134700 }, { "epoch": 6.842355977376057, "grad_norm": 5.769856929779053, "learning_rate": 3.772407267320698e-07, "loss": 4.2624, "step": 1134750 }, { "epoch": 6.842657469157391, "grad_norm": 3.9571945667266846, "learning_rate": 3.7579982958294007e-07, "loss": 4.0357, "step": 1134800 }, { "epoch": 6.8429589609387245, "grad_norm": 5.722338676452637, "learning_rate": 3.743616860536758e-07, "loss": 3.8945, "step": 1134850 }, { "epoch": 6.843260452720059, "grad_norm": 3.7767889499664307, "learning_rate": 3.7292629617077243e-07, "loss": 4.106, "step": 1134900 }, { "epoch": 6.843561944501393, "grad_norm": 3.9923439025878906, "learning_rate": 3.714936599606255e-07, "loss": 4.0543, "step": 1134950 }, { "epoch": 6.843863436282727, "grad_norm": 4.386150360107422, "learning_rate": 3.7006377744959737e-07, "loss": 4.373, "step": 1135000 }, { "epoch": 6.844164928064061, "grad_norm": 3.9716358184814453, "learning_rate": 3.686366486640335e-07, "loss": 4.0282, "step": 1135050 }, { "epoch": 6.844466419845395, "grad_norm": 6.561348915100098, "learning_rate": 3.6721227363016303e-07, "loss": 4.0225, "step": 1135100 }, { "epoch": 6.844767911626729, "grad_norm": 6.454498767852783, "learning_rate": 3.657906523742149e-07, "loss": 4.2346, "step": 1135150 }, { "epoch": 6.845069403408063, "grad_norm": 5.332611083984375, "learning_rate": 3.6437178492235153e-07, "loss": 3.7719, "step": 1135200 }, { "epoch": 6.845370895189397, "grad_norm": 4.894257068634033, "learning_rate": 3.629556713006687e-07, "loss": 4.2284, "step": 1135250 }, { "epoch": 6.845672386970731, "grad_norm": 6.591974258422852, "learning_rate": 3.6154231153526226e-07, "loss": 4.2538, "step": 1135300 }, { "epoch": 6.845973878752066, "grad_norm": 6.608980655670166, "learning_rate": 3.6013170565211137e-07, "loss": 4.1733, "step": 1135350 }, { "epoch": 6.846275370533399, "grad_norm": 3.9639945030212402, "learning_rate": 3.5872385367719524e-07, "loss": 4.4549, "step": 1135400 }, { "epoch": 6.846576862314733, "grad_norm": 2.952678918838501, "learning_rate": 3.573187556364265e-07, "loss": 3.6538, "step": 1135450 }, { "epoch": 6.846878354096067, "grad_norm": 2.651174306869507, "learning_rate": 3.5591641155563453e-07, "loss": 4.0941, "step": 1135500 }, { "epoch": 6.847179845877402, "grad_norm": 6.066713809967041, "learning_rate": 3.545168214606653e-07, "loss": 3.8799, "step": 1135550 }, { "epoch": 6.847481337658736, "grad_norm": 4.080566883087158, "learning_rate": 3.5311998537724817e-07, "loss": 4.4643, "step": 1135600 }, { "epoch": 6.847782829440069, "grad_norm": 6.6817731857299805, "learning_rate": 3.517259033311126e-07, "loss": 4.2761, "step": 1135650 }, { "epoch": 6.848084321221403, "grad_norm": 4.6522674560546875, "learning_rate": 3.503345753478881e-07, "loss": 3.8519, "step": 1135700 }, { "epoch": 6.8483858130027375, "grad_norm": 3.9563896656036377, "learning_rate": 3.489460014532042e-07, "loss": 4.1943, "step": 1135750 }, { "epoch": 6.848687304784072, "grad_norm": 7.361268520355225, "learning_rate": 3.4756018167259037e-07, "loss": 4.2928, "step": 1135800 }, { "epoch": 6.848988796565406, "grad_norm": 4.499571323394775, "learning_rate": 3.461771160315763e-07, "loss": 3.9931, "step": 1135850 }, { "epoch": 6.84929028834674, "grad_norm": 4.3933210372924805, "learning_rate": 3.447968045556082e-07, "loss": 3.8029, "step": 1135900 }, { "epoch": 6.8495917801280735, "grad_norm": 4.834592819213867, "learning_rate": 3.434192472700825e-07, "loss": 4.1219, "step": 1135950 }, { "epoch": 6.849893271909408, "grad_norm": 6.101498603820801, "learning_rate": 3.420444442003456e-07, "loss": 3.8225, "step": 1136000 }, { "epoch": 6.850194763690742, "grad_norm": 4.384661674499512, "learning_rate": 3.406723953717272e-07, "loss": 4.1684, "step": 1136050 }, { "epoch": 6.850496255472076, "grad_norm": 7.198178291320801, "learning_rate": 3.3930310080944046e-07, "loss": 4.0416, "step": 1136100 }, { "epoch": 6.850797747253409, "grad_norm": 6.258450508117676, "learning_rate": 3.379365605387152e-07, "loss": 4.2858, "step": 1136150 }, { "epoch": 6.851099239034744, "grad_norm": 7.395784378051758, "learning_rate": 3.3657277458468133e-07, "loss": 4.4068, "step": 1136200 }, { "epoch": 6.851400730816078, "grad_norm": 5.87676477432251, "learning_rate": 3.3521174297243545e-07, "loss": 3.7755, "step": 1136250 }, { "epoch": 6.851702222597412, "grad_norm": 5.793233394622803, "learning_rate": 3.3385346572704084e-07, "loss": 4.1295, "step": 1136300 }, { "epoch": 6.852003714378746, "grad_norm": 5.780034065246582, "learning_rate": 3.324979428734942e-07, "loss": 4.161, "step": 1136350 }, { "epoch": 6.85230520616008, "grad_norm": 4.376260757446289, "learning_rate": 3.311451744367255e-07, "loss": 4.0095, "step": 1136400 }, { "epoch": 6.852606697941414, "grad_norm": 4.439761638641357, "learning_rate": 3.2979516044166486e-07, "loss": 4.6399, "step": 1136450 }, { "epoch": 6.852908189722748, "grad_norm": 4.559563159942627, "learning_rate": 3.2844790091310915e-07, "loss": 3.9242, "step": 1136500 }, { "epoch": 6.853209681504082, "grad_norm": 5.95785665512085, "learning_rate": 3.271033958758884e-07, "loss": 3.8878, "step": 1136550 }, { "epoch": 6.853511173285416, "grad_norm": 4.4193010330200195, "learning_rate": 3.257616453547329e-07, "loss": 4.1895, "step": 1136600 }, { "epoch": 6.853812665066751, "grad_norm": 6.8796539306640625, "learning_rate": 3.2442264937432293e-07, "loss": 4.1159, "step": 1136650 }, { "epoch": 6.854114156848084, "grad_norm": 5.907749652862549, "learning_rate": 3.230864079593387e-07, "loss": 4.2187, "step": 1136700 }, { "epoch": 6.854415648629418, "grad_norm": 4.480772495269775, "learning_rate": 3.217529211343273e-07, "loss": 4.0603, "step": 1136750 }, { "epoch": 6.854717140410752, "grad_norm": 5.93387508392334, "learning_rate": 3.2042218892386895e-07, "loss": 4.3068, "step": 1136800 }, { "epoch": 6.8550186321920865, "grad_norm": 2.329770088195801, "learning_rate": 3.190942113524275e-07, "loss": 4.3059, "step": 1136850 }, { "epoch": 6.855320123973421, "grad_norm": 6.745229721069336, "learning_rate": 3.177689884444501e-07, "loss": 4.2559, "step": 1136900 }, { "epoch": 6.855621615754755, "grad_norm": 6.330860137939453, "learning_rate": 3.164465202243338e-07, "loss": 4.3572, "step": 1136950 }, { "epoch": 6.855923107536088, "grad_norm": 2.9852569103240967, "learning_rate": 3.1512680671640924e-07, "loss": 4.2488, "step": 1137000 }, { "epoch": 6.855923107536088, "eval_loss": 4.783166885375977, "eval_runtime": 39.0046, "eval_samples_per_second": 13.127, "eval_steps_per_second": 6.563, "eval_tts_loss": 8.169036834692072, "step": 1137000 }, { "epoch": 6.8562245993174225, "grad_norm": 2.8676555156707764, "learning_rate": 3.138098479449569e-07, "loss": 4.313, "step": 1137050 }, { "epoch": 6.856526091098757, "grad_norm": 4.588360786437988, "learning_rate": 3.124956439342241e-07, "loss": 4.0733, "step": 1137100 }, { "epoch": 6.856827582880091, "grad_norm": 2.8172035217285156, "learning_rate": 3.111841947083915e-07, "loss": 3.8503, "step": 1137150 }, { "epoch": 6.857129074661425, "grad_norm": 7.249810695648193, "learning_rate": 3.098755002915898e-07, "loss": 4.1455, "step": 1137200 }, { "epoch": 6.857430566442758, "grad_norm": 4.123479843139648, "learning_rate": 3.0856956070791637e-07, "loss": 4.0203, "step": 1137250 }, { "epoch": 6.857732058224093, "grad_norm": 4.521236896514893, "learning_rate": 3.07266375981402e-07, "loss": 4.3371, "step": 1137300 }, { "epoch": 6.858033550005427, "grad_norm": 4.0525898933410645, "learning_rate": 3.0596594613602755e-07, "loss": 4.1438, "step": 1137350 }, { "epoch": 6.858335041786761, "grad_norm": 7.717141628265381, "learning_rate": 3.046682711957238e-07, "loss": 3.9866, "step": 1137400 }, { "epoch": 6.858636533568095, "grad_norm": 4.760512828826904, "learning_rate": 3.033733511843717e-07, "loss": 4.4686, "step": 1137450 }, { "epoch": 6.858938025349429, "grad_norm": 2.527880907058716, "learning_rate": 3.020811861258021e-07, "loss": 4.1669, "step": 1137500 }, { "epoch": 6.859239517130763, "grad_norm": 5.924065589904785, "learning_rate": 3.0079177604381275e-07, "loss": 4.5215, "step": 1137550 }, { "epoch": 6.859541008912097, "grad_norm": 4.868409633636475, "learning_rate": 2.995051209621013e-07, "loss": 4.1251, "step": 1137600 }, { "epoch": 6.859842500693431, "grad_norm": 0.7099277973175049, "learning_rate": 2.9822122090438217e-07, "loss": 4.2578, "step": 1137650 }, { "epoch": 6.860143992474765, "grad_norm": 2.0302090644836426, "learning_rate": 2.969400758942364e-07, "loss": 4.1412, "step": 1137700 }, { "epoch": 6.860445484256099, "grad_norm": 6.856870651245117, "learning_rate": 2.9566168595529515e-07, "loss": 4.3389, "step": 1137750 }, { "epoch": 6.860746976037433, "grad_norm": 2.78745436668396, "learning_rate": 2.943860511110563e-07, "loss": 4.0256, "step": 1137800 }, { "epoch": 6.861048467818767, "grad_norm": 1.6518502235412598, "learning_rate": 2.93113171385001e-07, "loss": 4.3064, "step": 1137850 }, { "epoch": 6.861349959600101, "grad_norm": 2.167187452316284, "learning_rate": 2.9184304680054395e-07, "loss": 4.0254, "step": 1137900 }, { "epoch": 6.8616514513814355, "grad_norm": 5.60815954208374, "learning_rate": 2.90575677381083e-07, "loss": 4.3076, "step": 1137950 }, { "epoch": 6.86195294316277, "grad_norm": 5.05580472946167, "learning_rate": 2.8931106314991623e-07, "loss": 3.7053, "step": 1138000 }, { "epoch": 6.862254434944104, "grad_norm": 4.305138111114502, "learning_rate": 2.880492041303417e-07, "loss": 4.2091, "step": 1138050 }, { "epoch": 6.862555926725437, "grad_norm": 5.158095836639404, "learning_rate": 2.867901003455741e-07, "loss": 4.5562, "step": 1138100 }, { "epoch": 6.8628574185067714, "grad_norm": 4.21353006362915, "learning_rate": 2.8553375181877835e-07, "loss": 4.1408, "step": 1138150 }, { "epoch": 6.863158910288106, "grad_norm": 1.9722248315811157, "learning_rate": 2.842801585730692e-07, "loss": 3.9915, "step": 1138200 }, { "epoch": 6.86346040206944, "grad_norm": 4.311611175537109, "learning_rate": 2.8302932063154484e-07, "loss": 4.2815, "step": 1138250 }, { "epoch": 6.863761893850773, "grad_norm": 4.360747337341309, "learning_rate": 2.817812380172035e-07, "loss": 4.1713, "step": 1138300 }, { "epoch": 6.864063385632107, "grad_norm": 4.2335638999938965, "learning_rate": 2.805359107530103e-07, "loss": 4.2778, "step": 1138350 }, { "epoch": 6.864364877413442, "grad_norm": 5.020861625671387, "learning_rate": 2.7929333886189673e-07, "loss": 4.4187, "step": 1138400 }, { "epoch": 6.864666369194776, "grad_norm": 5.44598913192749, "learning_rate": 2.780535223667446e-07, "loss": 4.0386, "step": 1138450 }, { "epoch": 6.86496786097611, "grad_norm": 6.076391696929932, "learning_rate": 2.768164612903356e-07, "loss": 3.797, "step": 1138500 }, { "epoch": 6.865269352757444, "grad_norm": 5.380213260650635, "learning_rate": 2.755821556554516e-07, "loss": 4.2539, "step": 1138550 }, { "epoch": 6.8655708445387775, "grad_norm": 7.646152019500732, "learning_rate": 2.743506054848077e-07, "loss": 4.0025, "step": 1138600 }, { "epoch": 6.865872336320112, "grad_norm": 4.373816967010498, "learning_rate": 2.731218108010691e-07, "loss": 4.421, "step": 1138650 }, { "epoch": 6.866173828101446, "grad_norm": 5.996427059173584, "learning_rate": 2.718957716268677e-07, "loss": 4.0354, "step": 1138700 }, { "epoch": 6.86647531988278, "grad_norm": 6.562788963317871, "learning_rate": 2.706724879847355e-07, "loss": 4.1538, "step": 1138750 }, { "epoch": 6.866776811664114, "grad_norm": 9.438004493713379, "learning_rate": 2.694519598972211e-07, "loss": 4.156, "step": 1138800 }, { "epoch": 6.867078303445448, "grad_norm": 4.101658821105957, "learning_rate": 2.6823418738673995e-07, "loss": 3.793, "step": 1138850 }, { "epoch": 6.867379795226782, "grad_norm": 4.84734582901001, "learning_rate": 2.6701917047574074e-07, "loss": 4.1968, "step": 1138900 }, { "epoch": 6.867681287008116, "grad_norm": 3.8534436225891113, "learning_rate": 2.6580690918658884e-07, "loss": 4.3184, "step": 1138950 }, { "epoch": 6.86798277878945, "grad_norm": 5.5842108726501465, "learning_rate": 2.645974035415499e-07, "loss": 4.0211, "step": 1139000 }, { "epoch": 6.8682842705707845, "grad_norm": 4.578919410705566, "learning_rate": 2.6339065356293933e-07, "loss": 4.4101, "step": 1139050 }, { "epoch": 6.868585762352119, "grad_norm": 7.092202663421631, "learning_rate": 2.621866592729227e-07, "loss": 4.5453, "step": 1139100 }, { "epoch": 6.868887254133452, "grad_norm": 5.6471266746521, "learning_rate": 2.609854206936657e-07, "loss": 3.9483, "step": 1139150 }, { "epoch": 6.869188745914786, "grad_norm": 4.61777925491333, "learning_rate": 2.5978693784730054e-07, "loss": 4.0676, "step": 1139200 }, { "epoch": 6.86949023769612, "grad_norm": 4.2922868728637695, "learning_rate": 2.5859121075585963e-07, "loss": 4.251, "step": 1139250 }, { "epoch": 6.869791729477455, "grad_norm": 5.333596706390381, "learning_rate": 2.573982394413587e-07, "loss": 4.0128, "step": 1139300 }, { "epoch": 6.870093221258789, "grad_norm": 6.22287654876709, "learning_rate": 2.562080239257469e-07, "loss": 4.5482, "step": 1139350 }, { "epoch": 6.870394713040122, "grad_norm": 4.6989922523498535, "learning_rate": 2.550205642309233e-07, "loss": 4.2759, "step": 1139400 }, { "epoch": 6.870696204821456, "grad_norm": 4.681340217590332, "learning_rate": 2.538358603787538e-07, "loss": 4.6724, "step": 1139450 }, { "epoch": 6.870997696602791, "grad_norm": 2.9912967681884766, "learning_rate": 2.526539123910376e-07, "loss": 4.0574, "step": 1139500 }, { "epoch": 6.871299188384125, "grad_norm": 4.4907612800598145, "learning_rate": 2.51474720289524e-07, "loss": 4.1449, "step": 1139550 }, { "epoch": 6.871600680165459, "grad_norm": 6.294870853424072, "learning_rate": 2.502982840959122e-07, "loss": 4.3179, "step": 1139600 }, { "epoch": 6.871902171946793, "grad_norm": 4.4415202140808105, "learning_rate": 2.491246038318684e-07, "loss": 4.2833, "step": 1139650 }, { "epoch": 6.8722036637281265, "grad_norm": 5.186416149139404, "learning_rate": 2.479536795189585e-07, "loss": 4.2873, "step": 1139700 }, { "epoch": 6.872505155509461, "grad_norm": 4.427886962890625, "learning_rate": 2.467855111787653e-07, "loss": 4.3628, "step": 1139750 }, { "epoch": 6.872806647290795, "grad_norm": 4.418931007385254, "learning_rate": 2.456200988327883e-07, "loss": 4.0119, "step": 1139800 }, { "epoch": 6.873108139072129, "grad_norm": 5.747575283050537, "learning_rate": 2.444574425024437e-07, "loss": 4.3891, "step": 1139850 }, { "epoch": 6.8734096308534625, "grad_norm": 4.859823226928711, "learning_rate": 2.4329754220914767e-07, "loss": 4.2327, "step": 1139900 }, { "epoch": 6.873711122634797, "grad_norm": 6.338550567626953, "learning_rate": 2.4214039797426644e-07, "loss": 3.7657, "step": 1139950 }, { "epoch": 6.874012614416131, "grad_norm": 6.318322658538818, "learning_rate": 2.409860098190497e-07, "loss": 4.3114, "step": 1140000 }, { "epoch": 6.874012614416131, "eval_loss": 4.782625198364258, "eval_runtime": 39.1431, "eval_samples_per_second": 13.08, "eval_steps_per_second": 6.54, "eval_tts_loss": 8.166445844949553, "step": 1140000 }, { "epoch": 6.874314106197465, "grad_norm": 3.4365663528442383, "learning_rate": 2.3983437776478043e-07, "loss": 4.3726, "step": 1140050 }, { "epoch": 6.874615597978799, "grad_norm": 7.687468528747559, "learning_rate": 2.3868550183264166e-07, "loss": 4.6563, "step": 1140100 }, { "epoch": 6.8749170897601335, "grad_norm": 2.5928406715393066, "learning_rate": 2.3753938204376654e-07, "loss": 4.1439, "step": 1140150 }, { "epoch": 6.875218581541467, "grad_norm": 4.6149516105651855, "learning_rate": 2.3639601841927147e-07, "loss": 4.1552, "step": 1140200 }, { "epoch": 6.875520073322801, "grad_norm": 5.485229969024658, "learning_rate": 2.3525541098017298e-07, "loss": 4.043, "step": 1140250 }, { "epoch": 6.875821565104135, "grad_norm": 4.5974249839782715, "learning_rate": 2.3411755974747094e-07, "loss": 4.0293, "step": 1140300 }, { "epoch": 6.876123056885469, "grad_norm": 2.5241804122924805, "learning_rate": 2.3298246474211524e-07, "loss": 3.9571, "step": 1140350 }, { "epoch": 6.876424548666804, "grad_norm": 5.309401988983154, "learning_rate": 2.318501259849892e-07, "loss": 4.5935, "step": 1140400 }, { "epoch": 6.876726040448137, "grad_norm": 2.442695140838623, "learning_rate": 2.3072054349692615e-07, "loss": 4.2021, "step": 1140450 }, { "epoch": 6.877027532229471, "grad_norm": 5.259521961212158, "learning_rate": 2.2959371729870947e-07, "loss": 4.017, "step": 1140500 }, { "epoch": 6.877329024010805, "grad_norm": 4.643349647521973, "learning_rate": 2.2846964741110585e-07, "loss": 4.1042, "step": 1140550 }, { "epoch": 6.87763051579214, "grad_norm": 3.181889772415161, "learning_rate": 2.2734833385476548e-07, "loss": 4.0796, "step": 1140600 }, { "epoch": 6.877932007573474, "grad_norm": 6.4362874031066895, "learning_rate": 2.2622977665035514e-07, "loss": 4.2598, "step": 1140650 }, { "epoch": 6.878233499354808, "grad_norm": 5.456473350524902, "learning_rate": 2.251139758184417e-07, "loss": 4.6183, "step": 1140700 }, { "epoch": 6.878534991136141, "grad_norm": 3.26669979095459, "learning_rate": 2.2400093137957543e-07, "loss": 4.2636, "step": 1140750 }, { "epoch": 6.8788364829174755, "grad_norm": 6.111630916595459, "learning_rate": 2.2289064335420658e-07, "loss": 4.0377, "step": 1140800 }, { "epoch": 6.87913797469881, "grad_norm": 5.718616008758545, "learning_rate": 2.2178311176281881e-07, "loss": 4.1885, "step": 1140850 }, { "epoch": 6.879439466480144, "grad_norm": 3.593876838684082, "learning_rate": 2.2067833662574587e-07, "loss": 4.0446, "step": 1140900 }, { "epoch": 6.879740958261478, "grad_norm": 4.96671724319458, "learning_rate": 2.1957631796335473e-07, "loss": 3.8479, "step": 1140950 }, { "epoch": 6.8800424500428115, "grad_norm": 4.701363563537598, "learning_rate": 2.1847705579591257e-07, "loss": 3.8835, "step": 1141000 }, { "epoch": 6.880343941824146, "grad_norm": 4.036470413208008, "learning_rate": 2.1738055014363653e-07, "loss": 4.0291, "step": 1141050 }, { "epoch": 6.88064543360548, "grad_norm": 3.951066732406616, "learning_rate": 2.1628680102674377e-07, "loss": 3.8129, "step": 1141100 }, { "epoch": 6.880946925386814, "grad_norm": 4.637575149536133, "learning_rate": 2.1519580846531825e-07, "loss": 4.0998, "step": 1141150 }, { "epoch": 6.881248417168148, "grad_norm": 6.694678783416748, "learning_rate": 2.141075724794772e-07, "loss": 4.0399, "step": 1141200 }, { "epoch": 6.8815499089494825, "grad_norm": 4.203901290893555, "learning_rate": 2.1302209308922124e-07, "loss": 3.8812, "step": 1141250 }, { "epoch": 6.881851400730816, "grad_norm": 2.489539384841919, "learning_rate": 2.1193937031453446e-07, "loss": 4.4677, "step": 1141300 }, { "epoch": 6.88215289251215, "grad_norm": 5.275522232055664, "learning_rate": 2.1085940417535086e-07, "loss": 4.0417, "step": 1141350 }, { "epoch": 6.882454384293484, "grad_norm": 6.488619804382324, "learning_rate": 2.0978219469152125e-07, "loss": 4.3007, "step": 1141400 }, { "epoch": 6.882755876074818, "grad_norm": 6.039937496185303, "learning_rate": 2.0870774188291306e-07, "loss": 3.8615, "step": 1141450 }, { "epoch": 6.883057367856152, "grad_norm": 4.458571434020996, "learning_rate": 2.0763604576926052e-07, "loss": 4.4242, "step": 1141500 }, { "epoch": 6.883358859637486, "grad_norm": 4.142831802368164, "learning_rate": 2.0656710637029783e-07, "loss": 4.0531, "step": 1141550 }, { "epoch": 6.88366035141882, "grad_norm": 10.859737396240234, "learning_rate": 2.0550092370570925e-07, "loss": 4.3842, "step": 1141600 }, { "epoch": 6.883961843200154, "grad_norm": 5.616481304168701, "learning_rate": 2.0443749779511242e-07, "loss": 4.4236, "step": 1141650 }, { "epoch": 6.884263334981489, "grad_norm": 4.628053665161133, "learning_rate": 2.0337682865805838e-07, "loss": 4.3075, "step": 1141700 }, { "epoch": 6.884564826762823, "grad_norm": 6.360891819000244, "learning_rate": 2.0231891631409815e-07, "loss": 4.2339, "step": 1141750 }, { "epoch": 6.884866318544157, "grad_norm": 5.582727909088135, "learning_rate": 2.012637607826828e-07, "loss": 3.9159, "step": 1141800 }, { "epoch": 6.88516781032549, "grad_norm": 4.269870281219482, "learning_rate": 2.0021136208323018e-07, "loss": 4.403, "step": 1141850 }, { "epoch": 6.8854693021068245, "grad_norm": 3.9292919635772705, "learning_rate": 1.991617202351081e-07, "loss": 3.9506, "step": 1141900 }, { "epoch": 6.885770793888159, "grad_norm": 4.7740159034729, "learning_rate": 1.981148352576345e-07, "loss": 3.9965, "step": 1141950 }, { "epoch": 6.886072285669493, "grad_norm": 6.378256320953369, "learning_rate": 1.9707070717009388e-07, "loss": 4.0542, "step": 1142000 }, { "epoch": 6.886373777450826, "grad_norm": 3.99985671043396, "learning_rate": 1.9602933599167093e-07, "loss": 4.2636, "step": 1142050 }, { "epoch": 6.8866752692321604, "grad_norm": 4.536530017852783, "learning_rate": 1.949907217415503e-07, "loss": 4.2942, "step": 1142100 }, { "epoch": 6.886976761013495, "grad_norm": 7.848781585693359, "learning_rate": 1.9395486443885e-07, "loss": 4.2637, "step": 1142150 }, { "epoch": 6.887278252794829, "grad_norm": 6.138776779174805, "learning_rate": 1.929217641026215e-07, "loss": 4.4222, "step": 1142200 }, { "epoch": 6.887579744576163, "grad_norm": 7.051109790802002, "learning_rate": 1.918914207518829e-07, "loss": 4.0615, "step": 1142250 }, { "epoch": 6.887881236357497, "grad_norm": 3.8798458576202393, "learning_rate": 1.9086383440558573e-07, "loss": 4.256, "step": 1142300 }, { "epoch": 6.888182728138831, "grad_norm": 6.65895938873291, "learning_rate": 1.898390050826648e-07, "loss": 4.5948, "step": 1142350 }, { "epoch": 6.888484219920165, "grad_norm": 5.952325820922852, "learning_rate": 1.8881693280195508e-07, "loss": 4.0799, "step": 1142400 }, { "epoch": 6.888785711701499, "grad_norm": 4.534603595733643, "learning_rate": 1.8779761758229151e-07, "loss": 4.2722, "step": 1142450 }, { "epoch": 6.889087203482833, "grad_norm": 6.281252384185791, "learning_rate": 1.8678105944240905e-07, "loss": 4.0873, "step": 1142500 }, { "epoch": 6.889388695264167, "grad_norm": 4.343129634857178, "learning_rate": 1.8576725840102614e-07, "loss": 4.2669, "step": 1142550 }, { "epoch": 6.889690187045501, "grad_norm": 3.3421247005462646, "learning_rate": 1.8475621447679445e-07, "loss": 4.0714, "step": 1142600 }, { "epoch": 6.889991678826835, "grad_norm": 4.906384468078613, "learning_rate": 1.8374792768833247e-07, "loss": 4.2065, "step": 1142650 }, { "epoch": 6.890293170608169, "grad_norm": 4.2934112548828125, "learning_rate": 1.82742398054192e-07, "loss": 4.1154, "step": 1142700 }, { "epoch": 6.890594662389503, "grad_norm": 4.398365497589111, "learning_rate": 1.8173962559287492e-07, "loss": 4.3868, "step": 1142750 }, { "epoch": 6.890896154170838, "grad_norm": 7.859868049621582, "learning_rate": 1.807396103228498e-07, "loss": 3.9209, "step": 1142800 }, { "epoch": 6.891197645952172, "grad_norm": 5.41025972366333, "learning_rate": 1.797423522625019e-07, "loss": 4.3719, "step": 1142850 }, { "epoch": 6.891499137733505, "grad_norm": 6.613307952880859, "learning_rate": 1.7874785143018323e-07, "loss": 4.2124, "step": 1142900 }, { "epoch": 6.891800629514839, "grad_norm": 5.504396915435791, "learning_rate": 1.7775610784421247e-07, "loss": 4.1837, "step": 1142950 }, { "epoch": 6.8921021212961735, "grad_norm": 6.774323463439941, "learning_rate": 1.7676712152282502e-07, "loss": 3.7741, "step": 1143000 }, { "epoch": 6.8921021212961735, "eval_loss": 4.782520294189453, "eval_runtime": 38.9689, "eval_samples_per_second": 13.139, "eval_steps_per_second": 6.569, "eval_tts_loss": 8.163861955105967, "step": 1143000 }, { "epoch": 6.892403613077508, "grad_norm": 3.9016048908233643, "learning_rate": 1.75780892484223e-07, "loss": 3.9164, "step": 1143050 }, { "epoch": 6.892705104858842, "grad_norm": 5.4465203285217285, "learning_rate": 1.747974207465752e-07, "loss": 4.4682, "step": 1143100 }, { "epoch": 6.893006596640175, "grad_norm": 4.382339954376221, "learning_rate": 1.7381670632796718e-07, "loss": 4.2594, "step": 1143150 }, { "epoch": 6.893308088421509, "grad_norm": 5.991855144500732, "learning_rate": 1.7283874924645114e-07, "loss": 4.11, "step": 1143200 }, { "epoch": 6.893609580202844, "grad_norm": 4.283123016357422, "learning_rate": 1.7186354952002935e-07, "loss": 4.2327, "step": 1143250 }, { "epoch": 6.893911071984178, "grad_norm": 8.192548751831055, "learning_rate": 1.7089110716663746e-07, "loss": 4.1753, "step": 1143300 }, { "epoch": 6.894212563765512, "grad_norm": 8.42394733428955, "learning_rate": 1.6992142220416117e-07, "loss": 4.2033, "step": 1143350 }, { "epoch": 6.894514055546846, "grad_norm": 3.923403024673462, "learning_rate": 1.689544946504695e-07, "loss": 4.5813, "step": 1143400 }, { "epoch": 6.89481554732818, "grad_norm": 4.385291576385498, "learning_rate": 1.6799032452336492e-07, "loss": 4.2458, "step": 1143450 }, { "epoch": 6.895117039109514, "grad_norm": 2.1789391040802, "learning_rate": 1.6702891184056654e-07, "loss": 4.3982, "step": 1143500 }, { "epoch": 6.895418530890848, "grad_norm": 4.579166412353516, "learning_rate": 1.660702566197769e-07, "loss": 4.429, "step": 1143550 }, { "epoch": 6.895720022672182, "grad_norm": 4.321342945098877, "learning_rate": 1.6511435887864854e-07, "loss": 4.04, "step": 1143600 }, { "epoch": 6.8960215144535155, "grad_norm": 5.34592866897583, "learning_rate": 1.6416121863475075e-07, "loss": 4.1777, "step": 1143650 }, { "epoch": 6.89632300623485, "grad_norm": 3.1557912826538086, "learning_rate": 1.6321083590565276e-07, "loss": 4.2473, "step": 1143700 }, { "epoch": 6.896624498016184, "grad_norm": 6.537299156188965, "learning_rate": 1.6226321070882398e-07, "loss": 4.548, "step": 1143750 }, { "epoch": 6.896925989797518, "grad_norm": 5.954992771148682, "learning_rate": 1.6131834306170044e-07, "loss": 3.9174, "step": 1143800 }, { "epoch": 6.897227481578852, "grad_norm": 5.661499977111816, "learning_rate": 1.603762329816849e-07, "loss": 4.0887, "step": 1143850 }, { "epoch": 6.8975289733601866, "grad_norm": 5.311155319213867, "learning_rate": 1.5943688048613012e-07, "loss": 4.3107, "step": 1143900 }, { "epoch": 6.89783046514152, "grad_norm": 5.687238693237305, "learning_rate": 1.58500285592289e-07, "loss": 4.4607, "step": 1143950 }, { "epoch": 6.898131956922854, "grad_norm": 6.391983985900879, "learning_rate": 1.5756644831741438e-07, "loss": 4.2035, "step": 1144000 }, { "epoch": 6.898433448704188, "grad_norm": 5.532625198364258, "learning_rate": 1.566353686787092e-07, "loss": 3.961, "step": 1144050 }, { "epoch": 6.8987349404855225, "grad_norm": 4.395569801330566, "learning_rate": 1.557070466932764e-07, "loss": 4.2592, "step": 1144100 }, { "epoch": 6.899036432266857, "grad_norm": 5.617376804351807, "learning_rate": 1.5478148237821897e-07, "loss": 4.2228, "step": 1144150 }, { "epoch": 6.89933792404819, "grad_norm": 3.9534189701080322, "learning_rate": 1.5385867575057333e-07, "loss": 3.8698, "step": 1144200 }, { "epoch": 6.899639415829524, "grad_norm": 4.969107151031494, "learning_rate": 1.5293862682730919e-07, "loss": 4.3151, "step": 1144250 }, { "epoch": 6.899940907610858, "grad_norm": 6.2606072425842285, "learning_rate": 1.52021335625363e-07, "loss": 4.0338, "step": 1144300 }, { "epoch": 6.900242399392193, "grad_norm": 6.5139994621276855, "learning_rate": 1.5110680216163795e-07, "loss": 4.4132, "step": 1144350 }, { "epoch": 6.900543891173527, "grad_norm": 4.711843967437744, "learning_rate": 1.5019502645293723e-07, "loss": 4.4892, "step": 1144400 }, { "epoch": 6.900845382954861, "grad_norm": 5.007483959197998, "learning_rate": 1.492860085160308e-07, "loss": 4.3602, "step": 1144450 }, { "epoch": 6.901146874736194, "grad_norm": 1.12837553024292, "learning_rate": 1.483797483676885e-07, "loss": 3.9746, "step": 1144500 }, { "epoch": 6.901448366517529, "grad_norm": 3.917907238006592, "learning_rate": 1.4747624602456375e-07, "loss": 4.4357, "step": 1144550 }, { "epoch": 6.901749858298863, "grad_norm": 4.168631553649902, "learning_rate": 1.4657550150329323e-07, "loss": 3.7815, "step": 1144600 }, { "epoch": 6.902051350080197, "grad_norm": 1.768903374671936, "learning_rate": 1.4567751482043034e-07, "loss": 4.0031, "step": 1144650 }, { "epoch": 6.902352841861531, "grad_norm": 5.645401954650879, "learning_rate": 1.447822859925285e-07, "loss": 4.4229, "step": 1144700 }, { "epoch": 6.9026543336428645, "grad_norm": 5.157382488250732, "learning_rate": 1.438898150360579e-07, "loss": 3.8789, "step": 1144750 }, { "epoch": 6.902955825424199, "grad_norm": 7.107264995574951, "learning_rate": 1.4300010196743873e-07, "loss": 4.0967, "step": 1144800 }, { "epoch": 6.903257317205533, "grad_norm": 7.030604839324951, "learning_rate": 1.421131468030412e-07, "loss": 4.1504, "step": 1144850 }, { "epoch": 6.903558808986867, "grad_norm": 2.5958330631256104, "learning_rate": 1.4122894955920228e-07, "loss": 3.9897, "step": 1144900 }, { "epoch": 6.903860300768201, "grad_norm": 4.624429225921631, "learning_rate": 1.403475102521756e-07, "loss": 3.5638, "step": 1144950 }, { "epoch": 6.9041617925495355, "grad_norm": 2.1403260231018066, "learning_rate": 1.394688288981982e-07, "loss": 3.8492, "step": 1145000 }, { "epoch": 6.904463284330869, "grad_norm": 3.9222450256347656, "learning_rate": 1.3859290551342383e-07, "loss": 4.1375, "step": 1145050 }, { "epoch": 6.904764776112203, "grad_norm": 5.044134140014648, "learning_rate": 1.3771974011398957e-07, "loss": 3.9729, "step": 1145100 }, { "epoch": 6.905066267893537, "grad_norm": 4.085366725921631, "learning_rate": 1.3684933271594922e-07, "loss": 4.0982, "step": 1145150 }, { "epoch": 6.9053677596748715, "grad_norm": 4.395792007446289, "learning_rate": 1.3598168333534e-07, "loss": 3.9552, "step": 1145200 }, { "epoch": 6.905669251456205, "grad_norm": 2.561924695968628, "learning_rate": 1.3511679198811575e-07, "loss": 4.1596, "step": 1145250 }, { "epoch": 6.905970743237539, "grad_norm": 5.80685567855835, "learning_rate": 1.3425465869021378e-07, "loss": 4.4528, "step": 1145300 }, { "epoch": 6.906272235018873, "grad_norm": 5.215059757232666, "learning_rate": 1.333952834574714e-07, "loss": 4.2558, "step": 1145350 }, { "epoch": 6.906573726800207, "grad_norm": 6.133018493652344, "learning_rate": 1.325386663057093e-07, "loss": 4.0919, "step": 1145400 }, { "epoch": 6.906875218581542, "grad_norm": 4.3618879318237305, "learning_rate": 1.3168480725071484e-07, "loss": 4.3606, "step": 1145450 }, { "epoch": 6.907176710362876, "grad_norm": 3.3804433345794678, "learning_rate": 1.3083370630819213e-07, "loss": 3.7821, "step": 1145500 }, { "epoch": 6.90747820214421, "grad_norm": 4.120574474334717, "learning_rate": 1.299853634937953e-07, "loss": 4.2695, "step": 1145550 }, { "epoch": 6.907779693925543, "grad_norm": 2.065537214279175, "learning_rate": 1.2913977882312853e-07, "loss": 3.8033, "step": 1145600 }, { "epoch": 6.908081185706878, "grad_norm": 4.148277759552002, "learning_rate": 1.282969523117794e-07, "loss": 4.2018, "step": 1145650 }, { "epoch": 6.908382677488212, "grad_norm": 6.351994514465332, "learning_rate": 1.274568839752521e-07, "loss": 4.3281, "step": 1145700 }, { "epoch": 6.908684169269546, "grad_norm": 6.111281394958496, "learning_rate": 1.2661957382898436e-07, "loss": 4.0536, "step": 1145750 }, { "epoch": 6.908985661050879, "grad_norm": 7.403616428375244, "learning_rate": 1.2578502188841377e-07, "loss": 4.1683, "step": 1145800 }, { "epoch": 6.9092871528322135, "grad_norm": 1.326851487159729, "learning_rate": 1.2495322816889474e-07, "loss": 4.1991, "step": 1145850 }, { "epoch": 6.909588644613548, "grad_norm": 3.5006093978881836, "learning_rate": 1.2412419268571506e-07, "loss": 4.0388, "step": 1145900 }, { "epoch": 6.909890136394882, "grad_norm": 4.6142778396606445, "learning_rate": 1.232979154541458e-07, "loss": 3.5911, "step": 1145950 }, { "epoch": 6.910191628176216, "grad_norm": 4.960157871246338, "learning_rate": 1.224743964893915e-07, "loss": 4.1244, "step": 1146000 }, { "epoch": 6.910191628176216, "eval_loss": 4.782312393188477, "eval_runtime": 39.2102, "eval_samples_per_second": 13.058, "eval_steps_per_second": 6.529, "eval_tts_loss": 8.167014797683573, "step": 1146000 }, { "epoch": 6.91049311995755, "grad_norm": 6.374642848968506, "learning_rate": 1.2165363580662335e-07, "loss": 4.3032, "step": 1146050 }, { "epoch": 6.910794611738884, "grad_norm": 1.6824082136154175, "learning_rate": 1.2083563342092927e-07, "loss": 4.1832, "step": 1146100 }, { "epoch": 6.911096103520218, "grad_norm": 5.300570487976074, "learning_rate": 1.2002038934734727e-07, "loss": 4.1607, "step": 1146150 }, { "epoch": 6.911397595301552, "grad_norm": 4.456806659698486, "learning_rate": 1.1920790360091525e-07, "loss": 4.345, "step": 1146200 }, { "epoch": 6.911699087082886, "grad_norm": 4.102042198181152, "learning_rate": 1.1839817619657132e-07, "loss": 4.3426, "step": 1146250 }, { "epoch": 6.9120005788642205, "grad_norm": 5.005568027496338, "learning_rate": 1.1759120714920356e-07, "loss": 3.9092, "step": 1146300 }, { "epoch": 6.912302070645554, "grad_norm": 6.620627403259277, "learning_rate": 1.1678699647370005e-07, "loss": 3.9008, "step": 1146350 }, { "epoch": 6.912603562426888, "grad_norm": 4.357046604156494, "learning_rate": 1.1598554418481564e-07, "loss": 4.2195, "step": 1146400 }, { "epoch": 6.912905054208222, "grad_norm": 6.533871173858643, "learning_rate": 1.1518685029732189e-07, "loss": 3.7523, "step": 1146450 }, { "epoch": 6.913206545989556, "grad_norm": 4.195822238922119, "learning_rate": 1.1439091482592367e-07, "loss": 4.3032, "step": 1146500 }, { "epoch": 6.913508037770891, "grad_norm": 4.559008598327637, "learning_rate": 1.1359773778527592e-07, "loss": 4.4243, "step": 1146550 }, { "epoch": 6.913809529552225, "grad_norm": 4.592729091644287, "learning_rate": 1.1280731918993368e-07, "loss": 4.3034, "step": 1146600 }, { "epoch": 6.914111021333558, "grad_norm": 4.112634181976318, "learning_rate": 1.1201965905450194e-07, "loss": 4.1322, "step": 1146650 }, { "epoch": 6.914412513114892, "grad_norm": 3.9741945266723633, "learning_rate": 1.1123475739341914e-07, "loss": 4.4305, "step": 1146700 }, { "epoch": 6.914714004896227, "grad_norm": 4.448737144470215, "learning_rate": 1.1045261422117369e-07, "loss": 4.3576, "step": 1146750 }, { "epoch": 6.915015496677561, "grad_norm": 4.296876430511475, "learning_rate": 1.0967322955213742e-07, "loss": 3.838, "step": 1146800 }, { "epoch": 6.915316988458895, "grad_norm": 3.3190481662750244, "learning_rate": 1.0889660340066552e-07, "loss": 4.1394, "step": 1146850 }, { "epoch": 6.915618480240228, "grad_norm": 3.797161817550659, "learning_rate": 1.0812273578104658e-07, "loss": 4.1844, "step": 1146900 }, { "epoch": 6.9159199720215625, "grad_norm": 1.525134563446045, "learning_rate": 1.0735162670751918e-07, "loss": 3.8442, "step": 1146950 }, { "epoch": 6.916221463802897, "grad_norm": 1.5184028148651123, "learning_rate": 1.0658327619427199e-07, "loss": 4.0341, "step": 1147000 }, { "epoch": 6.916522955584231, "grad_norm": 4.951181411743164, "learning_rate": 1.0581768425546034e-07, "loss": 4.0236, "step": 1147050 }, { "epoch": 6.916824447365565, "grad_norm": 6.796804428100586, "learning_rate": 1.050548509051563e-07, "loss": 4.1528, "step": 1147100 }, { "epoch": 6.917125939146899, "grad_norm": 5.930273056030273, "learning_rate": 1.0429477615741533e-07, "loss": 4.1723, "step": 1147150 }, { "epoch": 6.917427430928233, "grad_norm": 4.083554267883301, "learning_rate": 1.0353746002620955e-07, "loss": 4.2222, "step": 1147200 }, { "epoch": 6.917728922709567, "grad_norm": 3.2631146907806396, "learning_rate": 1.0278290252547783e-07, "loss": 3.9145, "step": 1147250 }, { "epoch": 6.918030414490901, "grad_norm": 6.432422161102295, "learning_rate": 1.0203110366910904e-07, "loss": 3.7386, "step": 1147300 }, { "epoch": 6.918331906272235, "grad_norm": 4.2328925132751465, "learning_rate": 1.012820634709588e-07, "loss": 4.401, "step": 1147350 }, { "epoch": 6.918633398053569, "grad_norm": 5.206520080566406, "learning_rate": 1.0053578194478274e-07, "loss": 4.0117, "step": 1147400 }, { "epoch": 6.918934889834903, "grad_norm": 5.379257678985596, "learning_rate": 9.979225910433652e-08, "loss": 4.1957, "step": 1147450 }, { "epoch": 6.919236381616237, "grad_norm": 4.364268779754639, "learning_rate": 9.905149496327591e-08, "loss": 4.6041, "step": 1147500 }, { "epoch": 6.919537873397571, "grad_norm": 6.947967052459717, "learning_rate": 9.831348953527329e-08, "loss": 4.2499, "step": 1147550 }, { "epoch": 6.919839365178905, "grad_norm": 6.2209601402282715, "learning_rate": 9.757824283388449e-08, "loss": 3.576, "step": 1147600 }, { "epoch": 6.92014085696024, "grad_norm": 5.239034175872803, "learning_rate": 9.684575487263202e-08, "loss": 4.4534, "step": 1147650 }, { "epoch": 6.920442348741573, "grad_norm": 4.631996154785156, "learning_rate": 9.61160256650384e-08, "loss": 4.1526, "step": 1147700 }, { "epoch": 6.920743840522907, "grad_norm": 5.602205753326416, "learning_rate": 9.538905522447626e-08, "loss": 4.0519, "step": 1147750 }, { "epoch": 6.921045332304241, "grad_norm": 7.067890644073486, "learning_rate": 9.466484356436821e-08, "loss": 4.4796, "step": 1147800 }, { "epoch": 6.9213468240855756, "grad_norm": 4.559118270874023, "learning_rate": 9.394339069802026e-08, "loss": 4.1326, "step": 1147850 }, { "epoch": 6.92164831586691, "grad_norm": 6.78798770904541, "learning_rate": 9.322469663872179e-08, "loss": 4.1683, "step": 1147900 }, { "epoch": 6.921949807648243, "grad_norm": 4.451663017272949, "learning_rate": 9.250876139969554e-08, "loss": 4.2275, "step": 1147950 }, { "epoch": 6.922251299429577, "grad_norm": 5.604773044586182, "learning_rate": 9.179558499411433e-08, "loss": 4.0915, "step": 1148000 }, { "epoch": 6.9225527912109115, "grad_norm": 6.075748920440674, "learning_rate": 9.108516743510097e-08, "loss": 4.0552, "step": 1148050 }, { "epoch": 6.922854282992246, "grad_norm": 6.965340614318848, "learning_rate": 9.037750873574501e-08, "loss": 4.3554, "step": 1148100 }, { "epoch": 6.92315577477358, "grad_norm": 4.197390079498291, "learning_rate": 8.967260890903604e-08, "loss": 4.012, "step": 1148150 }, { "epoch": 6.923457266554914, "grad_norm": 3.2713472843170166, "learning_rate": 8.8970467967997e-08, "loss": 4.295, "step": 1148200 }, { "epoch": 6.923758758336247, "grad_norm": 2.1084916591644287, "learning_rate": 8.827108592550092e-08, "loss": 4.2401, "step": 1148250 }, { "epoch": 6.924060250117582, "grad_norm": 6.724650859832764, "learning_rate": 8.757446279445413e-08, "loss": 4.36, "step": 1148300 }, { "epoch": 6.924361741898916, "grad_norm": 4.166863441467285, "learning_rate": 8.68805985876464e-08, "loss": 4.0173, "step": 1148350 }, { "epoch": 6.92466323368025, "grad_norm": 6.459380626678467, "learning_rate": 8.618949331788417e-08, "loss": 3.968, "step": 1148400 }, { "epoch": 6.924964725461584, "grad_norm": 7.158838272094727, "learning_rate": 8.55011469978406e-08, "loss": 3.9338, "step": 1148450 }, { "epoch": 6.925266217242918, "grad_norm": 11.255460739135742, "learning_rate": 8.481555964023889e-08, "loss": 4.254, "step": 1148500 }, { "epoch": 6.925567709024252, "grad_norm": 4.011597156524658, "learning_rate": 8.413273125763564e-08, "loss": 4.012, "step": 1148550 }, { "epoch": 6.925869200805586, "grad_norm": 6.555157661437988, "learning_rate": 8.345266186265409e-08, "loss": 4.0682, "step": 1148600 }, { "epoch": 6.92617069258692, "grad_norm": 5.011341094970703, "learning_rate": 8.27753514677676e-08, "loss": 4.2804, "step": 1148650 }, { "epoch": 6.926472184368254, "grad_norm": 4.953328609466553, "learning_rate": 8.210080008546616e-08, "loss": 4.1136, "step": 1148700 }, { "epoch": 6.926773676149589, "grad_norm": 1.9181530475616455, "learning_rate": 8.142900772815652e-08, "loss": 4.141, "step": 1148750 }, { "epoch": 6.927075167930922, "grad_norm": 5.567940711975098, "learning_rate": 8.075997440819549e-08, "loss": 4.301, "step": 1148800 }, { "epoch": 6.927376659712256, "grad_norm": 4.907054901123047, "learning_rate": 8.009370013788984e-08, "loss": 4.3119, "step": 1148850 }, { "epoch": 6.92767815149359, "grad_norm": 6.218924045562744, "learning_rate": 7.94301849295298e-08, "loss": 3.919, "step": 1148900 }, { "epoch": 6.9279796432749245, "grad_norm": 1.9772869348526, "learning_rate": 7.876942879530557e-08, "loss": 4.0204, "step": 1148950 }, { "epoch": 6.928281135056259, "grad_norm": 6.146270275115967, "learning_rate": 7.811143174737411e-08, "loss": 4.149, "step": 1149000 }, { "epoch": 6.928281135056259, "eval_loss": 4.7824506759643555, "eval_runtime": 38.9939, "eval_samples_per_second": 13.13, "eval_steps_per_second": 6.565, "eval_tts_loss": 8.168793519551267, "step": 1149000 }, { "epoch": 6.928582626837592, "grad_norm": 4.207578182220459, "learning_rate": 7.745619379785906e-08, "loss": 4.116, "step": 1149050 }, { "epoch": 6.928884118618926, "grad_norm": 5.454368591308594, "learning_rate": 7.680371495881743e-08, "loss": 4.5226, "step": 1149100 }, { "epoch": 6.9291856104002605, "grad_norm": 2.111593246459961, "learning_rate": 7.615399524223963e-08, "loss": 3.9871, "step": 1149150 }, { "epoch": 6.929487102181595, "grad_norm": 8.527994155883789, "learning_rate": 7.550703466009944e-08, "loss": 4.3088, "step": 1149200 }, { "epoch": 6.929788593962929, "grad_norm": 5.2904157638549805, "learning_rate": 7.486283322428733e-08, "loss": 4.2461, "step": 1149250 }, { "epoch": 6.930090085744263, "grad_norm": 6.462740421295166, "learning_rate": 7.422139094669377e-08, "loss": 3.835, "step": 1149300 }, { "epoch": 6.930391577525596, "grad_norm": 1.3353164196014404, "learning_rate": 7.358270783907605e-08, "loss": 4.1227, "step": 1149350 }, { "epoch": 6.930693069306931, "grad_norm": 4.453554153442383, "learning_rate": 7.294678391322473e-08, "loss": 4.1248, "step": 1149400 }, { "epoch": 6.930994561088265, "grad_norm": 4.494970321655273, "learning_rate": 7.231361918083045e-08, "loss": 4.2021, "step": 1149450 }, { "epoch": 6.931296052869599, "grad_norm": 4.519551753997803, "learning_rate": 7.168321365353391e-08, "loss": 4.1956, "step": 1149500 }, { "epoch": 6.931597544650932, "grad_norm": 10.511944770812988, "learning_rate": 7.105556734295914e-08, "loss": 4.0067, "step": 1149550 }, { "epoch": 6.931899036432267, "grad_norm": 4.1927947998046875, "learning_rate": 7.043068026064691e-08, "loss": 4.01, "step": 1149600 }, { "epoch": 6.932200528213601, "grad_norm": 6.240715026855469, "learning_rate": 6.980855241810468e-08, "loss": 4.2167, "step": 1149650 }, { "epoch": 6.932502019994935, "grad_norm": 3.850280523300171, "learning_rate": 6.918918382675665e-08, "loss": 3.9612, "step": 1149700 }, { "epoch": 6.932803511776269, "grad_norm": 4.838010311126709, "learning_rate": 6.8572574498027e-08, "loss": 4.1648, "step": 1149750 }, { "epoch": 6.933105003557603, "grad_norm": 6.189878463745117, "learning_rate": 6.795872444325667e-08, "loss": 3.9541, "step": 1149800 }, { "epoch": 6.933406495338937, "grad_norm": 3.5870091915130615, "learning_rate": 6.73476336737533e-08, "loss": 3.8832, "step": 1149850 }, { "epoch": 6.933707987120271, "grad_norm": 4.593843460083008, "learning_rate": 6.673930220072454e-08, "loss": 4.132, "step": 1149900 }, { "epoch": 6.934009478901605, "grad_norm": 6.991671562194824, "learning_rate": 6.613373003541144e-08, "loss": 4.3336, "step": 1149950 }, { "epoch": 6.934310970682939, "grad_norm": 3.9716439247131348, "learning_rate": 6.553091718893844e-08, "loss": 4.0931, "step": 1150000 }, { "epoch": 6.9346124624642735, "grad_norm": 4.423027038574219, "learning_rate": 6.493086367239664e-08, "loss": 4.0873, "step": 1150050 }, { "epoch": 6.934913954245607, "grad_norm": 4.526165008544922, "learning_rate": 6.433356949682723e-08, "loss": 4.2789, "step": 1150100 }, { "epoch": 6.935215446026941, "grad_norm": 10.725101470947266, "learning_rate": 6.373903467323804e-08, "loss": 4.4358, "step": 1150150 }, { "epoch": 6.935516937808275, "grad_norm": 4.442194938659668, "learning_rate": 6.314725921257036e-08, "loss": 4.2427, "step": 1150200 }, { "epoch": 6.9358184295896095, "grad_norm": 6.272726535797119, "learning_rate": 6.255824312568214e-08, "loss": 4.4075, "step": 1150250 }, { "epoch": 6.936119921370944, "grad_norm": 4.288527011871338, "learning_rate": 6.197198642346468e-08, "loss": 4.2833, "step": 1150300 }, { "epoch": 6.936421413152278, "grad_norm": 4.353768348693848, "learning_rate": 6.138848911665939e-08, "loss": 4.4861, "step": 1150350 }, { "epoch": 6.936722904933611, "grad_norm": 4.466461181640625, "learning_rate": 6.080775121602433e-08, "loss": 3.878, "step": 1150400 }, { "epoch": 6.937024396714945, "grad_norm": 2.5923779010772705, "learning_rate": 6.022977273226759e-08, "loss": 3.5663, "step": 1150450 }, { "epoch": 6.93732588849628, "grad_norm": 9.219732284545898, "learning_rate": 5.965455367598071e-08, "loss": 4.3655, "step": 1150500 }, { "epoch": 6.937627380277614, "grad_norm": 4.6056413650512695, "learning_rate": 5.908209405778852e-08, "loss": 4.5935, "step": 1150550 }, { "epoch": 6.937928872058948, "grad_norm": 1.6401313543319702, "learning_rate": 5.851239388821594e-08, "loss": 4.2198, "step": 1150600 }, { "epoch": 6.938230363840281, "grad_norm": 6.07413911819458, "learning_rate": 5.794545317773791e-08, "loss": 4.1079, "step": 1150650 }, { "epoch": 6.938531855621616, "grad_norm": 1.371333360671997, "learning_rate": 5.738127193681274e-08, "loss": 3.9344, "step": 1150700 }, { "epoch": 6.93883334740295, "grad_norm": 4.857444763183594, "learning_rate": 5.681985017578217e-08, "loss": 3.7056, "step": 1150750 }, { "epoch": 6.939134839184284, "grad_norm": 6.751809120178223, "learning_rate": 5.6261187905021207e-08, "loss": 4.4445, "step": 1150800 }, { "epoch": 6.939436330965618, "grad_norm": 8.704606056213379, "learning_rate": 5.570528513477168e-08, "loss": 4.0991, "step": 1150850 }, { "epoch": 6.939737822746952, "grad_norm": 5.1956377029418945, "learning_rate": 5.515214187529204e-08, "loss": 4.2837, "step": 1150900 }, { "epoch": 6.940039314528286, "grad_norm": 4.9561662673950195, "learning_rate": 5.460175813675749e-08, "loss": 4.4909, "step": 1150950 }, { "epoch": 6.94034080630962, "grad_norm": 3.1928763389587402, "learning_rate": 5.4054133929293255e-08, "loss": 4.1752, "step": 1151000 }, { "epoch": 6.940642298090954, "grad_norm": 6.070968151092529, "learning_rate": 5.350926926299126e-08, "loss": 4.0793, "step": 1151050 }, { "epoch": 6.940943789872288, "grad_norm": 6.791683197021484, "learning_rate": 5.296716414784352e-08, "loss": 4.3299, "step": 1151100 }, { "epoch": 6.941245281653622, "grad_norm": 6.189631462097168, "learning_rate": 5.2427818593858697e-08, "loss": 4.2019, "step": 1151150 }, { "epoch": 6.941546773434956, "grad_norm": 6.659267902374268, "learning_rate": 5.189123261096218e-08, "loss": 4.1427, "step": 1151200 }, { "epoch": 6.94184826521629, "grad_norm": 3.710251569747925, "learning_rate": 5.135740620901274e-08, "loss": 4.22, "step": 1151250 }, { "epoch": 6.942149756997624, "grad_norm": 1.6287786960601807, "learning_rate": 5.082633939783587e-08, "loss": 4.0479, "step": 1151300 }, { "epoch": 6.9424512487789585, "grad_norm": 4.161162853240967, "learning_rate": 5.0298032187223724e-08, "loss": 4.3917, "step": 1151350 }, { "epoch": 6.942752740560293, "grad_norm": 8.819005012512207, "learning_rate": 4.977248458686856e-08, "loss": 4.2649, "step": 1151400 }, { "epoch": 6.943054232341626, "grad_norm": 4.038864612579346, "learning_rate": 4.924969660647926e-08, "loss": 4.1382, "step": 1151450 }, { "epoch": 6.94335572412296, "grad_norm": 7.6347432136535645, "learning_rate": 4.872966825564817e-08, "loss": 4.2902, "step": 1151500 }, { "epoch": 6.943657215904294, "grad_norm": 4.501208305358887, "learning_rate": 4.821239954395095e-08, "loss": 3.7708, "step": 1151550 }, { "epoch": 6.943958707685629, "grad_norm": 4.297950744628906, "learning_rate": 4.769789048092998e-08, "loss": 4.2185, "step": 1151600 }, { "epoch": 6.944260199466963, "grad_norm": 6.2066497802734375, "learning_rate": 4.718614107601104e-08, "loss": 4.0641, "step": 1151650 }, { "epoch": 6.944561691248296, "grad_norm": 4.167759895324707, "learning_rate": 4.667715133865324e-08, "loss": 4.1857, "step": 1151700 }, { "epoch": 6.94486318302963, "grad_norm": 5.164578437805176, "learning_rate": 4.61709212781991e-08, "loss": 4.4307, "step": 1151750 }, { "epoch": 6.9451646748109646, "grad_norm": 3.164074182510376, "learning_rate": 4.5667450903991154e-08, "loss": 4.0148, "step": 1151800 }, { "epoch": 6.945466166592299, "grad_norm": 2.623875141143799, "learning_rate": 4.516674022525535e-08, "loss": 4.0523, "step": 1151850 }, { "epoch": 6.945767658373633, "grad_norm": 4.945243835449219, "learning_rate": 4.46687892512343e-08, "loss": 4.0974, "step": 1151900 }, { "epoch": 6.946069150154967, "grad_norm": 3.4266550540924072, "learning_rate": 4.4173597991103983e-08, "loss": 4.2116, "step": 1151950 }, { "epoch": 6.9463706419363005, "grad_norm": 3.3703818321228027, "learning_rate": 4.368116645394049e-08, "loss": 4.288, "step": 1152000 }, { "epoch": 6.9463706419363005, "eval_loss": 4.782382488250732, "eval_runtime": 39.2318, "eval_samples_per_second": 13.051, "eval_steps_per_second": 6.525, "eval_tts_loss": 8.16800971006036, "step": 1152000 }, { "epoch": 6.946672133717635, "grad_norm": 5.101541519165039, "learning_rate": 4.319149464883653e-08, "loss": 4.1576, "step": 1152050 }, { "epoch": 6.946973625498969, "grad_norm": 5.898940086364746, "learning_rate": 4.270458258478493e-08, "loss": 4.2444, "step": 1152100 }, { "epoch": 6.947275117280303, "grad_norm": 4.352173328399658, "learning_rate": 4.222043027076183e-08, "loss": 4.3456, "step": 1152150 }, { "epoch": 6.947576609061637, "grad_norm": 4.76844596862793, "learning_rate": 4.173903771566012e-08, "loss": 4.5061, "step": 1152200 }, { "epoch": 6.947878100842971, "grad_norm": 1.8733909130096436, "learning_rate": 4.126040492835603e-08, "loss": 4.0661, "step": 1152250 }, { "epoch": 6.948179592624305, "grad_norm": 6.118303298950195, "learning_rate": 4.078453191765918e-08, "loss": 4.2521, "step": 1152300 }, { "epoch": 6.948481084405639, "grad_norm": 4.681636810302734, "learning_rate": 4.031141869231258e-08, "loss": 4.4375, "step": 1152350 }, { "epoch": 6.948782576186973, "grad_norm": 5.608218193054199, "learning_rate": 3.984106526102593e-08, "loss": 4.4403, "step": 1152400 }, { "epoch": 6.9490840679683075, "grad_norm": 5.087264060974121, "learning_rate": 3.9373471632475616e-08, "loss": 4.0588, "step": 1152450 }, { "epoch": 6.949385559749642, "grad_norm": 5.791478157043457, "learning_rate": 3.890863781523812e-08, "loss": 4.4435, "step": 1152500 }, { "epoch": 6.949687051530975, "grad_norm": 1.6737614870071411, "learning_rate": 3.844656381787325e-08, "loss": 4.5097, "step": 1152550 }, { "epoch": 6.949988543312309, "grad_norm": 4.901989459991455, "learning_rate": 3.798724964890754e-08, "loss": 4.0185, "step": 1152600 }, { "epoch": 6.950290035093643, "grad_norm": 6.451139450073242, "learning_rate": 3.753069531678421e-08, "loss": 4.2154, "step": 1152650 }, { "epoch": 6.950591526874978, "grad_norm": 5.135551929473877, "learning_rate": 3.7076900829879904e-08, "loss": 4.0858, "step": 1152700 }, { "epoch": 6.950893018656312, "grad_norm": 4.586382865905762, "learning_rate": 3.662586619658791e-08, "loss": 3.9222, "step": 1152750 }, { "epoch": 6.951194510437645, "grad_norm": 5.571926593780518, "learning_rate": 3.6177591425168294e-08, "loss": 4.2229, "step": 1152800 }, { "epoch": 6.951496002218979, "grad_norm": 3.811012029647827, "learning_rate": 3.5732076523897756e-08, "loss": 3.9261, "step": 1152850 }, { "epoch": 6.9517974940003135, "grad_norm": 5.476987838745117, "learning_rate": 3.528932150096975e-08, "loss": 4.3343, "step": 1152900 }, { "epoch": 6.952098985781648, "grad_norm": 4.54806661605835, "learning_rate": 3.484932636452775e-08, "loss": 4.2375, "step": 1152950 }, { "epoch": 6.952400477562982, "grad_norm": 5.426884174346924, "learning_rate": 3.441209112268195e-08, "loss": 4.0633, "step": 1153000 }, { "epoch": 6.952701969344316, "grad_norm": 6.249197006225586, "learning_rate": 3.3977615783459255e-08, "loss": 4.1587, "step": 1153050 }, { "epoch": 6.9530034611256495, "grad_norm": 3.149674654006958, "learning_rate": 3.354590035486993e-08, "loss": 4.2402, "step": 1153100 }, { "epoch": 6.953304952906984, "grad_norm": 8.129440307617188, "learning_rate": 3.311694484485761e-08, "loss": 4.0799, "step": 1153150 }, { "epoch": 6.953606444688318, "grad_norm": 5.509926795959473, "learning_rate": 3.269074926129933e-08, "loss": 3.8852, "step": 1153200 }, { "epoch": 6.953907936469652, "grad_norm": 4.3038225173950195, "learning_rate": 3.226731361207213e-08, "loss": 4.2128, "step": 1153250 }, { "epoch": 6.954209428250985, "grad_norm": 5.746734142303467, "learning_rate": 3.1846637904936464e-08, "loss": 4.5083, "step": 1153300 }, { "epoch": 6.95451092003232, "grad_norm": 2.7956185340881348, "learning_rate": 3.14287221476528e-08, "loss": 4.3269, "step": 1153350 }, { "epoch": 6.954812411813654, "grad_norm": 4.802638053894043, "learning_rate": 3.101356634789831e-08, "loss": 3.9238, "step": 1153400 }, { "epoch": 6.955113903594988, "grad_norm": 5.295772075653076, "learning_rate": 3.060117051333355e-08, "loss": 4.5188, "step": 1153450 }, { "epoch": 6.955415395376322, "grad_norm": 4.431416988372803, "learning_rate": 3.0191534651535785e-08, "loss": 4.4013, "step": 1153500 }, { "epoch": 6.9557168871576565, "grad_norm": 2.215867042541504, "learning_rate": 2.9784658770032334e-08, "loss": 4.3817, "step": 1153550 }, { "epoch": 6.95601837893899, "grad_norm": 5.680881500244141, "learning_rate": 2.9380542876333845e-08, "loss": 3.8641, "step": 1153600 }, { "epoch": 6.956319870720324, "grad_norm": 2.1821773052215576, "learning_rate": 2.8979186977867718e-08, "loss": 4.0438, "step": 1153650 }, { "epoch": 6.956621362501658, "grad_norm": 5.2685699462890625, "learning_rate": 2.8580591082011383e-08, "loss": 4.4203, "step": 1153700 }, { "epoch": 6.956922854282992, "grad_norm": 6.154561996459961, "learning_rate": 2.8184755196108964e-08, "loss": 3.915, "step": 1153750 }, { "epoch": 6.957224346064327, "grad_norm": 7.972964286804199, "learning_rate": 2.779167932745463e-08, "loss": 3.9334, "step": 1153800 }, { "epoch": 6.95752583784566, "grad_norm": 6.959815979003906, "learning_rate": 2.740136348327593e-08, "loss": 4.1535, "step": 1153850 }, { "epoch": 6.957827329626994, "grad_norm": 4.791515827178955, "learning_rate": 2.70138076707338e-08, "loss": 4.0873, "step": 1153900 }, { "epoch": 6.958128821408328, "grad_norm": 4.64670991897583, "learning_rate": 2.6629011896989184e-08, "loss": 4.2403, "step": 1153950 }, { "epoch": 6.9584303131896625, "grad_norm": 4.323456287384033, "learning_rate": 2.624697616911975e-08, "loss": 4.3455, "step": 1154000 }, { "epoch": 6.958731804970997, "grad_norm": 4.058629512786865, "learning_rate": 2.586770049415321e-08, "loss": 3.887, "step": 1154050 }, { "epoch": 6.959033296752331, "grad_norm": 5.486511707305908, "learning_rate": 2.5491184879050662e-08, "loss": 4.0196, "step": 1154100 }, { "epoch": 6.959334788533664, "grad_norm": 9.23521900177002, "learning_rate": 2.511742933077321e-08, "loss": 4.2466, "step": 1154150 }, { "epoch": 6.9596362803149985, "grad_norm": 7.341477394104004, "learning_rate": 2.474643385616537e-08, "loss": 4.1599, "step": 1154200 }, { "epoch": 6.959937772096333, "grad_norm": 2.751579761505127, "learning_rate": 2.4378198462088328e-08, "loss": 4.3505, "step": 1154250 }, { "epoch": 6.960239263877667, "grad_norm": 7.565325736999512, "learning_rate": 2.4012723155303338e-08, "loss": 3.9213, "step": 1154300 }, { "epoch": 6.960540755659001, "grad_norm": 6.169339179992676, "learning_rate": 2.3650007942538353e-08, "loss": 3.8058, "step": 1154350 }, { "epoch": 6.960842247440334, "grad_norm": 7.111403942108154, "learning_rate": 2.3290052830454708e-08, "loss": 4.0753, "step": 1154400 }, { "epoch": 6.961143739221669, "grad_norm": 11.635320663452148, "learning_rate": 2.2932857825697092e-08, "loss": 4.1694, "step": 1154450 }, { "epoch": 6.961445231003003, "grad_norm": 5.4062700271606445, "learning_rate": 2.2578422934843577e-08, "loss": 4.06, "step": 1154500 }, { "epoch": 6.961746722784337, "grad_norm": 6.187119007110596, "learning_rate": 2.2226748164388965e-08, "loss": 4.3374, "step": 1154550 }, { "epoch": 6.962048214565671, "grad_norm": 3.7226014137268066, "learning_rate": 2.1877833520828058e-08, "loss": 4.1121, "step": 1154600 }, { "epoch": 6.9623497063470055, "grad_norm": 4.805074691772461, "learning_rate": 2.15316790105724e-08, "loss": 4.2031, "step": 1154650 }, { "epoch": 6.962651198128339, "grad_norm": 5.703063011169434, "learning_rate": 2.118828464000022e-08, "loss": 3.9334, "step": 1154700 }, { "epoch": 6.962952689909673, "grad_norm": 5.518510818481445, "learning_rate": 2.084765041540648e-08, "loss": 4.3811, "step": 1154750 }, { "epoch": 6.963254181691007, "grad_norm": 4.527337074279785, "learning_rate": 2.0509776343102803e-08, "loss": 4.0217, "step": 1154800 }, { "epoch": 6.963555673472341, "grad_norm": 3.0101728439331055, "learning_rate": 2.0174662429284227e-08, "loss": 4.0687, "step": 1154850 }, { "epoch": 6.963857165253675, "grad_norm": 6.257022857666016, "learning_rate": 1.9842308680112494e-08, "loss": 4.226, "step": 1154900 }, { "epoch": 6.964158657035009, "grad_norm": 3.198573589324951, "learning_rate": 1.9512715101699384e-08, "loss": 3.9035, "step": 1154950 }, { "epoch": 6.964460148816343, "grad_norm": 2.553602933883667, "learning_rate": 1.9185881700140016e-08, "loss": 4.0334, "step": 1155000 }, { "epoch": 6.964460148816343, "eval_loss": 4.7823286056518555, "eval_runtime": 39.3257, "eval_samples_per_second": 13.019, "eval_steps_per_second": 6.51, "eval_tts_loss": 8.167234025587483, "step": 1155000 }, { "epoch": 6.964761640597677, "grad_norm": 5.926784515380859, "learning_rate": 1.8861808481412945e-08, "loss": 4.1692, "step": 1155050 }, { "epoch": 6.9650631323790115, "grad_norm": 7.326537132263184, "learning_rate": 1.854049545151337e-08, "loss": 4.0665, "step": 1155100 }, { "epoch": 6.965364624160346, "grad_norm": 4.221405982971191, "learning_rate": 1.8221942616336583e-08, "loss": 4.3515, "step": 1155150 }, { "epoch": 6.965666115941679, "grad_norm": 5.426081657409668, "learning_rate": 1.7906149981744554e-08, "loss": 3.9827, "step": 1155200 }, { "epoch": 6.965967607723013, "grad_norm": 5.4667439460754395, "learning_rate": 1.7593117553549307e-08, "loss": 4.5889, "step": 1155250 }, { "epoch": 6.9662690995043475, "grad_norm": 8.145644187927246, "learning_rate": 1.728284533752955e-08, "loss": 4.4457, "step": 1155300 }, { "epoch": 6.966570591285682, "grad_norm": 4.255272388458252, "learning_rate": 1.6975333339380726e-08, "loss": 4.3737, "step": 1155350 }, { "epoch": 6.966872083067016, "grad_norm": 5.5698676109313965, "learning_rate": 1.6670581564764974e-08, "loss": 4.0447, "step": 1155400 }, { "epoch": 6.967173574848349, "grad_norm": 6.209813594818115, "learning_rate": 1.6368590019277816e-08, "loss": 4.1948, "step": 1155450 }, { "epoch": 6.967475066629683, "grad_norm": 6.618574142456055, "learning_rate": 1.6069358708498126e-08, "loss": 4.0288, "step": 1155500 }, { "epoch": 6.967776558411018, "grad_norm": 4.418670654296875, "learning_rate": 1.5772887637921505e-08, "loss": 4.1425, "step": 1155550 }, { "epoch": 6.968078050192352, "grad_norm": 6.238652229309082, "learning_rate": 1.54791768129936e-08, "loss": 4.0433, "step": 1155600 }, { "epoch": 6.968379541973686, "grad_norm": 6.670811176300049, "learning_rate": 1.5188226239143397e-08, "loss": 4.0871, "step": 1155650 }, { "epoch": 6.96868103375502, "grad_norm": 4.962444305419922, "learning_rate": 1.4900035921716624e-08, "loss": 4.3589, "step": 1155700 }, { "epoch": 6.968982525536354, "grad_norm": 5.859117031097412, "learning_rate": 1.4614605866009043e-08, "loss": 4.1684, "step": 1155750 }, { "epoch": 6.969284017317688, "grad_norm": 4.462100505828857, "learning_rate": 1.4331936077283112e-08, "loss": 4.1706, "step": 1155800 }, { "epoch": 6.969585509099022, "grad_norm": 6.810337543487549, "learning_rate": 1.4052026560734675e-08, "loss": 4.2059, "step": 1155850 }, { "epoch": 6.969887000880356, "grad_norm": 4.0569233894348145, "learning_rate": 1.3774877321526267e-08, "loss": 4.2263, "step": 1155900 }, { "epoch": 6.97018849266169, "grad_norm": 5.754085540771484, "learning_rate": 1.350048836473716e-08, "loss": 4.1105, "step": 1155950 }, { "epoch": 6.970489984443024, "grad_norm": 7.360525608062744, "learning_rate": 1.3228859695429972e-08, "loss": 4.2643, "step": 1156000 }, { "epoch": 6.970791476224358, "grad_norm": 5.089325428009033, "learning_rate": 1.2959991318617358e-08, "loss": 4.0689, "step": 1156050 }, { "epoch": 6.971092968005692, "grad_norm": 5.541510581970215, "learning_rate": 1.2693883239212055e-08, "loss": 3.8335, "step": 1156100 }, { "epoch": 6.971394459787026, "grad_norm": 5.0140910148620605, "learning_rate": 1.2430535462143454e-08, "loss": 4.0842, "step": 1156150 }, { "epoch": 6.9716959515683605, "grad_norm": 6.396374225616455, "learning_rate": 1.2169947992257678e-08, "loss": 4.3378, "step": 1156200 }, { "epoch": 6.971997443349695, "grad_norm": 10.28768253326416, "learning_rate": 1.1912120834317584e-08, "loss": 4.2756, "step": 1156250 }, { "epoch": 6.972298935131028, "grad_norm": 7.4603142738342285, "learning_rate": 1.1657053993102682e-08, "loss": 4.3601, "step": 1156300 }, { "epoch": 6.972600426912362, "grad_norm": 5.85689640045166, "learning_rate": 1.1404747473292564e-08, "loss": 4.2913, "step": 1156350 }, { "epoch": 6.9729019186936965, "grad_norm": 3.2133193016052246, "learning_rate": 1.115520127953351e-08, "loss": 4.1918, "step": 1156400 }, { "epoch": 6.973203410475031, "grad_norm": 7.292247295379639, "learning_rate": 1.0908415416421846e-08, "loss": 4.1585, "step": 1156450 }, { "epoch": 6.973504902256365, "grad_norm": 5.691619873046875, "learning_rate": 1.066438988848728e-08, "loss": 3.8762, "step": 1156500 }, { "epoch": 6.973806394037698, "grad_norm": 1.56691312789917, "learning_rate": 1.0423124700242869e-08, "loss": 4.179, "step": 1156550 }, { "epoch": 6.974107885819032, "grad_norm": 5.8067827224731445, "learning_rate": 1.018461985610175e-08, "loss": 4.2926, "step": 1156600 }, { "epoch": 6.974409377600367, "grad_norm": 6.790717601776123, "learning_rate": 9.948875360460406e-09, "loss": 4.261, "step": 1156650 }, { "epoch": 6.974710869381701, "grad_norm": 5.033656120300293, "learning_rate": 9.715891217665361e-09, "loss": 4.3966, "step": 1156700 }, { "epoch": 6.975012361163035, "grad_norm": 4.425131320953369, "learning_rate": 9.485667432013178e-09, "loss": 4.0537, "step": 1156750 }, { "epoch": 6.975313852944369, "grad_norm": 6.458901405334473, "learning_rate": 9.258204007717152e-09, "loss": 4.1848, "step": 1156800 }, { "epoch": 6.9756153447257025, "grad_norm": 3.6940865516662598, "learning_rate": 9.033500948990579e-09, "loss": 4.078, "step": 1156850 }, { "epoch": 6.975916836507037, "grad_norm": 4.534050464630127, "learning_rate": 8.811558259946839e-09, "loss": 4.2381, "step": 1156900 }, { "epoch": 6.976218328288371, "grad_norm": 5.996420860290527, "learning_rate": 8.592375944665997e-09, "loss": 4.2298, "step": 1156950 }, { "epoch": 6.976519820069705, "grad_norm": 6.167656421661377, "learning_rate": 8.375954007194819e-09, "loss": 4.0617, "step": 1157000 }, { "epoch": 6.9768213118510385, "grad_norm": 6.134336948394775, "learning_rate": 8.162292451530106e-09, "loss": 4.1031, "step": 1157050 }, { "epoch": 6.977122803632373, "grad_norm": 4.798277854919434, "learning_rate": 7.951391281585395e-09, "loss": 4.1402, "step": 1157100 }, { "epoch": 6.977424295413707, "grad_norm": 6.026317119598389, "learning_rate": 7.743250501240916e-09, "loss": 4.1098, "step": 1157150 }, { "epoch": 6.977725787195041, "grad_norm": 4.426357746124268, "learning_rate": 7.537870114326938e-09, "loss": 4.3254, "step": 1157200 }, { "epoch": 6.978027278976375, "grad_norm": 5.523722171783447, "learning_rate": 7.335250124623771e-09, "loss": 4.1625, "step": 1157250 }, { "epoch": 6.9783287707577095, "grad_norm": 4.518901348114014, "learning_rate": 7.135390535878416e-09, "loss": 3.7568, "step": 1157300 }, { "epoch": 6.978630262539043, "grad_norm": 2.5616209506988525, "learning_rate": 6.938291351737957e-09, "loss": 4.018, "step": 1157350 }, { "epoch": 6.978931754320377, "grad_norm": 4.69639253616333, "learning_rate": 6.743952575866129e-09, "loss": 4.2154, "step": 1157400 }, { "epoch": 6.979233246101711, "grad_norm": 7.2813849449157715, "learning_rate": 6.552374211810097e-09, "loss": 3.9224, "step": 1157450 }, { "epoch": 6.9795347378830455, "grad_norm": 4.636859893798828, "learning_rate": 6.363556263100367e-09, "loss": 4.297, "step": 1157500 }, { "epoch": 6.97983622966438, "grad_norm": 6.032036304473877, "learning_rate": 6.177498733234143e-09, "loss": 4.3384, "step": 1157550 }, { "epoch": 6.980137721445713, "grad_norm": 5.284134864807129, "learning_rate": 5.9942016255920545e-09, "loss": 4.0495, "step": 1157600 }, { "epoch": 6.980439213227047, "grad_norm": 1.7680131196975708, "learning_rate": 5.8136649435880366e-09, "loss": 3.8634, "step": 1157650 }, { "epoch": 6.980740705008381, "grad_norm": 4.546878814697266, "learning_rate": 5.635888690536106e-09, "loss": 4.2091, "step": 1157700 }, { "epoch": 6.981042196789716, "grad_norm": 2.181340456008911, "learning_rate": 5.460872869700317e-09, "loss": 4.3707, "step": 1157750 }, { "epoch": 6.98134368857105, "grad_norm": 4.098128795623779, "learning_rate": 5.288617484294766e-09, "loss": 4.1093, "step": 1157800 }, { "epoch": 6.981645180352384, "grad_norm": 5.132857322692871, "learning_rate": 5.119122537500242e-09, "loss": 4.2615, "step": 1157850 }, { "epoch": 6.981946672133717, "grad_norm": 4.062800407409668, "learning_rate": 4.952388032447574e-09, "loss": 4.5273, "step": 1157900 }, { "epoch": 6.9822481639150515, "grad_norm": 5.4889092445373535, "learning_rate": 4.788413972184324e-09, "loss": 4.1123, "step": 1157950 }, { "epoch": 6.982549655696386, "grad_norm": 5.211069107055664, "learning_rate": 4.627200359724747e-09, "loss": 4.0374, "step": 1158000 }, { "epoch": 6.982549655696386, "eval_loss": 4.782321453094482, "eval_runtime": 39.0341, "eval_samples_per_second": 13.117, "eval_steps_per_second": 6.558, "eval_tts_loss": 8.167616244365936, "step": 1158000 }, { "epoch": 6.98285114747772, "grad_norm": 3.130988597869873, "learning_rate": 4.468747198049793e-09, "loss": 4.5367, "step": 1158050 }, { "epoch": 6.983152639259054, "grad_norm": 5.875298500061035, "learning_rate": 4.313054490090451e-09, "loss": 4.2573, "step": 1158100 }, { "epoch": 6.9834541310403875, "grad_norm": 6.210086345672607, "learning_rate": 4.160122238677787e-09, "loss": 4.1795, "step": 1158150 }, { "epoch": 6.983755622821722, "grad_norm": 5.1138997077941895, "learning_rate": 4.009950446642873e-09, "loss": 4.1073, "step": 1158200 }, { "epoch": 6.984057114603056, "grad_norm": 4.055771350860596, "learning_rate": 3.862539116750163e-09, "loss": 4.471, "step": 1158250 }, { "epoch": 6.98435860638439, "grad_norm": 6.0168657302856445, "learning_rate": 3.7178882517141516e-09, "loss": 3.9953, "step": 1158300 }, { "epoch": 6.984660098165724, "grad_norm": 5.187364101409912, "learning_rate": 3.575997854199375e-09, "loss": 3.9519, "step": 1158350 }, { "epoch": 6.9849615899470585, "grad_norm": 1.9440969228744507, "learning_rate": 3.4368679268037547e-09, "loss": 4.2647, "step": 1158400 }, { "epoch": 6.985263081728392, "grad_norm": 3.543307065963745, "learning_rate": 3.3004984720919058e-09, "loss": 4.3406, "step": 1158450 }, { "epoch": 6.985564573509726, "grad_norm": 6.869527816772461, "learning_rate": 3.1668894925951373e-09, "loss": 4.4135, "step": 1158500 }, { "epoch": 6.98586606529106, "grad_norm": 5.863481521606445, "learning_rate": 3.0360409907281835e-09, "loss": 4.2594, "step": 1158550 }, { "epoch": 6.9861675570723945, "grad_norm": 1.0975910425186157, "learning_rate": 2.90795296895574e-09, "loss": 4.0053, "step": 1158600 }, { "epoch": 6.986469048853728, "grad_norm": 4.283335208892822, "learning_rate": 2.7826254295759687e-09, "loss": 4.1382, "step": 1158650 }, { "epoch": 6.986770540635062, "grad_norm": 6.5138444900512695, "learning_rate": 2.6600583749536443e-09, "loss": 3.9797, "step": 1158700 }, { "epoch": 6.987072032416396, "grad_norm": 3.4378583431243896, "learning_rate": 2.5402518072870083e-09, "loss": 4.0356, "step": 1158750 }, { "epoch": 6.98737352419773, "grad_norm": 4.818868637084961, "learning_rate": 2.423205728824262e-09, "loss": 4.1105, "step": 1158800 }, { "epoch": 6.987675015979065, "grad_norm": 3.9935035705566406, "learning_rate": 2.3089201417136884e-09, "loss": 4.3676, "step": 1158850 }, { "epoch": 6.987976507760399, "grad_norm": 3.979335069656372, "learning_rate": 2.197395048036954e-09, "loss": 4.1805, "step": 1158900 }, { "epoch": 6.988277999541732, "grad_norm": 10.115994453430176, "learning_rate": 2.0886304498590743e-09, "loss": 4.0318, "step": 1158950 }, { "epoch": 6.988579491323066, "grad_norm": 7.086751461029053, "learning_rate": 1.982626349178451e-09, "loss": 4.2495, "step": 1159000 }, { "epoch": 6.9888809831044005, "grad_norm": 4.784701824188232, "learning_rate": 1.8793827479601786e-09, "loss": 4.4172, "step": 1159050 }, { "epoch": 6.989182474885735, "grad_norm": 4.428030490875244, "learning_rate": 1.778899648086085e-09, "loss": 4.0227, "step": 1159100 }, { "epoch": 6.989483966667069, "grad_norm": 4.9149370193481445, "learning_rate": 1.681177051404692e-09, "loss": 4.4127, "step": 1159150 }, { "epoch": 6.989785458448402, "grad_norm": 3.7468419075012207, "learning_rate": 1.5862149597312135e-09, "loss": 4.3774, "step": 1159200 }, { "epoch": 6.9900869502297365, "grad_norm": 2.3333184719085693, "learning_rate": 1.494013374797598e-09, "loss": 4.3672, "step": 1159250 }, { "epoch": 6.990388442011071, "grad_norm": 5.934176921844482, "learning_rate": 1.4045722983191398e-09, "loss": 4.026, "step": 1159300 }, { "epoch": 6.990689933792405, "grad_norm": 5.199026584625244, "learning_rate": 1.3178917319112136e-09, "loss": 4.1471, "step": 1159350 }, { "epoch": 6.990991425573739, "grad_norm": 4.796427249908447, "learning_rate": 1.2339716772058472e-09, "loss": 4.3093, "step": 1159400 }, { "epoch": 6.991292917355073, "grad_norm": 4.227102756500244, "learning_rate": 1.152812135718495e-09, "loss": 4.3971, "step": 1159450 }, { "epoch": 6.991594409136407, "grad_norm": 8.652019500732422, "learning_rate": 1.0744131089646112e-09, "loss": 4.1208, "step": 1159500 }, { "epoch": 6.991895900917741, "grad_norm": 2.6193325519561768, "learning_rate": 9.98774598376384e-10, "loss": 4.1547, "step": 1159550 }, { "epoch": 6.992197392699075, "grad_norm": 4.924525737762451, "learning_rate": 9.258966053360406e-10, "loss": 4.5735, "step": 1159600 }, { "epoch": 6.992498884480409, "grad_norm": 8.79789924621582, "learning_rate": 8.557791311925022e-10, "loss": 4.538, "step": 1159650 }, { "epoch": 6.9928003762617434, "grad_norm": 8.499534606933594, "learning_rate": 7.884221772447297e-10, "loss": 4.1695, "step": 1159700 }, { "epoch": 6.993101868043077, "grad_norm": 5.593395709991455, "learning_rate": 7.238257447250706e-10, "loss": 4.1962, "step": 1159750 }, { "epoch": 6.993403359824411, "grad_norm": 5.606229305267334, "learning_rate": 6.619898348325658e-10, "loss": 4.1306, "step": 1159800 }, { "epoch": 6.993704851605745, "grad_norm": 4.54372501373291, "learning_rate": 6.029144486829895e-10, "loss": 4.0402, "step": 1159850 }, { "epoch": 6.994006343387079, "grad_norm": 6.691610813140869, "learning_rate": 5.465995873754625e-10, "loss": 4.4296, "step": 1159900 }, { "epoch": 6.994307835168414, "grad_norm": 3.759533405303955, "learning_rate": 4.930452519424921e-10, "loss": 3.9047, "step": 1159950 }, { "epoch": 6.994609326949748, "grad_norm": 4.978980541229248, "learning_rate": 4.4225144338327914e-10, "loss": 4.3464, "step": 1160000 }, { "epoch": 6.994910818731081, "grad_norm": 3.159475326538086, "learning_rate": 3.9421816263041086e-10, "loss": 4.0948, "step": 1160050 }, { "epoch": 6.995212310512415, "grad_norm": 4.5121564865112305, "learning_rate": 3.489454105498612e-10, "loss": 4.4028, "step": 1160100 }, { "epoch": 6.9955138022937495, "grad_norm": 4.467411518096924, "learning_rate": 3.0643318799095095e-10, "loss": 4.4983, "step": 1160150 }, { "epoch": 6.995815294075084, "grad_norm": 4.467978000640869, "learning_rate": 2.666814957197338e-10, "loss": 3.8631, "step": 1160200 }, { "epoch": 6.996116785856418, "grad_norm": 5.578783988952637, "learning_rate": 2.2969033450226382e-10, "loss": 4.0839, "step": 1160250 }, { "epoch": 6.996418277637751, "grad_norm": 4.228992462158203, "learning_rate": 1.95459704971368e-10, "loss": 4.4463, "step": 1160300 }, { "epoch": 6.9967197694190855, "grad_norm": 7.178428649902344, "learning_rate": 1.6398960779318016e-10, "loss": 4.3632, "step": 1160350 }, { "epoch": 6.99702126120042, "grad_norm": 3.9393882751464844, "learning_rate": 1.352800435339141e-10, "loss": 4.0876, "step": 1160400 }, { "epoch": 6.997322752981754, "grad_norm": 3.5865554809570312, "learning_rate": 1.0933101272647682e-10, "loss": 4.4755, "step": 1160450 }, { "epoch": 6.997624244763088, "grad_norm": 6.693874359130859, "learning_rate": 8.614251585381537e-11, "loss": 3.9353, "step": 1160500 }, { "epoch": 6.997925736544422, "grad_norm": 10.336051940917969, "learning_rate": 6.571455331561004e-11, "loss": 3.8716, "step": 1160550 }, { "epoch": 6.998227228325756, "grad_norm": 5.562167644500732, "learning_rate": 4.8047125511541105e-11, "loss": 4.2787, "step": 1160600 }, { "epoch": 6.99852872010709, "grad_norm": 4.261462688446045, "learning_rate": 3.314023277467548e-11, "loss": 4.1402, "step": 1160650 }, { "epoch": 6.998830211888424, "grad_norm": 4.452030658721924, "learning_rate": 2.0993875354813338e-11, "loss": 4.3301, "step": 1160700 }, { "epoch": 6.999131703669758, "grad_norm": 3.8325891494750977, "learning_rate": 1.160805348510152e-11, "loss": 4.189, "step": 1160750 }, { "epoch": 6.9994331954510915, "grad_norm": 4.833970069885254, "learning_rate": 4.9827673487268235e-12, "loss": 4.09, "step": 1160800 }, { "epoch": 6.999734687232426, "grad_norm": 6.091032028198242, "learning_rate": 1.118017062262666e-12, "loss": 4.1536, "step": 1160850 } ], "logging_steps": 50, "max_steps": 1160894, "num_input_tokens_seen": 0, "num_train_epochs": 7, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.293557897033941e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }