{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.999804411104847, "eval_steps": 1000, "global_step": 108645, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002301045825327611, "grad_norm": 3.5780375003814697, "learning_rate": 4.7e-06, "loss": 13.6899, "step": 50 }, { "epoch": 0.004602091650655222, "grad_norm": 12.634012222290039, "learning_rate": 9.7e-06, "loss": 10.863, "step": 100 }, { "epoch": 0.006903137475982834, "grad_norm": 4.505175590515137, "learning_rate": 1.47e-05, "loss": 8.9306, "step": 150 }, { "epoch": 0.009204183301310445, "grad_norm": 4.666654109954834, "learning_rate": 1.9600000000000002e-05, "loss": 8.6804, "step": 200 }, { "epoch": 0.011505229126638056, "grad_norm": 12.136724472045898, "learning_rate": 2.46e-05, "loss": 8.6245, "step": 250 }, { "epoch": 0.013806274951965668, "grad_norm": 5.4161858558654785, "learning_rate": 2.96e-05, "loss": 8.5856, "step": 300 }, { "epoch": 0.01610732077729328, "grad_norm": 7.054779529571533, "learning_rate": 3.46e-05, "loss": 8.4581, "step": 350 }, { "epoch": 0.01840836660262089, "grad_norm": 4.1660475730896, "learning_rate": 3.960000000000001e-05, "loss": 8.387, "step": 400 }, { "epoch": 0.020709412427948503, "grad_norm": 7.97440767288208, "learning_rate": 4.46e-05, "loss": 8.2992, "step": 450 }, { "epoch": 0.023010458253276113, "grad_norm": 4.549796104431152, "learning_rate": 4.96e-05, "loss": 8.2301, "step": 500 }, { "epoch": 0.025311504078603726, "grad_norm": 4.553031921386719, "learning_rate": 5.4600000000000006e-05, "loss": 8.1272, "step": 550 }, { "epoch": 0.027612549903931336, "grad_norm": 5.876681327819824, "learning_rate": 5.96e-05, "loss": 8.0101, "step": 600 }, { "epoch": 0.02991359572925895, "grad_norm": 7.108986854553223, "learning_rate": 6.460000000000001e-05, "loss": 8.009, "step": 650 }, { "epoch": 0.03221464155458656, "grad_norm": 7.466887474060059, "learning_rate": 6.96e-05, "loss": 7.8225, "step": 700 }, { "epoch": 0.03451568737991417, "grad_norm": 5.25390625, "learning_rate": 7.46e-05, "loss": 7.8382, "step": 750 }, { "epoch": 0.03681673320524178, "grad_norm": 6.923309803009033, "learning_rate": 7.960000000000001e-05, "loss": 7.6769, "step": 800 }, { "epoch": 0.03911777903056939, "grad_norm": 7.7578816413879395, "learning_rate": 8.46e-05, "loss": 7.5925, "step": 850 }, { "epoch": 0.041418824855897006, "grad_norm": 7.556878566741943, "learning_rate": 8.960000000000001e-05, "loss": 7.5195, "step": 900 }, { "epoch": 0.04371987068122462, "grad_norm": 5.801334381103516, "learning_rate": 9.46e-05, "loss": 7.5253, "step": 950 }, { "epoch": 0.046020916506552226, "grad_norm": 7.2928948402404785, "learning_rate": 9.960000000000001e-05, "loss": 7.5399, "step": 1000 }, { "epoch": 0.046020916506552226, "eval_loss": 7.314383029937744, "eval_runtime": 32.585, "eval_samples_per_second": 11.785, "eval_steps_per_second": 5.892, "eval_tts_loss": 7.115667695799289, "step": 1000 }, { "epoch": 0.04832196233187984, "grad_norm": 5.685014247894287, "learning_rate": 9.999995494244432e-05, "loss": 7.4569, "step": 1050 }, { "epoch": 0.05062300815720745, "grad_norm": 5.959719657897949, "learning_rate": 9.999980375698306e-05, "loss": 7.4072, "step": 1100 }, { "epoch": 0.05292405398253506, "grad_norm": 6.67860221862793, "learning_rate": 9.999954610323825e-05, "loss": 7.4064, "step": 1150 }, { "epoch": 0.05522509980786267, "grad_norm": 4.586386203765869, "learning_rate": 9.999918198175851e-05, "loss": 7.3695, "step": 1200 }, { "epoch": 0.057526145633190286, "grad_norm": 3.8274621963500977, "learning_rate": 9.999871139331921e-05, "loss": 7.2462, "step": 1250 }, { "epoch": 0.0598271914585179, "grad_norm": 5.492649078369141, "learning_rate": 9.99981343389224e-05, "loss": 7.2843, "step": 1300 }, { "epoch": 0.062128237283845505, "grad_norm": 4.324922561645508, "learning_rate": 9.999745081979683e-05, "loss": 7.3252, "step": 1350 }, { "epoch": 0.06442928310917312, "grad_norm": 3.14571213722229, "learning_rate": 9.999666083739801e-05, "loss": 7.2732, "step": 1400 }, { "epoch": 0.06673032893450073, "grad_norm": 4.191070556640625, "learning_rate": 9.999576439340806e-05, "loss": 7.1295, "step": 1450 }, { "epoch": 0.06903137475982835, "grad_norm": 3.528020143508911, "learning_rate": 9.999476148973588e-05, "loss": 7.1829, "step": 1500 }, { "epoch": 0.07133242058515596, "grad_norm": 4.1083664894104, "learning_rate": 9.999365212851702e-05, "loss": 7.2347, "step": 1550 }, { "epoch": 0.07363346641048356, "grad_norm": 3.2220022678375244, "learning_rate": 9.99924363121137e-05, "loss": 7.1745, "step": 1600 }, { "epoch": 0.07593451223581117, "grad_norm": 4.157455921173096, "learning_rate": 9.999111404311488e-05, "loss": 7.128, "step": 1650 }, { "epoch": 0.07823555806113879, "grad_norm": 2.825235605239868, "learning_rate": 9.998968532433615e-05, "loss": 7.1691, "step": 1700 }, { "epoch": 0.0805366038864664, "grad_norm": 3.339789867401123, "learning_rate": 9.998815015881976e-05, "loss": 7.0882, "step": 1750 }, { "epoch": 0.08283764971179401, "grad_norm": 2.8071019649505615, "learning_rate": 9.998650854983472e-05, "loss": 7.0621, "step": 1800 }, { "epoch": 0.08513869553712163, "grad_norm": 3.1003334522247314, "learning_rate": 9.998476050087657e-05, "loss": 7.0141, "step": 1850 }, { "epoch": 0.08743974136244924, "grad_norm": 2.8613014221191406, "learning_rate": 9.998290601566759e-05, "loss": 7.0619, "step": 1900 }, { "epoch": 0.08974078718777684, "grad_norm": 2.567081928253174, "learning_rate": 9.998094509815665e-05, "loss": 7.0753, "step": 1950 }, { "epoch": 0.09204183301310445, "grad_norm": 2.5587222576141357, "learning_rate": 9.99788777525193e-05, "loss": 7.0433, "step": 2000 }, { "epoch": 0.09204183301310445, "eval_loss": 6.901029109954834, "eval_runtime": 33.3966, "eval_samples_per_second": 11.498, "eval_steps_per_second": 5.749, "eval_tts_loss": 7.299251959316763, "step": 2000 }, { "epoch": 0.09434287883843206, "grad_norm": 2.903456687927246, "learning_rate": 9.997670398315768e-05, "loss": 7.0923, "step": 2050 }, { "epoch": 0.09664392466375968, "grad_norm": 2.615713119506836, "learning_rate": 9.997442379470056e-05, "loss": 7.0888, "step": 2100 }, { "epoch": 0.09894497048908729, "grad_norm": 2.7937371730804443, "learning_rate": 9.997203719200331e-05, "loss": 6.9845, "step": 2150 }, { "epoch": 0.1012460163144149, "grad_norm": 2.678640365600586, "learning_rate": 9.99695441801479e-05, "loss": 7.0278, "step": 2200 }, { "epoch": 0.10354706213974252, "grad_norm": 2.9163057804107666, "learning_rate": 9.996694476444288e-05, "loss": 6.9738, "step": 2250 }, { "epoch": 0.10584810796507012, "grad_norm": 2.41644287109375, "learning_rate": 9.99642389504234e-05, "loss": 6.9378, "step": 2300 }, { "epoch": 0.10814915379039773, "grad_norm": 2.7268104553222656, "learning_rate": 9.996142674385114e-05, "loss": 7.0378, "step": 2350 }, { "epoch": 0.11045019961572534, "grad_norm": 2.242163896560669, "learning_rate": 9.995850815071433e-05, "loss": 6.8807, "step": 2400 }, { "epoch": 0.11275124544105296, "grad_norm": 2.39416241645813, "learning_rate": 9.995548317722775e-05, "loss": 6.9694, "step": 2450 }, { "epoch": 0.11505229126638057, "grad_norm": 2.4873769283294678, "learning_rate": 9.99523518298327e-05, "loss": 6.7956, "step": 2500 }, { "epoch": 0.11735333709170818, "grad_norm": 2.4013776779174805, "learning_rate": 9.994911411519699e-05, "loss": 6.8655, "step": 2550 }, { "epoch": 0.1196543829170358, "grad_norm": 3.897566080093384, "learning_rate": 9.994577004021493e-05, "loss": 6.822, "step": 2600 }, { "epoch": 0.1219554287423634, "grad_norm": 2.2062366008758545, "learning_rate": 9.994231961200731e-05, "loss": 6.7814, "step": 2650 }, { "epoch": 0.12425647456769101, "grad_norm": 2.6973462104797363, "learning_rate": 9.993876283792138e-05, "loss": 6.9139, "step": 2700 }, { "epoch": 0.12655752039301862, "grad_norm": 2.6067421436309814, "learning_rate": 9.993509972553084e-05, "loss": 6.8624, "step": 2750 }, { "epoch": 0.12885856621834624, "grad_norm": 2.6363160610198975, "learning_rate": 9.993133028263581e-05, "loss": 6.8297, "step": 2800 }, { "epoch": 0.13115961204367385, "grad_norm": 2.585055112838745, "learning_rate": 9.992745451726286e-05, "loss": 6.9063, "step": 2850 }, { "epoch": 0.13346065786900146, "grad_norm": 2.4903454780578613, "learning_rate": 9.992347243766494e-05, "loss": 6.8681, "step": 2900 }, { "epoch": 0.13576170369432908, "grad_norm": 2.526623249053955, "learning_rate": 9.99193840523214e-05, "loss": 6.8679, "step": 2950 }, { "epoch": 0.1380627495196567, "grad_norm": 2.4259722232818604, "learning_rate": 9.991518936993794e-05, "loss": 6.7638, "step": 3000 }, { "epoch": 0.1380627495196567, "eval_loss": 6.7133917808532715, "eval_runtime": 33.4323, "eval_samples_per_second": 11.486, "eval_steps_per_second": 5.743, "eval_tts_loss": 7.4207137375880805, "step": 3000 }, { "epoch": 0.1403637953449843, "grad_norm": 2.2677650451660156, "learning_rate": 9.991088839944658e-05, "loss": 6.8054, "step": 3050 }, { "epoch": 0.14266484117031192, "grad_norm": 1.8182429075241089, "learning_rate": 9.990648115000572e-05, "loss": 6.7162, "step": 3100 }, { "epoch": 0.14496588699563953, "grad_norm": 2.308943748474121, "learning_rate": 9.990196763100004e-05, "loss": 6.8139, "step": 3150 }, { "epoch": 0.14726693282096712, "grad_norm": 2.179896593093872, "learning_rate": 9.98973478520405e-05, "loss": 6.6905, "step": 3200 }, { "epoch": 0.14956797864629473, "grad_norm": 2.6917459964752197, "learning_rate": 9.989262182296435e-05, "loss": 6.7617, "step": 3250 }, { "epoch": 0.15186902447162234, "grad_norm": 2.362337589263916, "learning_rate": 9.988778955383506e-05, "loss": 6.8229, "step": 3300 }, { "epoch": 0.15417007029694996, "grad_norm": 2.33597731590271, "learning_rate": 9.988285105494236e-05, "loss": 6.7483, "step": 3350 }, { "epoch": 0.15647111612227757, "grad_norm": 2.041456460952759, "learning_rate": 9.987780633680213e-05, "loss": 6.6595, "step": 3400 }, { "epoch": 0.15877216194760518, "grad_norm": 2.1077539920806885, "learning_rate": 9.98726554101565e-05, "loss": 6.7904, "step": 3450 }, { "epoch": 0.1610732077729328, "grad_norm": 2.350275993347168, "learning_rate": 9.986739828597367e-05, "loss": 6.7545, "step": 3500 }, { "epoch": 0.1633742535982604, "grad_norm": 2.2597575187683105, "learning_rate": 9.986203497544808e-05, "loss": 6.7821, "step": 3550 }, { "epoch": 0.16567529942358802, "grad_norm": 2.205111026763916, "learning_rate": 9.985656549000018e-05, "loss": 6.7361, "step": 3600 }, { "epoch": 0.16797634524891564, "grad_norm": 2.0112202167510986, "learning_rate": 9.985098984127657e-05, "loss": 6.6824, "step": 3650 }, { "epoch": 0.17027739107424325, "grad_norm": 2.225270986557007, "learning_rate": 9.984530804114987e-05, "loss": 6.7456, "step": 3700 }, { "epoch": 0.17257843689957086, "grad_norm": 2.2781286239624023, "learning_rate": 9.98395201017188e-05, "loss": 6.7912, "step": 3750 }, { "epoch": 0.17487948272489848, "grad_norm": 2.5008931159973145, "learning_rate": 9.983362603530802e-05, "loss": 6.7475, "step": 3800 }, { "epoch": 0.1771805285502261, "grad_norm": 2.242917776107788, "learning_rate": 9.98276258544682e-05, "loss": 6.6998, "step": 3850 }, { "epoch": 0.17948157437555368, "grad_norm": 1.6901216506958008, "learning_rate": 9.982151957197598e-05, "loss": 6.6986, "step": 3900 }, { "epoch": 0.1817826202008813, "grad_norm": 2.2120909690856934, "learning_rate": 9.981530720083391e-05, "loss": 6.6784, "step": 3950 }, { "epoch": 0.1840836660262089, "grad_norm": 2.279149055480957, "learning_rate": 9.980898875427045e-05, "loss": 6.6557, "step": 4000 }, { "epoch": 0.1840836660262089, "eval_loss": 6.581401824951172, "eval_runtime": 32.9434, "eval_samples_per_second": 11.656, "eval_steps_per_second": 5.828, "eval_tts_loss": 7.501853841915528, "step": 4000 }, { "epoch": 0.18638471185153652, "grad_norm": 1.847670078277588, "learning_rate": 9.980256424573995e-05, "loss": 6.6079, "step": 4050 }, { "epoch": 0.18868575767686413, "grad_norm": 1.8807518482208252, "learning_rate": 9.979603368892259e-05, "loss": 6.705, "step": 4100 }, { "epoch": 0.19098680350219174, "grad_norm": 2.1693525314331055, "learning_rate": 9.978939709772435e-05, "loss": 6.6899, "step": 4150 }, { "epoch": 0.19328784932751936, "grad_norm": 1.934512734413147, "learning_rate": 9.978265448627702e-05, "loss": 6.6307, "step": 4200 }, { "epoch": 0.19558889515284697, "grad_norm": 2.0667643547058105, "learning_rate": 9.977580586893816e-05, "loss": 6.6746, "step": 4250 }, { "epoch": 0.19788994097817458, "grad_norm": 2.2924129962921143, "learning_rate": 9.976885126029103e-05, "loss": 6.6497, "step": 4300 }, { "epoch": 0.2001909868035022, "grad_norm": 2.1842291355133057, "learning_rate": 9.976179067514459e-05, "loss": 6.5531, "step": 4350 }, { "epoch": 0.2024920326288298, "grad_norm": 2.3213446140289307, "learning_rate": 9.975462412853345e-05, "loss": 6.6737, "step": 4400 }, { "epoch": 0.20479307845415742, "grad_norm": 1.937312126159668, "learning_rate": 9.974735163571789e-05, "loss": 6.6511, "step": 4450 }, { "epoch": 0.20709412427948504, "grad_norm": 1.7467703819274902, "learning_rate": 9.973997321218376e-05, "loss": 6.5941, "step": 4500 }, { "epoch": 0.20939517010481265, "grad_norm": 1.9643598794937134, "learning_rate": 9.973248887364251e-05, "loss": 6.561, "step": 4550 }, { "epoch": 0.21169621593014024, "grad_norm": 2.028480052947998, "learning_rate": 9.972489863603102e-05, "loss": 6.5878, "step": 4600 }, { "epoch": 0.21399726175546785, "grad_norm": 1.9181138277053833, "learning_rate": 9.971720251551184e-05, "loss": 6.63, "step": 4650 }, { "epoch": 0.21629830758079546, "grad_norm": 1.8953630924224854, "learning_rate": 9.970940052847283e-05, "loss": 6.6879, "step": 4700 }, { "epoch": 0.21859935340612308, "grad_norm": 2.008338451385498, "learning_rate": 9.970149269152736e-05, "loss": 6.5496, "step": 4750 }, { "epoch": 0.2209003992314507, "grad_norm": 2.563183546066284, "learning_rate": 9.969347902151416e-05, "loss": 6.5905, "step": 4800 }, { "epoch": 0.2232014450567783, "grad_norm": 2.335991382598877, "learning_rate": 9.968535953549734e-05, "loss": 6.6685, "step": 4850 }, { "epoch": 0.22550249088210592, "grad_norm": 1.9713267087936401, "learning_rate": 9.967713425076631e-05, "loss": 6.6486, "step": 4900 }, { "epoch": 0.22780353670743353, "grad_norm": 1.923415184020996, "learning_rate": 9.966880318483577e-05, "loss": 6.6034, "step": 4950 }, { "epoch": 0.23010458253276114, "grad_norm": 2.1620161533355713, "learning_rate": 9.966036635544573e-05, "loss": 6.5386, "step": 5000 }, { "epoch": 0.23010458253276114, "eval_loss": 6.485872745513916, "eval_runtime": 34.773, "eval_samples_per_second": 11.043, "eval_steps_per_second": 5.522, "eval_tts_loss": 7.581498966822206, "step": 5000 }, { "epoch": 0.23240562835808876, "grad_norm": 2.058232307434082, "learning_rate": 9.96518237805613e-05, "loss": 6.5541, "step": 5050 }, { "epoch": 0.23470667418341637, "grad_norm": 2.401785373687744, "learning_rate": 9.964317547837283e-05, "loss": 6.5717, "step": 5100 }, { "epoch": 0.23700772000874398, "grad_norm": 1.8926520347595215, "learning_rate": 9.963442146729583e-05, "loss": 6.5713, "step": 5150 }, { "epoch": 0.2393087658340716, "grad_norm": 2.0552031993865967, "learning_rate": 9.962556176597079e-05, "loss": 6.5887, "step": 5200 }, { "epoch": 0.2416098116593992, "grad_norm": 2.0118703842163086, "learning_rate": 9.961659639326338e-05, "loss": 6.5794, "step": 5250 }, { "epoch": 0.2439108574847268, "grad_norm": 1.9087331295013428, "learning_rate": 9.960752536826422e-05, "loss": 6.5892, "step": 5300 }, { "epoch": 0.2462119033100544, "grad_norm": 1.9817070960998535, "learning_rate": 9.95983487102889e-05, "loss": 6.6203, "step": 5350 }, { "epoch": 0.24851294913538202, "grad_norm": 1.9609904289245605, "learning_rate": 9.9589066438878e-05, "loss": 6.6003, "step": 5400 }, { "epoch": 0.25081399496070966, "grad_norm": 1.7898999452590942, "learning_rate": 9.957967857379689e-05, "loss": 6.5957, "step": 5450 }, { "epoch": 0.25311504078603725, "grad_norm": 2.2270612716674805, "learning_rate": 9.957018513503589e-05, "loss": 6.548, "step": 5500 }, { "epoch": 0.2554160866113649, "grad_norm": 2.224034547805786, "learning_rate": 9.956058614281006e-05, "loss": 6.5198, "step": 5550 }, { "epoch": 0.2577171324366925, "grad_norm": 1.9815775156021118, "learning_rate": 9.955088161755925e-05, "loss": 6.5712, "step": 5600 }, { "epoch": 0.26001817826202006, "grad_norm": 2.1270036697387695, "learning_rate": 9.954107157994802e-05, "loss": 6.5586, "step": 5650 }, { "epoch": 0.2623192240873477, "grad_norm": 1.7809762954711914, "learning_rate": 9.953115605086564e-05, "loss": 6.5644, "step": 5700 }, { "epoch": 0.2646202699126753, "grad_norm": 1.6975411176681519, "learning_rate": 9.952113505142593e-05, "loss": 6.5975, "step": 5750 }, { "epoch": 0.26692131573800293, "grad_norm": 1.9254316091537476, "learning_rate": 9.951100860296738e-05, "loss": 6.5063, "step": 5800 }, { "epoch": 0.2692223615633305, "grad_norm": 1.7933043241500854, "learning_rate": 9.950077672705298e-05, "loss": 6.4768, "step": 5850 }, { "epoch": 0.27152340738865816, "grad_norm": 2.008695363998413, "learning_rate": 9.94904394454702e-05, "loss": 6.5563, "step": 5900 }, { "epoch": 0.27382445321398574, "grad_norm": 2.1302218437194824, "learning_rate": 9.947999678023103e-05, "loss": 6.5424, "step": 5950 }, { "epoch": 0.2761254990393134, "grad_norm": 1.78187894821167, "learning_rate": 9.946944875357174e-05, "loss": 6.61, "step": 6000 }, { "epoch": 0.2761254990393134, "eval_loss": 6.425212860107422, "eval_runtime": 35.036, "eval_samples_per_second": 10.96, "eval_steps_per_second": 5.48, "eval_tts_loss": 7.612353266970973, "step": 6000 }, { "epoch": 0.27842654486464097, "grad_norm": 1.921924352645874, "learning_rate": 9.945879538795309e-05, "loss": 6.5369, "step": 6050 }, { "epoch": 0.2807275906899686, "grad_norm": 1.6859031915664673, "learning_rate": 9.944803670606004e-05, "loss": 6.5272, "step": 6100 }, { "epoch": 0.2830286365152962, "grad_norm": 1.9181588888168335, "learning_rate": 9.943717273080186e-05, "loss": 6.527, "step": 6150 }, { "epoch": 0.28532968234062384, "grad_norm": 1.921380877494812, "learning_rate": 9.942620348531201e-05, "loss": 6.5185, "step": 6200 }, { "epoch": 0.2876307281659514, "grad_norm": 2.4337542057037354, "learning_rate": 9.941512899294813e-05, "loss": 6.5091, "step": 6250 }, { "epoch": 0.28993177399127906, "grad_norm": 1.9452122449874878, "learning_rate": 9.940394927729194e-05, "loss": 6.5382, "step": 6300 }, { "epoch": 0.29223281981660665, "grad_norm": 1.84445321559906, "learning_rate": 9.939266436214925e-05, "loss": 6.5412, "step": 6350 }, { "epoch": 0.29453386564193423, "grad_norm": 2.0815675258636475, "learning_rate": 9.938127427154986e-05, "loss": 6.5187, "step": 6400 }, { "epoch": 0.2968349114672619, "grad_norm": 1.921200156211853, "learning_rate": 9.936977902974753e-05, "loss": 6.5182, "step": 6450 }, { "epoch": 0.29913595729258946, "grad_norm": 2.276503086090088, "learning_rate": 9.935841169867299e-05, "loss": 6.4457, "step": 6500 }, { "epoch": 0.3014370031179171, "grad_norm": 2.267094612121582, "learning_rate": 9.934670832991861e-05, "loss": 6.5038, "step": 6550 }, { "epoch": 0.3037380489432447, "grad_norm": 1.982542872428894, "learning_rate": 9.93348998835651e-05, "loss": 6.572, "step": 6600 }, { "epoch": 0.30603909476857233, "grad_norm": 2.0786359310150146, "learning_rate": 9.932298638475707e-05, "loss": 6.54, "step": 6650 }, { "epoch": 0.3083401405938999, "grad_norm": 2.178440809249878, "learning_rate": 9.931096785886282e-05, "loss": 6.4728, "step": 6700 }, { "epoch": 0.31064118641922756, "grad_norm": 1.7472290992736816, "learning_rate": 9.92988443314743e-05, "loss": 6.4496, "step": 6750 }, { "epoch": 0.31294223224455514, "grad_norm": 2.0170915126800537, "learning_rate": 9.9286615828407e-05, "loss": 6.5349, "step": 6800 }, { "epoch": 0.3152432780698828, "grad_norm": 1.9643365144729614, "learning_rate": 9.927428237569998e-05, "loss": 6.472, "step": 6850 }, { "epoch": 0.31754432389521037, "grad_norm": 1.955383539199829, "learning_rate": 9.92618439996158e-05, "loss": 6.5057, "step": 6900 }, { "epoch": 0.319845369720538, "grad_norm": 1.6880296468734741, "learning_rate": 9.92493007266404e-05, "loss": 6.4687, "step": 6950 }, { "epoch": 0.3221464155458656, "grad_norm": 1.9719414710998535, "learning_rate": 9.923665258348311e-05, "loss": 6.4366, "step": 7000 }, { "epoch": 0.3221464155458656, "eval_loss": 6.379239559173584, "eval_runtime": 33.7763, "eval_samples_per_second": 11.369, "eval_steps_per_second": 5.684, "eval_tts_loss": 7.613140820339867, "step": 7000 }, { "epoch": 0.3244474613711932, "grad_norm": 2.014941930770874, "learning_rate": 9.922389959707654e-05, "loss": 6.4835, "step": 7050 }, { "epoch": 0.3267485071965208, "grad_norm": 1.9350296258926392, "learning_rate": 9.921104179457658e-05, "loss": 6.4587, "step": 7100 }, { "epoch": 0.3290495530218484, "grad_norm": 1.7552827596664429, "learning_rate": 9.91980792033623e-05, "loss": 6.4796, "step": 7150 }, { "epoch": 0.33135059884717605, "grad_norm": 1.9260075092315674, "learning_rate": 9.91850118510359e-05, "loss": 6.4548, "step": 7200 }, { "epoch": 0.33365164467250363, "grad_norm": 1.632943868637085, "learning_rate": 9.917183976542268e-05, "loss": 6.4872, "step": 7250 }, { "epoch": 0.3359526904978313, "grad_norm": 1.9595330953598022, "learning_rate": 9.915856297457091e-05, "loss": 6.4717, "step": 7300 }, { "epoch": 0.33825373632315886, "grad_norm": 1.787189245223999, "learning_rate": 9.914518150675185e-05, "loss": 6.4497, "step": 7350 }, { "epoch": 0.3405547821484865, "grad_norm": 1.894849419593811, "learning_rate": 9.913169539045966e-05, "loss": 6.4711, "step": 7400 }, { "epoch": 0.3428558279738141, "grad_norm": 1.8900083303451538, "learning_rate": 9.911810465441131e-05, "loss": 6.4507, "step": 7450 }, { "epoch": 0.34515687379914173, "grad_norm": 2.1415016651153564, "learning_rate": 9.910440932754658e-05, "loss": 6.4268, "step": 7500 }, { "epoch": 0.3474579196244693, "grad_norm": 2.0673065185546875, "learning_rate": 9.909060943902793e-05, "loss": 6.3687, "step": 7550 }, { "epoch": 0.34975896544979695, "grad_norm": 1.9016964435577393, "learning_rate": 9.90767050182405e-05, "loss": 6.4239, "step": 7600 }, { "epoch": 0.35206001127512454, "grad_norm": 1.8827308416366577, "learning_rate": 9.906269609479198e-05, "loss": 6.4469, "step": 7650 }, { "epoch": 0.3543610571004522, "grad_norm": 1.8428524732589722, "learning_rate": 9.904858269851261e-05, "loss": 6.5031, "step": 7700 }, { "epoch": 0.35666210292577977, "grad_norm": 2.421943426132202, "learning_rate": 9.903436485945512e-05, "loss": 6.4074, "step": 7750 }, { "epoch": 0.35896314875110735, "grad_norm": 1.8993301391601562, "learning_rate": 9.902004260789457e-05, "loss": 6.4325, "step": 7800 }, { "epoch": 0.361264194576435, "grad_norm": 1.8504090309143066, "learning_rate": 9.90056159743284e-05, "loss": 6.4851, "step": 7850 }, { "epoch": 0.3635652404017626, "grad_norm": 2.0040860176086426, "learning_rate": 9.899108498947634e-05, "loss": 6.5079, "step": 7900 }, { "epoch": 0.3658662862270902, "grad_norm": 1.930400013923645, "learning_rate": 9.897644968428025e-05, "loss": 6.4366, "step": 7950 }, { "epoch": 0.3681673320524178, "grad_norm": 2.233633279800415, "learning_rate": 9.896171008990418e-05, "loss": 6.3755, "step": 8000 }, { "epoch": 0.3681673320524178, "eval_loss": 6.341970920562744, "eval_runtime": 33.2129, "eval_samples_per_second": 11.562, "eval_steps_per_second": 5.781, "eval_tts_loss": 7.688191944505792, "step": 8000 }, { "epoch": 0.37046837787774545, "grad_norm": 2.337127447128296, "learning_rate": 9.894686623773426e-05, "loss": 6.4258, "step": 8050 }, { "epoch": 0.37276942370307303, "grad_norm": 1.9181020259857178, "learning_rate": 9.893191815937857e-05, "loss": 6.4511, "step": 8100 }, { "epoch": 0.3750704695284007, "grad_norm": 1.8250304460525513, "learning_rate": 9.89168658866672e-05, "loss": 6.471, "step": 8150 }, { "epoch": 0.37737151535372826, "grad_norm": 1.7581539154052734, "learning_rate": 9.890170945165203e-05, "loss": 6.3565, "step": 8200 }, { "epoch": 0.3796725611790559, "grad_norm": 1.6649830341339111, "learning_rate": 9.88864488866068e-05, "loss": 6.4983, "step": 8250 }, { "epoch": 0.3819736070043835, "grad_norm": 1.8759455680847168, "learning_rate": 9.887108422402696e-05, "loss": 6.384, "step": 8300 }, { "epoch": 0.3842746528297111, "grad_norm": 2.1728053092956543, "learning_rate": 9.88556154966296e-05, "loss": 6.4939, "step": 8350 }, { "epoch": 0.3865756986550387, "grad_norm": 2.9008238315582275, "learning_rate": 9.884004273735347e-05, "loss": 6.375, "step": 8400 }, { "epoch": 0.38887674448036635, "grad_norm": 1.6754823923110962, "learning_rate": 9.882436597935877e-05, "loss": 6.3691, "step": 8450 }, { "epoch": 0.39117779030569394, "grad_norm": 1.9855177402496338, "learning_rate": 9.880858525602719e-05, "loss": 6.3651, "step": 8500 }, { "epoch": 0.3934788361310215, "grad_norm": 1.8004143238067627, "learning_rate": 9.879270060096178e-05, "loss": 6.446, "step": 8550 }, { "epoch": 0.39577988195634917, "grad_norm": 2.030467987060547, "learning_rate": 9.877671204798693e-05, "loss": 6.4362, "step": 8600 }, { "epoch": 0.39808092778167675, "grad_norm": 1.5885698795318604, "learning_rate": 9.876061963114823e-05, "loss": 6.3598, "step": 8650 }, { "epoch": 0.4003819736070044, "grad_norm": 1.9746173620224, "learning_rate": 9.874442338471246e-05, "loss": 6.4072, "step": 8700 }, { "epoch": 0.402683019432332, "grad_norm": 1.8697595596313477, "learning_rate": 9.872845036096635e-05, "loss": 6.3998, "step": 8750 }, { "epoch": 0.4049840652576596, "grad_norm": 1.9580832719802856, "learning_rate": 9.871204863388751e-05, "loss": 6.4055, "step": 8800 }, { "epoch": 0.4072851110829872, "grad_norm": 2.037536859512329, "learning_rate": 9.869554318063743e-05, "loss": 6.3919, "step": 8850 }, { "epoch": 0.40958615690831485, "grad_norm": 2.1049587726593018, "learning_rate": 9.867893403636237e-05, "loss": 6.437, "step": 8900 }, { "epoch": 0.41188720273364243, "grad_norm": 1.8628839254379272, "learning_rate": 9.866222123642946e-05, "loss": 6.4381, "step": 8950 }, { "epoch": 0.4141882485589701, "grad_norm": 1.5722795724868774, "learning_rate": 9.864540481642644e-05, "loss": 6.3435, "step": 9000 }, { "epoch": 0.4141882485589701, "eval_loss": 6.309738636016846, "eval_runtime": 33.7952, "eval_samples_per_second": 11.363, "eval_steps_per_second": 5.681, "eval_tts_loss": 7.738714543004772, "step": 9000 }, { "epoch": 0.41648929438429766, "grad_norm": 1.810716152191162, "learning_rate": 9.86284848121618e-05, "loss": 6.4626, "step": 9050 }, { "epoch": 0.4187903402096253, "grad_norm": 1.6294385194778442, "learning_rate": 9.861146125966455e-05, "loss": 6.3863, "step": 9100 }, { "epoch": 0.4210913860349529, "grad_norm": 1.7128016948699951, "learning_rate": 9.85943341951842e-05, "loss": 6.4224, "step": 9150 }, { "epoch": 0.42339243186028047, "grad_norm": 1.9421542882919312, "learning_rate": 9.857710365519066e-05, "loss": 6.4133, "step": 9200 }, { "epoch": 0.4256934776856081, "grad_norm": 1.792728066444397, "learning_rate": 9.855976967637422e-05, "loss": 6.3259, "step": 9250 }, { "epoch": 0.4279945235109357, "grad_norm": 1.9310187101364136, "learning_rate": 9.85423322956454e-05, "loss": 6.3347, "step": 9300 }, { "epoch": 0.43029556933626334, "grad_norm": 2.0094540119171143, "learning_rate": 9.85247915501349e-05, "loss": 6.3609, "step": 9350 }, { "epoch": 0.4325966151615909, "grad_norm": 1.99038827419281, "learning_rate": 9.85071474771935e-05, "loss": 6.3763, "step": 9400 }, { "epoch": 0.43489766098691857, "grad_norm": 1.6113660335540771, "learning_rate": 9.848940011439207e-05, "loss": 6.4504, "step": 9450 }, { "epoch": 0.43719870681224615, "grad_norm": 1.7859371900558472, "learning_rate": 9.847154949952138e-05, "loss": 6.4023, "step": 9500 }, { "epoch": 0.4394997526375738, "grad_norm": 2.0613961219787598, "learning_rate": 9.845359567059205e-05, "loss": 6.4021, "step": 9550 }, { "epoch": 0.4418007984629014, "grad_norm": 1.7589733600616455, "learning_rate": 9.84355386658345e-05, "loss": 6.3691, "step": 9600 }, { "epoch": 0.444101844288229, "grad_norm": 1.9880051612854004, "learning_rate": 9.841737852369884e-05, "loss": 6.3675, "step": 9650 }, { "epoch": 0.4464028901135566, "grad_norm": 2.1406807899475098, "learning_rate": 9.839911528285484e-05, "loss": 6.4018, "step": 9700 }, { "epoch": 0.44870393593888425, "grad_norm": 1.5996681451797485, "learning_rate": 9.838074898219171e-05, "loss": 6.3984, "step": 9750 }, { "epoch": 0.45100498176421183, "grad_norm": 2.1226611137390137, "learning_rate": 9.836227966081823e-05, "loss": 6.407, "step": 9800 }, { "epoch": 0.4533060275895395, "grad_norm": 1.7142243385314941, "learning_rate": 9.834370735806248e-05, "loss": 6.3776, "step": 9850 }, { "epoch": 0.45560707341486706, "grad_norm": 1.8948795795440674, "learning_rate": 9.832503211347183e-05, "loss": 6.2953, "step": 9900 }, { "epoch": 0.45790811924019464, "grad_norm": 1.875388264656067, "learning_rate": 9.830625396681286e-05, "loss": 6.2976, "step": 9950 }, { "epoch": 0.4602091650655223, "grad_norm": 1.710487723350525, "learning_rate": 9.828737295807126e-05, "loss": 6.3461, "step": 10000 }, { "epoch": 0.4602091650655223, "eval_loss": 6.289060115814209, "eval_runtime": 34.3884, "eval_samples_per_second": 11.167, "eval_steps_per_second": 5.583, "eval_tts_loss": 7.814920344420079, "step": 10000 }, { "epoch": 0.46251021089084987, "grad_norm": 1.7555044889450073, "learning_rate": 9.826838912745178e-05, "loss": 6.4022, "step": 10050 }, { "epoch": 0.4648112567161775, "grad_norm": 1.9916635751724243, "learning_rate": 9.824930251537808e-05, "loss": 6.4684, "step": 10100 }, { "epoch": 0.4671123025415051, "grad_norm": 1.9527746438980103, "learning_rate": 9.823011316249271e-05, "loss": 6.3466, "step": 10150 }, { "epoch": 0.46941334836683274, "grad_norm": 1.668678879737854, "learning_rate": 9.821082110965696e-05, "loss": 6.3642, "step": 10200 }, { "epoch": 0.4717143941921603, "grad_norm": 1.691568374633789, "learning_rate": 9.819142639795085e-05, "loss": 6.3586, "step": 10250 }, { "epoch": 0.47401544001748797, "grad_norm": 1.8639135360717773, "learning_rate": 9.817192906867295e-05, "loss": 6.314, "step": 10300 }, { "epoch": 0.47631648584281555, "grad_norm": 1.7101272344589233, "learning_rate": 9.815232916334039e-05, "loss": 6.3314, "step": 10350 }, { "epoch": 0.4786175316681432, "grad_norm": 1.6885664463043213, "learning_rate": 9.813262672368868e-05, "loss": 6.4618, "step": 10400 }, { "epoch": 0.4809185774934708, "grad_norm": 1.7570464611053467, "learning_rate": 9.81128217916717e-05, "loss": 6.3258, "step": 10450 }, { "epoch": 0.4832196233187984, "grad_norm": 1.5871278047561646, "learning_rate": 9.809291440946154e-05, "loss": 6.3315, "step": 10500 }, { "epoch": 0.485520669144126, "grad_norm": 1.6280348300933838, "learning_rate": 9.807290461944849e-05, "loss": 6.3455, "step": 10550 }, { "epoch": 0.4878217149694536, "grad_norm": 1.7376903295516968, "learning_rate": 9.805279246424085e-05, "loss": 6.3455, "step": 10600 }, { "epoch": 0.49012276079478123, "grad_norm": 1.9705889225006104, "learning_rate": 9.803257798666493e-05, "loss": 6.2758, "step": 10650 }, { "epoch": 0.4924238066201088, "grad_norm": 1.764738917350769, "learning_rate": 9.801226122976492e-05, "loss": 6.2885, "step": 10700 }, { "epoch": 0.49472485244543646, "grad_norm": 1.740444302558899, "learning_rate": 9.799184223680278e-05, "loss": 6.3493, "step": 10750 }, { "epoch": 0.49702589827076404, "grad_norm": 1.464941143989563, "learning_rate": 9.79713210512582e-05, "loss": 6.3239, "step": 10800 }, { "epoch": 0.4993269440960917, "grad_norm": 1.7715671062469482, "learning_rate": 9.795069771682845e-05, "loss": 6.3002, "step": 10850 }, { "epoch": 0.5016279899214193, "grad_norm": 1.8907268047332764, "learning_rate": 9.792997227742833e-05, "loss": 6.2998, "step": 10900 }, { "epoch": 0.5039290357467469, "grad_norm": 1.9487262964248657, "learning_rate": 9.790914477719005e-05, "loss": 6.358, "step": 10950 }, { "epoch": 0.5062300815720745, "grad_norm": 1.8834967613220215, "learning_rate": 9.788821526046312e-05, "loss": 6.3576, "step": 11000 }, { "epoch": 0.5062300815720745, "eval_loss": 6.265079498291016, "eval_runtime": 34.881, "eval_samples_per_second": 11.009, "eval_steps_per_second": 5.504, "eval_tts_loss": 7.740538769862218, "step": 11000 }, { "epoch": 0.5085311273974021, "grad_norm": 1.8796509504318237, "learning_rate": 9.786718377181437e-05, "loss": 6.3623, "step": 11050 }, { "epoch": 0.5108321732227298, "grad_norm": 1.6013100147247314, "learning_rate": 9.784605035602764e-05, "loss": 6.3532, "step": 11100 }, { "epoch": 0.5131332190480573, "grad_norm": 1.9460678100585938, "learning_rate": 9.782481505810393e-05, "loss": 6.35, "step": 11150 }, { "epoch": 0.515434264873385, "grad_norm": 1.8577035665512085, "learning_rate": 9.780347792326108e-05, "loss": 6.4018, "step": 11200 }, { "epoch": 0.5177353106987126, "grad_norm": 2.1580724716186523, "learning_rate": 9.778203899693388e-05, "loss": 6.3821, "step": 11250 }, { "epoch": 0.5200363565240401, "grad_norm": 1.9228373765945435, "learning_rate": 9.77604983247738e-05, "loss": 6.3247, "step": 11300 }, { "epoch": 0.5223374023493678, "grad_norm": 1.6796767711639404, "learning_rate": 9.7738855952649e-05, "loss": 6.3384, "step": 11350 }, { "epoch": 0.5246384481746954, "grad_norm": 2.2095754146575928, "learning_rate": 9.771711192664417e-05, "loss": 6.3121, "step": 11400 }, { "epoch": 0.526939494000023, "grad_norm": 1.9185891151428223, "learning_rate": 9.769526629306046e-05, "loss": 6.2757, "step": 11450 }, { "epoch": 0.5292405398253506, "grad_norm": 1.7003620862960815, "learning_rate": 9.767331909841544e-05, "loss": 6.3201, "step": 11500 }, { "epoch": 0.5315415856506782, "grad_norm": 1.6581804752349854, "learning_rate": 9.765127038944285e-05, "loss": 6.3326, "step": 11550 }, { "epoch": 0.5338426314760059, "grad_norm": 1.8560444116592407, "learning_rate": 9.762912021309268e-05, "loss": 6.3576, "step": 11600 }, { "epoch": 0.5361436773013335, "grad_norm": 1.8938666582107544, "learning_rate": 9.76068686165309e-05, "loss": 6.3133, "step": 11650 }, { "epoch": 0.538444723126661, "grad_norm": 1.8633164167404175, "learning_rate": 9.758451564713951e-05, "loss": 6.3787, "step": 11700 }, { "epoch": 0.5407457689519887, "grad_norm": 1.7775530815124512, "learning_rate": 9.756206135251633e-05, "loss": 6.3219, "step": 11750 }, { "epoch": 0.5430468147773163, "grad_norm": 1.8082398176193237, "learning_rate": 9.75399578841242e-05, "loss": 6.2549, "step": 11800 }, { "epoch": 0.545347860602644, "grad_norm": 1.79505455493927, "learning_rate": 9.751730310680957e-05, "loss": 6.3603, "step": 11850 }, { "epoch": 0.5476489064279715, "grad_norm": 1.7274733781814575, "learning_rate": 9.749454714738381e-05, "loss": 6.3104, "step": 11900 }, { "epoch": 0.5499499522532991, "grad_norm": 1.908056378364563, "learning_rate": 9.747169005430285e-05, "loss": 6.3052, "step": 11950 }, { "epoch": 0.5522509980786268, "grad_norm": 1.8075348138809204, "learning_rate": 9.744873187623798e-05, "loss": 6.2878, "step": 12000 }, { "epoch": 0.5522509980786268, "eval_loss": 6.244420528411865, "eval_runtime": 33.9046, "eval_samples_per_second": 11.326, "eval_steps_per_second": 5.663, "eval_tts_loss": 7.759218772150092, "step": 12000 }, { "epoch": 0.5545520439039543, "grad_norm": 1.722110629081726, "learning_rate": 9.742567266207576e-05, "loss": 6.3152, "step": 12050 }, { "epoch": 0.5568530897292819, "grad_norm": 1.9611711502075195, "learning_rate": 9.74025124609179e-05, "loss": 6.2803, "step": 12100 }, { "epoch": 0.5591541355546096, "grad_norm": 2.127253532409668, "learning_rate": 9.737925132208108e-05, "loss": 6.3487, "step": 12150 }, { "epoch": 0.5614551813799372, "grad_norm": 9.285630226135254, "learning_rate": 9.735588929509701e-05, "loss": 6.2823, "step": 12200 }, { "epoch": 0.5637562272052647, "grad_norm": 2.125577688217163, "learning_rate": 9.733242642971218e-05, "loss": 6.2491, "step": 12250 }, { "epoch": 0.5660572730305924, "grad_norm": 1.9916199445724487, "learning_rate": 9.730886277588777e-05, "loss": 6.3205, "step": 12300 }, { "epoch": 0.56835831885592, "grad_norm": 1.5257749557495117, "learning_rate": 9.728519838379966e-05, "loss": 6.3426, "step": 12350 }, { "epoch": 0.5706593646812477, "grad_norm": 1.864864706993103, "learning_rate": 9.726143330383818e-05, "loss": 6.3545, "step": 12400 }, { "epoch": 0.5729604105065752, "grad_norm": 1.7556637525558472, "learning_rate": 9.723756758660805e-05, "loss": 6.3174, "step": 12450 }, { "epoch": 0.5752614563319028, "grad_norm": 1.8196032047271729, "learning_rate": 9.721360128292832e-05, "loss": 6.2831, "step": 12500 }, { "epoch": 0.5775625021572305, "grad_norm": 1.5368512868881226, "learning_rate": 9.718953444383223e-05, "loss": 6.2929, "step": 12550 }, { "epoch": 0.5798635479825581, "grad_norm": 1.7474430799484253, "learning_rate": 9.716536712056707e-05, "loss": 6.2911, "step": 12600 }, { "epoch": 0.5821645938078857, "grad_norm": 1.7803459167480469, "learning_rate": 9.71410993645941e-05, "loss": 6.2586, "step": 12650 }, { "epoch": 0.5844656396332133, "grad_norm": 1.7328126430511475, "learning_rate": 9.711673122758846e-05, "loss": 6.2789, "step": 12700 }, { "epoch": 0.5867666854585409, "grad_norm": 2.044720411300659, "learning_rate": 9.709226276143903e-05, "loss": 6.3616, "step": 12750 }, { "epoch": 0.5890677312838685, "grad_norm": 1.506191372871399, "learning_rate": 9.706769401824832e-05, "loss": 6.3557, "step": 12800 }, { "epoch": 0.5913687771091961, "grad_norm": 1.9317868947982788, "learning_rate": 9.704302505033239e-05, "loss": 6.2725, "step": 12850 }, { "epoch": 0.5936698229345237, "grad_norm": 1.6422539949417114, "learning_rate": 9.701825591022067e-05, "loss": 6.2711, "step": 12900 }, { "epoch": 0.5959708687598514, "grad_norm": 1.9548308849334717, "learning_rate": 9.699338665065595e-05, "loss": 6.2829, "step": 12950 }, { "epoch": 0.5982719145851789, "grad_norm": 1.6549190282821655, "learning_rate": 9.696841732459418e-05, "loss": 6.3148, "step": 13000 }, { "epoch": 0.5982719145851789, "eval_loss": 6.220914840698242, "eval_runtime": 34.7579, "eval_samples_per_second": 11.048, "eval_steps_per_second": 5.524, "eval_tts_loss": 7.842876913280491, "step": 13000 }, { "epoch": 0.6005729604105066, "grad_norm": 1.860897421836853, "learning_rate": 9.694334798520438e-05, "loss": 6.3158, "step": 13050 }, { "epoch": 0.6028740062358342, "grad_norm": 1.664921522140503, "learning_rate": 9.691817868586856e-05, "loss": 6.3374, "step": 13100 }, { "epoch": 0.6051750520611618, "grad_norm": 1.67441987991333, "learning_rate": 9.689290948018159e-05, "loss": 6.2999, "step": 13150 }, { "epoch": 0.6074760978864894, "grad_norm": 1.8464356660842896, "learning_rate": 9.686754042195102e-05, "loss": 6.3205, "step": 13200 }, { "epoch": 0.609777143711817, "grad_norm": 1.770418405532837, "learning_rate": 9.68420715651971e-05, "loss": 6.3312, "step": 13250 }, { "epoch": 0.6120781895371447, "grad_norm": 1.8769677877426147, "learning_rate": 9.681650296415256e-05, "loss": 6.3142, "step": 13300 }, { "epoch": 0.6143792353624723, "grad_norm": 1.7331925630569458, "learning_rate": 9.679083467326247e-05, "loss": 6.2706, "step": 13350 }, { "epoch": 0.6166802811877998, "grad_norm": 2.1047513484954834, "learning_rate": 9.676506674718426e-05, "loss": 6.2648, "step": 13400 }, { "epoch": 0.6189813270131275, "grad_norm": 2.162065267562866, "learning_rate": 9.673919924078745e-05, "loss": 6.2775, "step": 13450 }, { "epoch": 0.6212823728384551, "grad_norm": 1.890669345855713, "learning_rate": 9.671323220915367e-05, "loss": 6.3046, "step": 13500 }, { "epoch": 0.6235834186637826, "grad_norm": 1.7662500143051147, "learning_rate": 9.668716570757644e-05, "loss": 6.336, "step": 13550 }, { "epoch": 0.6258844644891103, "grad_norm": 1.6510889530181885, "learning_rate": 9.666099979156106e-05, "loss": 6.3583, "step": 13600 }, { "epoch": 0.6281855103144379, "grad_norm": 1.7694858312606812, "learning_rate": 9.663473451682457e-05, "loss": 6.2774, "step": 13650 }, { "epoch": 0.6304865561397656, "grad_norm": 1.576418161392212, "learning_rate": 9.660836993929557e-05, "loss": 6.2338, "step": 13700 }, { "epoch": 0.6327876019650931, "grad_norm": 2.029823064804077, "learning_rate": 9.658190611511411e-05, "loss": 6.3032, "step": 13750 }, { "epoch": 0.6350886477904207, "grad_norm": 1.8009357452392578, "learning_rate": 9.655534310063155e-05, "loss": 6.2281, "step": 13800 }, { "epoch": 0.6373896936157484, "grad_norm": 1.6074466705322266, "learning_rate": 9.65286809524105e-05, "loss": 6.3638, "step": 13850 }, { "epoch": 0.639690739441076, "grad_norm": 1.4809355735778809, "learning_rate": 9.650191972722464e-05, "loss": 6.3167, "step": 13900 }, { "epoch": 0.6419917852664035, "grad_norm": 1.7859705686569214, "learning_rate": 9.647505948205864e-05, "loss": 6.1846, "step": 13950 }, { "epoch": 0.6442928310917312, "grad_norm": 1.7351763248443604, "learning_rate": 9.644810027410798e-05, "loss": 6.2017, "step": 14000 }, { "epoch": 0.6442928310917312, "eval_loss": 6.187459468841553, "eval_runtime": 35.2876, "eval_samples_per_second": 10.882, "eval_steps_per_second": 5.441, "eval_tts_loss": 7.89450947834087, "step": 14000 }, { "epoch": 0.6465938769170588, "grad_norm": 1.6009070873260498, "learning_rate": 9.642104216077894e-05, "loss": 6.27, "step": 14050 }, { "epoch": 0.6488949227423864, "grad_norm": 1.9434318542480469, "learning_rate": 9.639388519968831e-05, "loss": 6.246, "step": 14100 }, { "epoch": 0.651195968567714, "grad_norm": 1.9468326568603516, "learning_rate": 9.636662944866346e-05, "loss": 6.277, "step": 14150 }, { "epoch": 0.6534970143930416, "grad_norm": 1.8342995643615723, "learning_rate": 9.633927496574207e-05, "loss": 6.2374, "step": 14200 }, { "epoch": 0.6557980602183693, "grad_norm": 1.743561029434204, "learning_rate": 9.631182180917204e-05, "loss": 6.2423, "step": 14250 }, { "epoch": 0.6580991060436968, "grad_norm": 1.6310720443725586, "learning_rate": 9.628427003741145e-05, "loss": 6.2321, "step": 14300 }, { "epoch": 0.6604001518690245, "grad_norm": 1.880783200263977, "learning_rate": 9.625661970912829e-05, "loss": 6.1948, "step": 14350 }, { "epoch": 0.6627011976943521, "grad_norm": 1.8415277004241943, "learning_rate": 9.622887088320049e-05, "loss": 6.2053, "step": 14400 }, { "epoch": 0.6650022435196797, "grad_norm": 2.1539406776428223, "learning_rate": 9.620102361871564e-05, "loss": 6.2076, "step": 14450 }, { "epoch": 0.6673032893450073, "grad_norm": 1.9872362613677979, "learning_rate": 9.617307797497099e-05, "loss": 6.2048, "step": 14500 }, { "epoch": 0.6696043351703349, "grad_norm": 2.019122838973999, "learning_rate": 9.614503401147328e-05, "loss": 6.2539, "step": 14550 }, { "epoch": 0.6719053809956625, "grad_norm": 1.6095964908599854, "learning_rate": 9.61168917879386e-05, "loss": 6.3102, "step": 14600 }, { "epoch": 0.6742064268209902, "grad_norm": 1.7596853971481323, "learning_rate": 9.608865136429226e-05, "loss": 6.2507, "step": 14650 }, { "epoch": 0.6765074726463177, "grad_norm": 1.6878117322921753, "learning_rate": 9.60603128006687e-05, "loss": 6.2378, "step": 14700 }, { "epoch": 0.6788085184716454, "grad_norm": 1.8443124294281006, "learning_rate": 9.603187615741129e-05, "loss": 6.2655, "step": 14750 }, { "epoch": 0.681109564296973, "grad_norm": 1.8682785034179688, "learning_rate": 9.60033414950723e-05, "loss": 6.1037, "step": 14800 }, { "epoch": 0.6834106101223005, "grad_norm": 1.932824730873108, "learning_rate": 9.597470887441271e-05, "loss": 6.208, "step": 14850 }, { "epoch": 0.6857116559476282, "grad_norm": 1.65823495388031, "learning_rate": 9.594597835640207e-05, "loss": 6.2669, "step": 14900 }, { "epoch": 0.6880127017729558, "grad_norm": 1.7727546691894531, "learning_rate": 9.591715000221838e-05, "loss": 6.1912, "step": 14950 }, { "epoch": 0.6903137475982835, "grad_norm": 1.9301639795303345, "learning_rate": 9.588822387324801e-05, "loss": 6.2252, "step": 15000 }, { "epoch": 0.6903137475982835, "eval_loss": 6.159511566162109, "eval_runtime": 32.3151, "eval_samples_per_second": 11.883, "eval_steps_per_second": 5.941, "eval_tts_loss": 7.847364975218921, "step": 15000 }, { "epoch": 0.692614793423611, "grad_norm": 1.5824342966079712, "learning_rate": 9.58592000310855e-05, "loss": 6.2213, "step": 15050 }, { "epoch": 0.6949158392489386, "grad_norm": 1.6961408853530884, "learning_rate": 9.583007853753347e-05, "loss": 6.2481, "step": 15100 }, { "epoch": 0.6972168850742663, "grad_norm": 1.7308921813964844, "learning_rate": 9.580085945460248e-05, "loss": 6.2166, "step": 15150 }, { "epoch": 0.6995179308995939, "grad_norm": 1.796118974685669, "learning_rate": 9.577154284451087e-05, "loss": 6.2793, "step": 15200 }, { "epoch": 0.7018189767249214, "grad_norm": 1.796897053718567, "learning_rate": 9.574212876968467e-05, "loss": 6.1633, "step": 15250 }, { "epoch": 0.7041200225502491, "grad_norm": 1.9501514434814453, "learning_rate": 9.571261729275745e-05, "loss": 6.2368, "step": 15300 }, { "epoch": 0.7064210683755767, "grad_norm": 2.005586862564087, "learning_rate": 9.568300847657019e-05, "loss": 6.1743, "step": 15350 }, { "epoch": 0.7087221142009044, "grad_norm": 1.8299839496612549, "learning_rate": 9.565330238417112e-05, "loss": 6.1763, "step": 15400 }, { "epoch": 0.7110231600262319, "grad_norm": 1.7406200170516968, "learning_rate": 9.562349907881563e-05, "loss": 6.2527, "step": 15450 }, { "epoch": 0.7133242058515595, "grad_norm": 1.6860395669937134, "learning_rate": 9.55935986239661e-05, "loss": 6.2061, "step": 15500 }, { "epoch": 0.7156252516768872, "grad_norm": 1.9635874032974243, "learning_rate": 9.55636010832918e-05, "loss": 6.1784, "step": 15550 }, { "epoch": 0.7179262975022147, "grad_norm": 1.8998063802719116, "learning_rate": 9.55335065206687e-05, "loss": 6.2378, "step": 15600 }, { "epoch": 0.7202273433275423, "grad_norm": 1.7541154623031616, "learning_rate": 9.550331500017937e-05, "loss": 6.238, "step": 15650 }, { "epoch": 0.72252838915287, "grad_norm": 1.7203463315963745, "learning_rate": 9.547302658611287e-05, "loss": 6.1993, "step": 15700 }, { "epoch": 0.7248294349781976, "grad_norm": 1.8435782194137573, "learning_rate": 9.544264134296455e-05, "loss": 6.2433, "step": 15750 }, { "epoch": 0.7271304808035252, "grad_norm": 1.6835908889770508, "learning_rate": 9.541215933543595e-05, "loss": 6.219, "step": 15800 }, { "epoch": 0.7294315266288528, "grad_norm": 1.8939025402069092, "learning_rate": 9.538219314980872e-05, "loss": 6.2096, "step": 15850 }, { "epoch": 0.7317325724541804, "grad_norm": 1.8113733530044556, "learning_rate": 9.535151974049603e-05, "loss": 6.214, "step": 15900 }, { "epoch": 0.7340336182795081, "grad_norm": 1.863580584526062, "learning_rate": 9.532074976083503e-05, "loss": 6.2043, "step": 15950 }, { "epoch": 0.7363346641048356, "grad_norm": 1.762188196182251, "learning_rate": 9.528988327634654e-05, "loss": 6.2466, "step": 16000 }, { "epoch": 0.7363346641048356, "eval_loss": 6.129319667816162, "eval_runtime": 33.6388, "eval_samples_per_second": 11.415, "eval_steps_per_second": 5.708, "eval_tts_loss": 7.892954236130724, "step": 16000 }, { "epoch": 0.7386357099301633, "grad_norm": 1.8499406576156616, "learning_rate": 9.525892035275682e-05, "loss": 6.2367, "step": 16050 }, { "epoch": 0.7409367557554909, "grad_norm": 1.8135892152786255, "learning_rate": 9.522786105599754e-05, "loss": 6.1914, "step": 16100 }, { "epoch": 0.7432378015808185, "grad_norm": 1.6833168268203735, "learning_rate": 9.519670545220556e-05, "loss": 6.1262, "step": 16150 }, { "epoch": 0.7455388474061461, "grad_norm": 1.6315560340881348, "learning_rate": 9.516545360772282e-05, "loss": 6.1558, "step": 16200 }, { "epoch": 0.7478398932314737, "grad_norm": 1.9072283506393433, "learning_rate": 9.513410558909617e-05, "loss": 6.1518, "step": 16250 }, { "epoch": 0.7501409390568013, "grad_norm": 1.840588092803955, "learning_rate": 9.510266146307725e-05, "loss": 6.2151, "step": 16300 }, { "epoch": 0.7524419848821289, "grad_norm": 1.820553183555603, "learning_rate": 9.507112129662242e-05, "loss": 6.2398, "step": 16350 }, { "epoch": 0.7547430307074565, "grad_norm": 1.8674460649490356, "learning_rate": 9.503948515689243e-05, "loss": 6.1383, "step": 16400 }, { "epoch": 0.7570440765327842, "grad_norm": 2.053492784500122, "learning_rate": 9.500775311125247e-05, "loss": 6.1975, "step": 16450 }, { "epoch": 0.7593451223581118, "grad_norm": 1.882986307144165, "learning_rate": 9.497592522727197e-05, "loss": 6.1723, "step": 16500 }, { "epoch": 0.7616461681834393, "grad_norm": 1.6095331907272339, "learning_rate": 9.494400157272438e-05, "loss": 6.1988, "step": 16550 }, { "epoch": 0.763947214008767, "grad_norm": 1.6419812440872192, "learning_rate": 9.491198221558708e-05, "loss": 6.1695, "step": 16600 }, { "epoch": 0.7662482598340946, "grad_norm": 1.678825855255127, "learning_rate": 9.487986722404127e-05, "loss": 6.1044, "step": 16650 }, { "epoch": 0.7685493056594223, "grad_norm": 1.7096123695373535, "learning_rate": 9.48476566664718e-05, "loss": 6.1511, "step": 16700 }, { "epoch": 0.7708503514847498, "grad_norm": 2.1413791179656982, "learning_rate": 9.481535061146698e-05, "loss": 6.1708, "step": 16750 }, { "epoch": 0.7731513973100774, "grad_norm": 1.9227194786071777, "learning_rate": 9.478294912781847e-05, "loss": 6.1949, "step": 16800 }, { "epoch": 0.7754524431354051, "grad_norm": 1.8362767696380615, "learning_rate": 9.475045228452115e-05, "loss": 6.1788, "step": 16850 }, { "epoch": 0.7777534889607327, "grad_norm": 1.8240971565246582, "learning_rate": 9.471786015077298e-05, "loss": 6.1563, "step": 16900 }, { "epoch": 0.7800545347860602, "grad_norm": 2.078509569168091, "learning_rate": 9.468517279597477e-05, "loss": 6.1832, "step": 16950 }, { "epoch": 0.7823555806113879, "grad_norm": 2.1859052181243896, "learning_rate": 9.465239028973017e-05, "loss": 6.2047, "step": 17000 }, { "epoch": 0.7823555806113879, "eval_loss": 6.096608638763428, "eval_runtime": 34.1394, "eval_samples_per_second": 11.248, "eval_steps_per_second": 5.624, "eval_tts_loss": 7.896921396746157, "step": 17000 }, { "epoch": 0.7846566264367155, "grad_norm": 1.922098159790039, "learning_rate": 9.461951270184534e-05, "loss": 6.1982, "step": 17050 }, { "epoch": 0.786957672262043, "grad_norm": 1.8664427995681763, "learning_rate": 9.458654010232901e-05, "loss": 6.1964, "step": 17100 }, { "epoch": 0.7892587180873707, "grad_norm": 1.889997124671936, "learning_rate": 9.455347256139215e-05, "loss": 6.1673, "step": 17150 }, { "epoch": 0.7915597639126983, "grad_norm": 1.7203173637390137, "learning_rate": 9.452031014944792e-05, "loss": 6.1519, "step": 17200 }, { "epoch": 0.793860809738026, "grad_norm": 2.062267303466797, "learning_rate": 9.448705293711151e-05, "loss": 6.1475, "step": 17250 }, { "epoch": 0.7961618555633535, "grad_norm": 1.6810802221298218, "learning_rate": 9.445370099519998e-05, "loss": 6.1636, "step": 17300 }, { "epoch": 0.7984629013886811, "grad_norm": 1.6828442811965942, "learning_rate": 9.442025439473207e-05, "loss": 6.1879, "step": 17350 }, { "epoch": 0.8007639472140088, "grad_norm": 1.7201191186904907, "learning_rate": 9.438671320692809e-05, "loss": 6.2027, "step": 17400 }, { "epoch": 0.8030649930393364, "grad_norm": 1.9775705337524414, "learning_rate": 9.435307750320979e-05, "loss": 6.1444, "step": 17450 }, { "epoch": 0.805366038864664, "grad_norm": 1.6285938024520874, "learning_rate": 9.431934735520019e-05, "loss": 6.0895, "step": 17500 }, { "epoch": 0.8076670846899916, "grad_norm": 2.1120924949645996, "learning_rate": 9.428552283472333e-05, "loss": 6.0707, "step": 17550 }, { "epoch": 0.8099681305153192, "grad_norm": 1.9405549764633179, "learning_rate": 9.425160401380431e-05, "loss": 6.1812, "step": 17600 }, { "epoch": 0.8122691763406468, "grad_norm": 1.8059804439544678, "learning_rate": 9.421759096466897e-05, "loss": 6.2241, "step": 17650 }, { "epoch": 0.8145702221659744, "grad_norm": 1.6747099161148071, "learning_rate": 9.418348375974384e-05, "loss": 6.1178, "step": 17700 }, { "epoch": 0.816871267991302, "grad_norm": 1.9652963876724243, "learning_rate": 9.414928247165587e-05, "loss": 6.1801, "step": 17750 }, { "epoch": 0.8191723138166297, "grad_norm": 1.6845933198928833, "learning_rate": 9.411498717323244e-05, "loss": 6.1187, "step": 17800 }, { "epoch": 0.8214733596419572, "grad_norm": 2.096360206604004, "learning_rate": 9.408059793750103e-05, "loss": 6.1353, "step": 17850 }, { "epoch": 0.8237744054672849, "grad_norm": 1.7789738178253174, "learning_rate": 9.404611483768922e-05, "loss": 6.0715, "step": 17900 }, { "epoch": 0.8260754512926125, "grad_norm": 2.0107250213623047, "learning_rate": 9.401153794722441e-05, "loss": 6.1532, "step": 17950 }, { "epoch": 0.8283764971179401, "grad_norm": 1.776755452156067, "learning_rate": 9.397686733973372e-05, "loss": 6.1186, "step": 18000 }, { "epoch": 0.8283764971179401, "eval_loss": 6.063215732574463, "eval_runtime": 33.1049, "eval_samples_per_second": 11.599, "eval_steps_per_second": 5.8, "eval_tts_loss": 7.988993335909586, "step": 18000 }, { "epoch": 0.8306775429432677, "grad_norm": 1.5279197692871094, "learning_rate": 9.394210308904386e-05, "loss": 6.1574, "step": 18050 }, { "epoch": 0.8329785887685953, "grad_norm": 1.7856347560882568, "learning_rate": 9.390724526918091e-05, "loss": 6.0909, "step": 18100 }, { "epoch": 0.835279634593923, "grad_norm": 1.844883680343628, "learning_rate": 9.387229395437023e-05, "loss": 6.1327, "step": 18150 }, { "epoch": 0.8375806804192506, "grad_norm": 1.9506162405014038, "learning_rate": 9.38372492190362e-05, "loss": 6.0869, "step": 18200 }, { "epoch": 0.8398817262445781, "grad_norm": 1.6576199531555176, "learning_rate": 9.380211113780222e-05, "loss": 6.1195, "step": 18250 }, { "epoch": 0.8421827720699058, "grad_norm": 1.8321812152862549, "learning_rate": 9.376687978549037e-05, "loss": 6.1679, "step": 18300 }, { "epoch": 0.8444838178952334, "grad_norm": 2.062652587890625, "learning_rate": 9.373155523712138e-05, "loss": 6.0964, "step": 18350 }, { "epoch": 0.8467848637205609, "grad_norm": 1.7374162673950195, "learning_rate": 9.369613756791445e-05, "loss": 6.1764, "step": 18400 }, { "epoch": 0.8490859095458886, "grad_norm": 1.9153779745101929, "learning_rate": 9.366062685328703e-05, "loss": 6.1341, "step": 18450 }, { "epoch": 0.8513869553712162, "grad_norm": 1.7411820888519287, "learning_rate": 9.362502316885469e-05, "loss": 6.1364, "step": 18500 }, { "epoch": 0.8536880011965439, "grad_norm": 2.0500361919403076, "learning_rate": 9.358932659043102e-05, "loss": 6.0828, "step": 18550 }, { "epoch": 0.8559890470218714, "grad_norm": 1.7861090898513794, "learning_rate": 9.35535371940274e-05, "loss": 6.138, "step": 18600 }, { "epoch": 0.858290092847199, "grad_norm": 2.1005265712738037, "learning_rate": 9.351837360699179e-05, "loss": 6.1243, "step": 18650 }, { "epoch": 0.8605911386725267, "grad_norm": 1.8733184337615967, "learning_rate": 9.348240065600997e-05, "loss": 6.0982, "step": 18700 }, { "epoch": 0.8628921844978543, "grad_norm": 1.5661227703094482, "learning_rate": 9.344633511473347e-05, "loss": 6.0685, "step": 18750 }, { "epoch": 0.8651932303231818, "grad_norm": 2.270904064178467, "learning_rate": 9.341017705995936e-05, "loss": 6.0827, "step": 18800 }, { "epoch": 0.8674942761485095, "grad_norm": 1.945278525352478, "learning_rate": 9.337392656868162e-05, "loss": 6.0418, "step": 18850 }, { "epoch": 0.8697953219738371, "grad_norm": 1.6341246366500854, "learning_rate": 9.333758371809113e-05, "loss": 6.1227, "step": 18900 }, { "epoch": 0.8720963677991648, "grad_norm": 1.9579964876174927, "learning_rate": 9.330114858557541e-05, "loss": 6.1275, "step": 18950 }, { "epoch": 0.8743974136244923, "grad_norm": 2.0281152725219727, "learning_rate": 9.326462124871846e-05, "loss": 6.0869, "step": 19000 }, { "epoch": 0.8743974136244923, "eval_loss": 6.040208339691162, "eval_runtime": 34.65, "eval_samples_per_second": 11.082, "eval_steps_per_second": 5.541, "eval_tts_loss": 8.007550918315454, "step": 19000 }, { "epoch": 0.87669845944982, "grad_norm": 1.834215521812439, "learning_rate": 9.322800178530069e-05, "loss": 6.1467, "step": 19050 }, { "epoch": 0.8789995052751476, "grad_norm": 1.8311221599578857, "learning_rate": 9.31912902732986e-05, "loss": 6.0677, "step": 19100 }, { "epoch": 0.8813005511004751, "grad_norm": 1.7210936546325684, "learning_rate": 9.315448679088472e-05, "loss": 6.0809, "step": 19150 }, { "epoch": 0.8836015969258028, "grad_norm": 1.8017765283584595, "learning_rate": 9.311759141642746e-05, "loss": 6.0724, "step": 19200 }, { "epoch": 0.8859026427511304, "grad_norm": 1.822521448135376, "learning_rate": 9.308060422849085e-05, "loss": 6.0447, "step": 19250 }, { "epoch": 0.888203688576458, "grad_norm": 2.083303451538086, "learning_rate": 9.304352530583447e-05, "loss": 6.1126, "step": 19300 }, { "epoch": 0.8905047344017856, "grad_norm": 1.7161856889724731, "learning_rate": 9.30063547274132e-05, "loss": 6.1106, "step": 19350 }, { "epoch": 0.8928057802271132, "grad_norm": 2.1552722454071045, "learning_rate": 9.296909257237707e-05, "loss": 6.0917, "step": 19400 }, { "epoch": 0.8951068260524409, "grad_norm": 1.898194432258606, "learning_rate": 9.293173892007122e-05, "loss": 6.1337, "step": 19450 }, { "epoch": 0.8974078718777685, "grad_norm": 1.6852689981460571, "learning_rate": 9.28942938500355e-05, "loss": 6.1289, "step": 19500 }, { "epoch": 0.899708917703096, "grad_norm": 2.3115508556365967, "learning_rate": 9.285675744200447e-05, "loss": 6.1953, "step": 19550 }, { "epoch": 0.9020099635284237, "grad_norm": 1.9407068490982056, "learning_rate": 9.281912977590719e-05, "loss": 6.0774, "step": 19600 }, { "epoch": 0.9043110093537513, "grad_norm": 1.9050291776657104, "learning_rate": 9.278141093186703e-05, "loss": 6.077, "step": 19650 }, { "epoch": 0.906612055179079, "grad_norm": 1.570671796798706, "learning_rate": 9.274360099020154e-05, "loss": 6.0458, "step": 19700 }, { "epoch": 0.9089131010044065, "grad_norm": 1.8827368021011353, "learning_rate": 9.270570003142218e-05, "loss": 6.0504, "step": 19750 }, { "epoch": 0.9112141468297341, "grad_norm": 1.8248310089111328, "learning_rate": 9.266770813623432e-05, "loss": 6.0229, "step": 19800 }, { "epoch": 0.9135151926550618, "grad_norm": 1.6708036661148071, "learning_rate": 9.262962538553689e-05, "loss": 6.0974, "step": 19850 }, { "epoch": 0.9158162384803893, "grad_norm": 1.7397429943084717, "learning_rate": 9.259145186042231e-05, "loss": 6.0746, "step": 19900 }, { "epoch": 0.9181172843057169, "grad_norm": 1.5898857116699219, "learning_rate": 9.255318764217629e-05, "loss": 6.0868, "step": 19950 }, { "epoch": 0.9204183301310446, "grad_norm": 1.8456883430480957, "learning_rate": 9.251483281227767e-05, "loss": 6.0739, "step": 20000 }, { "epoch": 0.9204183301310446, "eval_loss": 6.005526065826416, "eval_runtime": 34.2528, "eval_samples_per_second": 11.211, "eval_steps_per_second": 5.605, "eval_tts_loss": 7.991646247745309, "step": 20000 }, { "epoch": 0.9227193759563722, "grad_norm": 1.8995615243911743, "learning_rate": 9.247638745239822e-05, "loss": 6.1174, "step": 20050 }, { "epoch": 0.9250204217816997, "grad_norm": 1.6712597608566284, "learning_rate": 9.243785164440251e-05, "loss": 6.0987, "step": 20100 }, { "epoch": 0.9273214676070274, "grad_norm": 1.762939691543579, "learning_rate": 9.239922547034765e-05, "loss": 5.9939, "step": 20150 }, { "epoch": 0.929622513432355, "grad_norm": 1.6403757333755493, "learning_rate": 9.236050901248325e-05, "loss": 6.0611, "step": 20200 }, { "epoch": 0.9319235592576827, "grad_norm": 2.0031139850616455, "learning_rate": 9.232170235325113e-05, "loss": 6.1203, "step": 20250 }, { "epoch": 0.9342246050830102, "grad_norm": 1.9146487712860107, "learning_rate": 9.228280557528516e-05, "loss": 6.0571, "step": 20300 }, { "epoch": 0.9365256509083378, "grad_norm": 1.7245732545852661, "learning_rate": 9.224381876141113e-05, "loss": 6.0659, "step": 20350 }, { "epoch": 0.9388266967336655, "grad_norm": 1.967654824256897, "learning_rate": 9.220474199464657e-05, "loss": 6.0195, "step": 20400 }, { "epoch": 0.941127742558993, "grad_norm": 1.929421067237854, "learning_rate": 9.216557535820052e-05, "loss": 6.0432, "step": 20450 }, { "epoch": 0.9434287883843206, "grad_norm": 1.8827496767044067, "learning_rate": 9.212631893547339e-05, "loss": 6.0518, "step": 20500 }, { "epoch": 0.9457298342096483, "grad_norm": 1.6459333896636963, "learning_rate": 9.20869728100568e-05, "loss": 6.0171, "step": 20550 }, { "epoch": 0.9480308800349759, "grad_norm": 2.004542827606201, "learning_rate": 9.204753706573335e-05, "loss": 6.047, "step": 20600 }, { "epoch": 0.9503319258603035, "grad_norm": 1.9624686241149902, "learning_rate": 9.200801178647652e-05, "loss": 6.0593, "step": 20650 }, { "epoch": 0.9526329716856311, "grad_norm": 1.5215227603912354, "learning_rate": 9.196839705645039e-05, "loss": 6.0869, "step": 20700 }, { "epoch": 0.9549340175109587, "grad_norm": 1.7919971942901611, "learning_rate": 9.192869296000954e-05, "loss": 5.9958, "step": 20750 }, { "epoch": 0.9572350633362864, "grad_norm": 1.7097506523132324, "learning_rate": 9.188889958169884e-05, "loss": 6.0802, "step": 20800 }, { "epoch": 0.9595361091616139, "grad_norm": 1.670233130455017, "learning_rate": 9.184901700625328e-05, "loss": 6.0703, "step": 20850 }, { "epoch": 0.9618371549869416, "grad_norm": 2.0614235401153564, "learning_rate": 9.18090453185978e-05, "loss": 6.0206, "step": 20900 }, { "epoch": 0.9641382008122692, "grad_norm": 1.811000108718872, "learning_rate": 9.176898460384704e-05, "loss": 5.9914, "step": 20950 }, { "epoch": 0.9664392466375968, "grad_norm": 1.9324144124984741, "learning_rate": 9.172883494730526e-05, "loss": 6.0325, "step": 21000 }, { "epoch": 0.9664392466375968, "eval_loss": 5.979398250579834, "eval_runtime": 33.2361, "eval_samples_per_second": 11.554, "eval_steps_per_second": 5.777, "eval_tts_loss": 8.089930703009033, "step": 21000 }, { "epoch": 0.9687402924629244, "grad_norm": 1.9032566547393799, "learning_rate": 9.16885964344661e-05, "loss": 6.0419, "step": 21050 }, { "epoch": 0.971041338288252, "grad_norm": 2.04426908493042, "learning_rate": 9.164826915101239e-05, "loss": 6.0501, "step": 21100 }, { "epoch": 0.9733423841135797, "grad_norm": 1.8017916679382324, "learning_rate": 9.160785318281601e-05, "loss": 6.0282, "step": 21150 }, { "epoch": 0.9756434299389072, "grad_norm": 1.8715277910232544, "learning_rate": 9.156815957498479e-05, "loss": 6.0933, "step": 21200 }, { "epoch": 0.9779444757642348, "grad_norm": 1.7041345834732056, "learning_rate": 9.152756826507607e-05, "loss": 6.0776, "step": 21250 }, { "epoch": 0.9802455215895625, "grad_norm": 1.8159674406051636, "learning_rate": 9.148688852744201e-05, "loss": 6.0045, "step": 21300 }, { "epoch": 0.9825465674148901, "grad_norm": 2.1322624683380127, "learning_rate": 9.144612044870497e-05, "loss": 6.0547, "step": 21350 }, { "epoch": 0.9848476132402176, "grad_norm": 1.899408221244812, "learning_rate": 9.14052641156755e-05, "loss": 6.0359, "step": 21400 }, { "epoch": 0.9871486590655453, "grad_norm": 1.912842869758606, "learning_rate": 9.136431961535193e-05, "loss": 6.032, "step": 21450 }, { "epoch": 0.9894497048908729, "grad_norm": 2.2110729217529297, "learning_rate": 9.132328703492048e-05, "loss": 6.0135, "step": 21500 }, { "epoch": 0.9917507507162006, "grad_norm": 1.8513312339782715, "learning_rate": 9.128216646175482e-05, "loss": 5.9792, "step": 21550 }, { "epoch": 0.9940517965415281, "grad_norm": 1.685534954071045, "learning_rate": 9.124095798341607e-05, "loss": 5.9946, "step": 21600 }, { "epoch": 0.9963528423668557, "grad_norm": 1.8378783464431763, "learning_rate": 9.119966168765246e-05, "loss": 5.9931, "step": 21650 }, { "epoch": 0.9986538881921834, "grad_norm": 1.9953211545944214, "learning_rate": 9.115827766239925e-05, "loss": 6.0143, "step": 21700 }, { "epoch": 1.000920418330131, "grad_norm": 1.8958375453948975, "learning_rate": 9.111680599577854e-05, "loss": 5.9748, "step": 21750 }, { "epoch": 1.0032214641554587, "grad_norm": 1.59896719455719, "learning_rate": 9.1075246776099e-05, "loss": 5.8144, "step": 21800 }, { "epoch": 1.0055225099807863, "grad_norm": 1.7725744247436523, "learning_rate": 9.103360009185574e-05, "loss": 5.8584, "step": 21850 }, { "epoch": 1.007823555806114, "grad_norm": 2.1068501472473145, "learning_rate": 9.099186603173016e-05, "loss": 5.9851, "step": 21900 }, { "epoch": 1.0101246016314416, "grad_norm": 2.0774874687194824, "learning_rate": 9.095004468458965e-05, "loss": 5.8846, "step": 21950 }, { "epoch": 1.012425647456769, "grad_norm": 2.4163296222686768, "learning_rate": 9.090813613948752e-05, "loss": 5.9117, "step": 22000 }, { "epoch": 1.012425647456769, "eval_loss": 5.969619274139404, "eval_runtime": 34.7535, "eval_samples_per_second": 11.049, "eval_steps_per_second": 5.525, "eval_tts_loss": 8.110409582485643, "step": 22000 }, { "epoch": 1.0147266932820966, "grad_norm": 2.0644235610961914, "learning_rate": 9.086614048566273e-05, "loss": 5.9255, "step": 22050 }, { "epoch": 1.0170277391074243, "grad_norm": 1.7027440071105957, "learning_rate": 9.082405781253973e-05, "loss": 5.8677, "step": 22100 }, { "epoch": 1.019328784932752, "grad_norm": 2.0452287197113037, "learning_rate": 9.07818882097283e-05, "loss": 5.9431, "step": 22150 }, { "epoch": 1.0216298307580796, "grad_norm": 1.6214035749435425, "learning_rate": 9.073963176702324e-05, "loss": 5.9106, "step": 22200 }, { "epoch": 1.0239308765834072, "grad_norm": 1.7072349786758423, "learning_rate": 9.06972885744044e-05, "loss": 5.8734, "step": 22250 }, { "epoch": 1.0262319224087348, "grad_norm": 1.7980152368545532, "learning_rate": 9.06548587220362e-05, "loss": 5.9673, "step": 22300 }, { "epoch": 1.0285329682340625, "grad_norm": 2.011014938354492, "learning_rate": 9.061234230026771e-05, "loss": 5.9262, "step": 22350 }, { "epoch": 1.03083401405939, "grad_norm": 2.122554302215576, "learning_rate": 9.05697393996323e-05, "loss": 5.9159, "step": 22400 }, { "epoch": 1.0331350598847175, "grad_norm": 1.9338388442993164, "learning_rate": 9.052705011084743e-05, "loss": 5.8578, "step": 22450 }, { "epoch": 1.0354361057100452, "grad_norm": 2.0208539962768555, "learning_rate": 9.048427452481462e-05, "loss": 5.9803, "step": 22500 }, { "epoch": 1.0377371515353728, "grad_norm": 1.9141596555709839, "learning_rate": 9.044141273261906e-05, "loss": 5.9241, "step": 22550 }, { "epoch": 1.0400381973607005, "grad_norm": 1.8084018230438232, "learning_rate": 9.039846482552953e-05, "loss": 5.9547, "step": 22600 }, { "epoch": 1.042339243186028, "grad_norm": 1.9551680088043213, "learning_rate": 9.03554308949982e-05, "loss": 5.9616, "step": 22650 }, { "epoch": 1.0446402890113557, "grad_norm": 1.8424718379974365, "learning_rate": 9.031231103266044e-05, "loss": 5.9839, "step": 22700 }, { "epoch": 1.0469413348366832, "grad_norm": 1.7704012393951416, "learning_rate": 9.02691053303345e-05, "loss": 5.9764, "step": 22750 }, { "epoch": 1.0492423806620108, "grad_norm": 2.2767720222473145, "learning_rate": 9.022581388002152e-05, "loss": 5.9506, "step": 22800 }, { "epoch": 1.0515434264873385, "grad_norm": 1.8481018543243408, "learning_rate": 9.018243677390518e-05, "loss": 5.8892, "step": 22850 }, { "epoch": 1.053844472312666, "grad_norm": 1.9781994819641113, "learning_rate": 9.013897410435157e-05, "loss": 5.8882, "step": 22900 }, { "epoch": 1.0561455181379937, "grad_norm": 2.307131767272949, "learning_rate": 9.009542596390896e-05, "loss": 5.8953, "step": 22950 }, { "epoch": 1.0584465639633214, "grad_norm": 1.830885648727417, "learning_rate": 9.005179244530764e-05, "loss": 5.902, "step": 23000 }, { "epoch": 1.0584465639633214, "eval_loss": 5.945216655731201, "eval_runtime": 34.7154, "eval_samples_per_second": 11.061, "eval_steps_per_second": 5.531, "eval_tts_loss": 8.16352996774057, "step": 23000 }, { "epoch": 1.060747609788649, "grad_norm": 2.0037059783935547, "learning_rate": 9.000807364145967e-05, "loss": 5.8711, "step": 23050 }, { "epoch": 1.0630486556139767, "grad_norm": 1.691949725151062, "learning_rate": 8.996426964545876e-05, "loss": 5.887, "step": 23100 }, { "epoch": 1.065349701439304, "grad_norm": 1.8151764869689941, "learning_rate": 8.992038055058e-05, "loss": 5.9154, "step": 23150 }, { "epoch": 1.0676507472646317, "grad_norm": 1.672204613685608, "learning_rate": 8.987640645027966e-05, "loss": 5.9166, "step": 23200 }, { "epoch": 1.0699517930899594, "grad_norm": 1.6701843738555908, "learning_rate": 8.983234743819508e-05, "loss": 5.907, "step": 23250 }, { "epoch": 1.072252838915287, "grad_norm": 1.888508677482605, "learning_rate": 8.978820360814436e-05, "loss": 5.8915, "step": 23300 }, { "epoch": 1.0745538847406146, "grad_norm": 2.2695815563201904, "learning_rate": 8.974397505412624e-05, "loss": 5.9346, "step": 23350 }, { "epoch": 1.0768549305659423, "grad_norm": 1.7576676607131958, "learning_rate": 8.969966187031985e-05, "loss": 5.9133, "step": 23400 }, { "epoch": 1.07915597639127, "grad_norm": 1.5666314363479614, "learning_rate": 8.965526415108449e-05, "loss": 5.9285, "step": 23450 }, { "epoch": 1.0814570222165973, "grad_norm": 1.5109493732452393, "learning_rate": 8.961078199095957e-05, "loss": 5.9068, "step": 23500 }, { "epoch": 1.083758068041925, "grad_norm": 1.9691708087921143, "learning_rate": 8.956621548466421e-05, "loss": 5.9005, "step": 23550 }, { "epoch": 1.0860591138672526, "grad_norm": 1.978844404220581, "learning_rate": 8.952245856729633e-05, "loss": 5.9172, "step": 23600 }, { "epoch": 1.0883601596925803, "grad_norm": 1.8941922187805176, "learning_rate": 8.947772533572675e-05, "loss": 5.8864, "step": 23650 }, { "epoch": 1.090661205517908, "grad_norm": 1.850315809249878, "learning_rate": 8.943290804131414e-05, "loss": 5.9281, "step": 23700 }, { "epoch": 1.0929622513432355, "grad_norm": 1.6464835405349731, "learning_rate": 8.938800677949127e-05, "loss": 5.9169, "step": 23750 }, { "epoch": 1.0952632971685632, "grad_norm": 1.6319020986557007, "learning_rate": 8.934302164586972e-05, "loss": 5.8943, "step": 23800 }, { "epoch": 1.0975643429938908, "grad_norm": 1.853194236755371, "learning_rate": 8.929795273623967e-05, "loss": 5.9478, "step": 23850 }, { "epoch": 1.0998653888192182, "grad_norm": 2.057925224304199, "learning_rate": 8.925280014656967e-05, "loss": 5.947, "step": 23900 }, { "epoch": 1.102166434644546, "grad_norm": 1.695457935333252, "learning_rate": 8.920756397300651e-05, "loss": 5.8765, "step": 23950 }, { "epoch": 1.1044674804698735, "grad_norm": 2.0463249683380127, "learning_rate": 8.916224431187487e-05, "loss": 5.9068, "step": 24000 }, { "epoch": 1.1044674804698735, "eval_loss": 5.919960021972656, "eval_runtime": 34.2903, "eval_samples_per_second": 11.199, "eval_steps_per_second": 5.599, "eval_tts_loss": 8.153686654926947, "step": 24000 }, { "epoch": 1.1067685262952012, "grad_norm": 1.822067379951477, "learning_rate": 8.91168412596773e-05, "loss": 5.9224, "step": 24050 }, { "epoch": 1.1090695721205288, "grad_norm": 1.8195558786392212, "learning_rate": 8.907135491309386e-05, "loss": 5.8628, "step": 24100 }, { "epoch": 1.1113706179458565, "grad_norm": 1.7702739238739014, "learning_rate": 8.902578536898202e-05, "loss": 5.843, "step": 24150 }, { "epoch": 1.113671663771184, "grad_norm": 1.775231957435608, "learning_rate": 8.898013272437635e-05, "loss": 5.9523, "step": 24200 }, { "epoch": 1.1159727095965115, "grad_norm": 1.9853113889694214, "learning_rate": 8.893439707648843e-05, "loss": 5.9059, "step": 24250 }, { "epoch": 1.1182737554218392, "grad_norm": 1.9251091480255127, "learning_rate": 8.888857852270655e-05, "loss": 5.8562, "step": 24300 }, { "epoch": 1.1205748012471668, "grad_norm": 1.6936697959899902, "learning_rate": 8.884267716059556e-05, "loss": 5.8354, "step": 24350 }, { "epoch": 1.1228758470724944, "grad_norm": 1.7089874744415283, "learning_rate": 8.87966930878966e-05, "loss": 5.8841, "step": 24400 }, { "epoch": 1.125176892897822, "grad_norm": 1.789920687675476, "learning_rate": 8.8750626402527e-05, "loss": 5.9337, "step": 24450 }, { "epoch": 1.1274779387231497, "grad_norm": 1.7186089754104614, "learning_rate": 8.870447720257994e-05, "loss": 5.9068, "step": 24500 }, { "epoch": 1.1297789845484774, "grad_norm": 2.3741872310638428, "learning_rate": 8.865824558632431e-05, "loss": 5.9303, "step": 24550 }, { "epoch": 1.132080030373805, "grad_norm": 1.9989960193634033, "learning_rate": 8.861193165220456e-05, "loss": 5.9017, "step": 24600 }, { "epoch": 1.1343810761991324, "grad_norm": 1.9211207628250122, "learning_rate": 8.856553549884034e-05, "loss": 5.9274, "step": 24650 }, { "epoch": 1.13668212202446, "grad_norm": 2.168541193008423, "learning_rate": 8.851905722502643e-05, "loss": 5.8614, "step": 24700 }, { "epoch": 1.1389831678497877, "grad_norm": 1.7189666032791138, "learning_rate": 8.847249692973245e-05, "loss": 5.8518, "step": 24750 }, { "epoch": 1.1412842136751153, "grad_norm": 1.8728166818618774, "learning_rate": 8.842585471210271e-05, "loss": 5.924, "step": 24800 }, { "epoch": 1.143585259500443, "grad_norm": 2.2806296348571777, "learning_rate": 8.837913067145589e-05, "loss": 5.8643, "step": 24850 }, { "epoch": 1.1458863053257706, "grad_norm": 2.4408116340637207, "learning_rate": 8.833232490728499e-05, "loss": 5.8813, "step": 24900 }, { "epoch": 1.1481873511510983, "grad_norm": 2.0429434776306152, "learning_rate": 8.828543751925697e-05, "loss": 5.8622, "step": 24950 }, { "epoch": 1.1504883969764257, "grad_norm": 1.8052806854248047, "learning_rate": 8.823846860721262e-05, "loss": 5.8601, "step": 25000 }, { "epoch": 1.1504883969764257, "eval_loss": 5.919439792633057, "eval_runtime": 35.1925, "eval_samples_per_second": 10.911, "eval_steps_per_second": 5.456, "eval_tts_loss": 8.2057244715313, "step": 25000 }, { "epoch": 1.1527894428017533, "grad_norm": 1.8507080078125, "learning_rate": 8.819141827116633e-05, "loss": 5.8693, "step": 25050 }, { "epoch": 1.155090488627081, "grad_norm": 1.8380216360092163, "learning_rate": 8.814428661130584e-05, "loss": 5.8838, "step": 25100 }, { "epoch": 1.1573915344524086, "grad_norm": 1.9006010293960571, "learning_rate": 8.80970737279921e-05, "loss": 5.9039, "step": 25150 }, { "epoch": 1.1596925802777363, "grad_norm": 2.1745150089263916, "learning_rate": 8.804977972175901e-05, "loss": 5.937, "step": 25200 }, { "epoch": 1.161993626103064, "grad_norm": 1.970659613609314, "learning_rate": 8.800240469331315e-05, "loss": 5.87, "step": 25250 }, { "epoch": 1.1642946719283915, "grad_norm": 1.9034065008163452, "learning_rate": 8.795494874353373e-05, "loss": 5.8825, "step": 25300 }, { "epoch": 1.1665957177537192, "grad_norm": 1.7885934114456177, "learning_rate": 8.79074119734722e-05, "loss": 5.9146, "step": 25350 }, { "epoch": 1.1688967635790466, "grad_norm": 1.9243812561035156, "learning_rate": 8.785979448435213e-05, "loss": 5.8155, "step": 25400 }, { "epoch": 1.1711978094043742, "grad_norm": 1.7413113117218018, "learning_rate": 8.781209637756894e-05, "loss": 5.9454, "step": 25450 }, { "epoch": 1.1734988552297019, "grad_norm": 1.7900454998016357, "learning_rate": 8.776431775468974e-05, "loss": 5.8851, "step": 25500 }, { "epoch": 1.1757999010550295, "grad_norm": 2.080112934112549, "learning_rate": 8.771645871745311e-05, "loss": 5.8682, "step": 25550 }, { "epoch": 1.1781009468803572, "grad_norm": 2.1360979080200195, "learning_rate": 8.766851936776886e-05, "loss": 5.8843, "step": 25600 }, { "epoch": 1.1804019927056848, "grad_norm": 1.5990478992462158, "learning_rate": 8.762049980771773e-05, "loss": 5.8718, "step": 25650 }, { "epoch": 1.1827030385310124, "grad_norm": 1.6874678134918213, "learning_rate": 8.75724001395514e-05, "loss": 5.8586, "step": 25700 }, { "epoch": 1.1850040843563399, "grad_norm": 1.6797226667404175, "learning_rate": 8.752422046569198e-05, "loss": 5.8377, "step": 25750 }, { "epoch": 1.1873051301816675, "grad_norm": 1.8640986680984497, "learning_rate": 8.747596088873208e-05, "loss": 5.9506, "step": 25800 }, { "epoch": 1.1896061760069951, "grad_norm": 1.6982637643814087, "learning_rate": 8.742762151143436e-05, "loss": 5.9223, "step": 25850 }, { "epoch": 1.1919072218323228, "grad_norm": 1.8827178478240967, "learning_rate": 8.737920243673143e-05, "loss": 5.8191, "step": 25900 }, { "epoch": 1.1942082676576504, "grad_norm": 1.896240472793579, "learning_rate": 8.733070376772562e-05, "loss": 5.8389, "step": 25950 }, { "epoch": 1.196509313482978, "grad_norm": 2.3180794715881348, "learning_rate": 8.728212560768874e-05, "loss": 5.8701, "step": 26000 }, { "epoch": 1.196509313482978, "eval_loss": 5.89279317855835, "eval_runtime": 34.4045, "eval_samples_per_second": 11.161, "eval_steps_per_second": 5.581, "eval_tts_loss": 8.16692597775296, "step": 26000 }, { "epoch": 1.1988103593083057, "grad_norm": 1.9639935493469238, "learning_rate": 8.723346806006186e-05, "loss": 5.8741, "step": 26050 }, { "epoch": 1.2011114051336333, "grad_norm": 1.6462650299072266, "learning_rate": 8.718473122845508e-05, "loss": 5.8613, "step": 26100 }, { "epoch": 1.2034124509589608, "grad_norm": 1.7803330421447754, "learning_rate": 8.713591521664737e-05, "loss": 5.8446, "step": 26150 }, { "epoch": 1.2057134967842884, "grad_norm": 1.8175888061523438, "learning_rate": 8.708799880462161e-05, "loss": 5.895, "step": 26200 }, { "epoch": 1.208014542609616, "grad_norm": 1.984127402305603, "learning_rate": 8.703902632284438e-05, "loss": 5.8236, "step": 26250 }, { "epoch": 1.2103155884349437, "grad_norm": 1.7971287965774536, "learning_rate": 8.698997497112645e-05, "loss": 5.8432, "step": 26300 }, { "epoch": 1.2126166342602713, "grad_norm": 1.832613229751587, "learning_rate": 8.694084485391647e-05, "loss": 5.8908, "step": 26350 }, { "epoch": 1.214917680085599, "grad_norm": 1.8921012878417969, "learning_rate": 8.689163607583087e-05, "loss": 5.8555, "step": 26400 }, { "epoch": 1.2172187259109264, "grad_norm": 1.7966539859771729, "learning_rate": 8.684234874165354e-05, "loss": 5.9059, "step": 26450 }, { "epoch": 1.219519771736254, "grad_norm": 1.910406231880188, "learning_rate": 8.679298295633565e-05, "loss": 5.9553, "step": 26500 }, { "epoch": 1.2218208175615817, "grad_norm": 1.7525932788848877, "learning_rate": 8.67435388249954e-05, "loss": 5.8169, "step": 26550 }, { "epoch": 1.2241218633869093, "grad_norm": 1.8935388326644897, "learning_rate": 8.669401645291787e-05, "loss": 5.8138, "step": 26600 }, { "epoch": 1.226422909212237, "grad_norm": 2.199080467224121, "learning_rate": 8.664441594555468e-05, "loss": 5.8429, "step": 26650 }, { "epoch": 1.2287239550375646, "grad_norm": 1.8432965278625488, "learning_rate": 8.659473740852388e-05, "loss": 5.847, "step": 26700 }, { "epoch": 1.2310250008628922, "grad_norm": 1.6469765901565552, "learning_rate": 8.654498094760967e-05, "loss": 5.8197, "step": 26750 }, { "epoch": 1.2333260466882199, "grad_norm": 1.7215397357940674, "learning_rate": 8.649514666876214e-05, "loss": 5.8082, "step": 26800 }, { "epoch": 1.2356270925135475, "grad_norm": 1.6139415502548218, "learning_rate": 8.644523467809714e-05, "loss": 5.8674, "step": 26850 }, { "epoch": 1.237928138338875, "grad_norm": 2.062239646911621, "learning_rate": 8.639524508189591e-05, "loss": 5.8828, "step": 26900 }, { "epoch": 1.2402291841642026, "grad_norm": 1.7671411037445068, "learning_rate": 8.634517798660507e-05, "loss": 5.911, "step": 26950 }, { "epoch": 1.2425302299895302, "grad_norm": 1.7960079908370972, "learning_rate": 8.629503349883614e-05, "loss": 5.8929, "step": 27000 }, { "epoch": 1.2425302299895302, "eval_loss": 5.874563694000244, "eval_runtime": 34.9389, "eval_samples_per_second": 10.991, "eval_steps_per_second": 5.495, "eval_tts_loss": 8.201237089489519, "step": 27000 }, { "epoch": 1.2448312758148579, "grad_norm": 1.8907201290130615, "learning_rate": 8.624481172536551e-05, "loss": 5.8062, "step": 27050 }, { "epoch": 1.2471323216401855, "grad_norm": 1.8463737964630127, "learning_rate": 8.619451277313413e-05, "loss": 5.8612, "step": 27100 }, { "epoch": 1.2494333674655131, "grad_norm": 1.5661523342132568, "learning_rate": 8.614413674924726e-05, "loss": 5.7927, "step": 27150 }, { "epoch": 1.2517344132908406, "grad_norm": 1.546129584312439, "learning_rate": 8.60936837609743e-05, "loss": 5.8647, "step": 27200 }, { "epoch": 1.2540354591161682, "grad_norm": 1.8839119672775269, "learning_rate": 8.604315391574856e-05, "loss": 5.8938, "step": 27250 }, { "epoch": 1.2563365049414958, "grad_norm": 1.8147374391555786, "learning_rate": 8.599254732116694e-05, "loss": 5.8561, "step": 27300 }, { "epoch": 1.2586375507668235, "grad_norm": 1.9511573314666748, "learning_rate": 8.594186408498984e-05, "loss": 5.8655, "step": 27350 }, { "epoch": 1.2609385965921511, "grad_norm": 1.7783595323562622, "learning_rate": 8.58911043151408e-05, "loss": 5.8455, "step": 27400 }, { "epoch": 1.2632396424174788, "grad_norm": 1.8595982789993286, "learning_rate": 8.584026811970637e-05, "loss": 5.8557, "step": 27450 }, { "epoch": 1.2655406882428064, "grad_norm": 1.6024532318115234, "learning_rate": 8.578935560693585e-05, "loss": 5.8431, "step": 27500 }, { "epoch": 1.267841734068134, "grad_norm": 1.6488127708435059, "learning_rate": 8.573836688524099e-05, "loss": 5.8319, "step": 27550 }, { "epoch": 1.2701427798934617, "grad_norm": 1.9752188920974731, "learning_rate": 8.568730206319585e-05, "loss": 5.8757, "step": 27600 }, { "epoch": 1.2724438257187891, "grad_norm": 1.7887974977493286, "learning_rate": 8.563616124953654e-05, "loss": 5.8791, "step": 27650 }, { "epoch": 1.2747448715441168, "grad_norm": 1.8554754257202148, "learning_rate": 8.5584944553161e-05, "loss": 5.9141, "step": 27700 }, { "epoch": 1.2770459173694444, "grad_norm": 2.1256062984466553, "learning_rate": 8.553365208312869e-05, "loss": 5.8593, "step": 27750 }, { "epoch": 1.279346963194772, "grad_norm": 1.798427700996399, "learning_rate": 8.548228394866052e-05, "loss": 5.8325, "step": 27800 }, { "epoch": 1.2816480090200997, "grad_norm": 2.006601095199585, "learning_rate": 8.54308402591384e-05, "loss": 5.8436, "step": 27850 }, { "epoch": 1.2839490548454273, "grad_norm": 1.8625519275665283, "learning_rate": 8.537932112410522e-05, "loss": 5.9255, "step": 27900 }, { "epoch": 1.2862501006707547, "grad_norm": 1.655990481376648, "learning_rate": 8.532772665326452e-05, "loss": 5.821, "step": 27950 }, { "epoch": 1.2885511464960824, "grad_norm": 2.096052408218384, "learning_rate": 8.527605695648016e-05, "loss": 5.9169, "step": 28000 }, { "epoch": 1.2885511464960824, "eval_loss": 5.868961811065674, "eval_runtime": 34.5303, "eval_samples_per_second": 11.121, "eval_steps_per_second": 5.56, "eval_tts_loss": 8.222427133380569, "step": 28000 }, { "epoch": 1.29085219232141, "grad_norm": 1.7661573886871338, "learning_rate": 8.52243121437763e-05, "loss": 5.8294, "step": 28050 }, { "epoch": 1.2931532381467377, "grad_norm": 1.6404305696487427, "learning_rate": 8.517249232533697e-05, "loss": 5.8047, "step": 28100 }, { "epoch": 1.2954542839720653, "grad_norm": 1.677655816078186, "learning_rate": 8.512059761150597e-05, "loss": 5.8431, "step": 28150 }, { "epoch": 1.297755329797393, "grad_norm": 1.4250967502593994, "learning_rate": 8.506862811278655e-05, "loss": 5.8489, "step": 28200 }, { "epoch": 1.3000563756227206, "grad_norm": 1.6657822132110596, "learning_rate": 8.501658393984118e-05, "loss": 5.8159, "step": 28250 }, { "epoch": 1.3023574214480482, "grad_norm": 1.8584389686584473, "learning_rate": 8.496446520349142e-05, "loss": 5.8423, "step": 28300 }, { "epoch": 1.3046584672733759, "grad_norm": 1.83997642993927, "learning_rate": 8.491227201471752e-05, "loss": 5.8523, "step": 28350 }, { "epoch": 1.3069595130987033, "grad_norm": 1.6837650537490845, "learning_rate": 8.48600044846583e-05, "loss": 5.8619, "step": 28400 }, { "epoch": 1.309260558924031, "grad_norm": 1.858853816986084, "learning_rate": 8.480766272461091e-05, "loss": 5.8936, "step": 28450 }, { "epoch": 1.3115616047493586, "grad_norm": 1.9012060165405273, "learning_rate": 8.47552468460305e-05, "loss": 5.8445, "step": 28500 }, { "epoch": 1.3138626505746862, "grad_norm": 1.7135539054870605, "learning_rate": 8.47027569605301e-05, "loss": 5.8654, "step": 28550 }, { "epoch": 1.3161636964000138, "grad_norm": 1.8688374757766724, "learning_rate": 8.465019317988029e-05, "loss": 5.8217, "step": 28600 }, { "epoch": 1.3184647422253415, "grad_norm": 1.7405012845993042, "learning_rate": 8.459755561600906e-05, "loss": 5.9095, "step": 28650 }, { "epoch": 1.320765788050669, "grad_norm": 1.8719056844711304, "learning_rate": 8.454484438100143e-05, "loss": 5.8759, "step": 28700 }, { "epoch": 1.3230668338759966, "grad_norm": 1.9225695133209229, "learning_rate": 8.449205958709936e-05, "loss": 5.8396, "step": 28750 }, { "epoch": 1.3253678797013242, "grad_norm": 2.018826484680176, "learning_rate": 8.44392013467014e-05, "loss": 5.9031, "step": 28800 }, { "epoch": 1.3276689255266518, "grad_norm": 1.7171567678451538, "learning_rate": 8.438626977236253e-05, "loss": 5.889, "step": 28850 }, { "epoch": 1.3299699713519795, "grad_norm": 1.7654656171798706, "learning_rate": 8.433326497679386e-05, "loss": 5.8685, "step": 28900 }, { "epoch": 1.3322710171773071, "grad_norm": 1.6935468912124634, "learning_rate": 8.428018707286242e-05, "loss": 5.7898, "step": 28950 }, { "epoch": 1.3345720630026348, "grad_norm": 1.7551668882369995, "learning_rate": 8.422703617359096e-05, "loss": 5.8487, "step": 29000 }, { "epoch": 1.3345720630026348, "eval_loss": 5.853118419647217, "eval_runtime": 34.41, "eval_samples_per_second": 11.16, "eval_steps_per_second": 5.58, "eval_tts_loss": 8.169616448533585, "step": 29000 }, { "epoch": 1.3368731088279624, "grad_norm": 2.0664875507354736, "learning_rate": 8.417381239215756e-05, "loss": 5.8392, "step": 29050 }, { "epoch": 1.33917415465329, "grad_norm": 1.8029478788375854, "learning_rate": 8.41205158418956e-05, "loss": 5.89, "step": 29100 }, { "epoch": 1.3414752004786175, "grad_norm": 1.7948838472366333, "learning_rate": 8.406714663629337e-05, "loss": 5.8001, "step": 29150 }, { "epoch": 1.343776246303945, "grad_norm": 1.8767132759094238, "learning_rate": 8.401370488899385e-05, "loss": 5.8363, "step": 29200 }, { "epoch": 1.3460772921292727, "grad_norm": 2.1527884006500244, "learning_rate": 8.396019071379453e-05, "loss": 5.8177, "step": 29250 }, { "epoch": 1.3483783379546004, "grad_norm": 1.8454744815826416, "learning_rate": 8.390660422464709e-05, "loss": 5.8218, "step": 29300 }, { "epoch": 1.350679383779928, "grad_norm": 1.9853605031967163, "learning_rate": 8.385294553565723e-05, "loss": 5.8137, "step": 29350 }, { "epoch": 1.3529804296052557, "grad_norm": 1.7211130857467651, "learning_rate": 8.379921476108432e-05, "loss": 5.8233, "step": 29400 }, { "epoch": 1.355281475430583, "grad_norm": 1.9382165670394897, "learning_rate": 8.374541201534132e-05, "loss": 5.8378, "step": 29450 }, { "epoch": 1.3575825212559107, "grad_norm": 1.642452597618103, "learning_rate": 8.36915374129944e-05, "loss": 5.819, "step": 29500 }, { "epoch": 1.3598835670812384, "grad_norm": 1.6594552993774414, "learning_rate": 8.363759106876273e-05, "loss": 5.8647, "step": 29550 }, { "epoch": 1.362184612906566, "grad_norm": 1.715219497680664, "learning_rate": 8.358357309751823e-05, "loss": 5.8451, "step": 29600 }, { "epoch": 1.3644856587318936, "grad_norm": 2.000636100769043, "learning_rate": 8.352948361428538e-05, "loss": 5.8495, "step": 29650 }, { "epoch": 1.3667867045572213, "grad_norm": 1.6294361352920532, "learning_rate": 8.347532273424095e-05, "loss": 5.8231, "step": 29700 }, { "epoch": 1.369087750382549, "grad_norm": 1.901503324508667, "learning_rate": 8.342109057271368e-05, "loss": 5.8407, "step": 29750 }, { "epoch": 1.3713887962078766, "grad_norm": 2.1467387676239014, "learning_rate": 8.336678724518414e-05, "loss": 5.8054, "step": 29800 }, { "epoch": 1.3736898420332042, "grad_norm": 1.8240388631820679, "learning_rate": 8.331241286728442e-05, "loss": 5.8444, "step": 29850 }, { "epoch": 1.3759908878585316, "grad_norm": 1.9498894214630127, "learning_rate": 8.325796755479788e-05, "loss": 5.878, "step": 29900 }, { "epoch": 1.3782919336838593, "grad_norm": 1.9778785705566406, "learning_rate": 8.3204542439554e-05, "loss": 5.8187, "step": 29950 }, { "epoch": 1.380592979509187, "grad_norm": 1.907461404800415, "learning_rate": 8.314995701876072e-05, "loss": 5.7968, "step": 30000 }, { "epoch": 1.380592979509187, "eval_loss": 5.843836307525635, "eval_runtime": 35.0246, "eval_samples_per_second": 10.964, "eval_steps_per_second": 5.482, "eval_tts_loss": 8.287326435213084, "step": 30000 }, { "epoch": 1.3828940253345146, "grad_norm": 1.7704988718032837, "learning_rate": 8.309530100930993e-05, "loss": 5.8421, "step": 30050 }, { "epoch": 1.3851950711598422, "grad_norm": 1.8107035160064697, "learning_rate": 8.304057452758469e-05, "loss": 5.8385, "step": 30100 }, { "epoch": 1.3874961169851698, "grad_norm": 1.9480881690979004, "learning_rate": 8.298577769011817e-05, "loss": 5.7843, "step": 30150 }, { "epoch": 1.3897971628104973, "grad_norm": 1.8153917789459229, "learning_rate": 8.29309106135933e-05, "loss": 5.8488, "step": 30200 }, { "epoch": 1.392098208635825, "grad_norm": 1.7765768766403198, "learning_rate": 8.287597341484265e-05, "loss": 5.7928, "step": 30250 }, { "epoch": 1.3943992544611525, "grad_norm": 1.750103235244751, "learning_rate": 8.282096621084802e-05, "loss": 5.8483, "step": 30300 }, { "epoch": 1.3967003002864802, "grad_norm": 1.730452299118042, "learning_rate": 8.276588911874037e-05, "loss": 5.7923, "step": 30350 }, { "epoch": 1.3990013461118078, "grad_norm": 1.7049480676651, "learning_rate": 8.27107422557994e-05, "loss": 5.8009, "step": 30400 }, { "epoch": 1.4013023919371355, "grad_norm": 1.7733477354049683, "learning_rate": 8.26555257394534e-05, "loss": 5.7999, "step": 30450 }, { "epoch": 1.403603437762463, "grad_norm": 1.7672901153564453, "learning_rate": 8.2600239687279e-05, "loss": 5.8059, "step": 30500 }, { "epoch": 1.4059044835877907, "grad_norm": 1.7486193180084229, "learning_rate": 8.254488421700089e-05, "loss": 5.8793, "step": 30550 }, { "epoch": 1.4082055294131184, "grad_norm": 1.9197908639907837, "learning_rate": 8.248945944649156e-05, "loss": 5.8231, "step": 30600 }, { "epoch": 1.4105065752384458, "grad_norm": 1.6453993320465088, "learning_rate": 8.243396549377106e-05, "loss": 5.9293, "step": 30650 }, { "epoch": 1.4128076210637734, "grad_norm": 1.6542948484420776, "learning_rate": 8.237840247700682e-05, "loss": 5.8378, "step": 30700 }, { "epoch": 1.415108666889101, "grad_norm": 1.7614498138427734, "learning_rate": 8.232277051451323e-05, "loss": 5.7473, "step": 30750 }, { "epoch": 1.4174097127144287, "grad_norm": 1.842435598373413, "learning_rate": 8.226706972475158e-05, "loss": 5.8296, "step": 30800 }, { "epoch": 1.4197107585397564, "grad_norm": 2.2376012802124023, "learning_rate": 8.221130022632968e-05, "loss": 5.9024, "step": 30850 }, { "epoch": 1.422011804365084, "grad_norm": 1.8776355981826782, "learning_rate": 8.215546213800163e-05, "loss": 5.8814, "step": 30900 }, { "epoch": 1.4243128501904114, "grad_norm": 1.7883305549621582, "learning_rate": 8.209955557866764e-05, "loss": 5.7814, "step": 30950 }, { "epoch": 1.426613896015739, "grad_norm": 1.9317328929901123, "learning_rate": 8.204358066737368e-05, "loss": 5.8059, "step": 31000 }, { "epoch": 1.426613896015739, "eval_loss": 5.836706638336182, "eval_runtime": 35.3062, "eval_samples_per_second": 10.876, "eval_steps_per_second": 5.438, "eval_tts_loss": 8.241923826243525, "step": 31000 }, { "epoch": 1.4289149418410667, "grad_norm": 1.7335458993911743, "learning_rate": 8.198753752331125e-05, "loss": 5.863, "step": 31050 }, { "epoch": 1.4312159876663944, "grad_norm": 2.1010098457336426, "learning_rate": 8.193142626581717e-05, "loss": 5.8385, "step": 31100 }, { "epoch": 1.433517033491722, "grad_norm": 1.677432894706726, "learning_rate": 8.18752470143733e-05, "loss": 5.8509, "step": 31150 }, { "epoch": 1.4358180793170496, "grad_norm": 1.8363360166549683, "learning_rate": 8.181899988860629e-05, "loss": 5.8043, "step": 31200 }, { "epoch": 1.4381191251423773, "grad_norm": 1.850675106048584, "learning_rate": 8.176268500828728e-05, "loss": 5.7705, "step": 31250 }, { "epoch": 1.440420170967705, "grad_norm": 1.644543170928955, "learning_rate": 8.170630249333175e-05, "loss": 5.773, "step": 31300 }, { "epoch": 1.4427212167930326, "grad_norm": 2.056574821472168, "learning_rate": 8.164985246379915e-05, "loss": 5.9044, "step": 31350 }, { "epoch": 1.44502226261836, "grad_norm": 1.9397711753845215, "learning_rate": 8.159333503989268e-05, "loss": 5.7941, "step": 31400 }, { "epoch": 1.4473233084436876, "grad_norm": 1.8340749740600586, "learning_rate": 8.153675034195911e-05, "loss": 5.807, "step": 31450 }, { "epoch": 1.4496243542690153, "grad_norm": 1.899632453918457, "learning_rate": 8.148009849048841e-05, "loss": 5.8034, "step": 31500 }, { "epoch": 1.451925400094343, "grad_norm": 1.5245773792266846, "learning_rate": 8.142337960611357e-05, "loss": 5.7948, "step": 31550 }, { "epoch": 1.4542264459196705, "grad_norm": 1.7319273948669434, "learning_rate": 8.136659380961031e-05, "loss": 5.7877, "step": 31600 }, { "epoch": 1.4565274917449982, "grad_norm": 2.0220251083374023, "learning_rate": 8.130974122189681e-05, "loss": 5.7578, "step": 31650 }, { "epoch": 1.4588285375703256, "grad_norm": 1.8379764556884766, "learning_rate": 8.125282196403353e-05, "loss": 5.8232, "step": 31700 }, { "epoch": 1.4611295833956532, "grad_norm": 1.728026032447815, "learning_rate": 8.119583615722279e-05, "loss": 5.8564, "step": 31750 }, { "epoch": 1.4634306292209809, "grad_norm": 1.9760767221450806, "learning_rate": 8.113878392280877e-05, "loss": 5.8495, "step": 31800 }, { "epoch": 1.4657316750463085, "grad_norm": 1.6762609481811523, "learning_rate": 8.108166538227698e-05, "loss": 5.8023, "step": 31850 }, { "epoch": 1.4680327208716362, "grad_norm": 1.9471601247787476, "learning_rate": 8.102448065725414e-05, "loss": 5.8703, "step": 31900 }, { "epoch": 1.4703337666969638, "grad_norm": 1.9490435123443604, "learning_rate": 8.096722986950795e-05, "loss": 5.8204, "step": 31950 }, { "epoch": 1.4726348125222914, "grad_norm": 1.9292274713516235, "learning_rate": 8.090991314094671e-05, "loss": 5.8201, "step": 32000 }, { "epoch": 1.4726348125222914, "eval_loss": 5.825372219085693, "eval_runtime": 32.5118, "eval_samples_per_second": 11.811, "eval_steps_per_second": 5.906, "eval_tts_loss": 8.242703154573366, "step": 32000 }, { "epoch": 1.474935858347619, "grad_norm": 1.584494948387146, "learning_rate": 8.085253059361923e-05, "loss": 5.8162, "step": 32050 }, { "epoch": 1.4772369041729467, "grad_norm": 1.9945138692855835, "learning_rate": 8.079508234971438e-05, "loss": 5.8252, "step": 32100 }, { "epoch": 1.4795379499982741, "grad_norm": 1.899518609046936, "learning_rate": 8.073756853156096e-05, "loss": 5.8475, "step": 32150 }, { "epoch": 1.4818389958236018, "grad_norm": 1.690568208694458, "learning_rate": 8.067998926162739e-05, "loss": 5.7999, "step": 32200 }, { "epoch": 1.4841400416489294, "grad_norm": 2.187692403793335, "learning_rate": 8.062234466252149e-05, "loss": 5.8321, "step": 32250 }, { "epoch": 1.486441087474257, "grad_norm": 1.73610520362854, "learning_rate": 8.056463485699019e-05, "loss": 5.7885, "step": 32300 }, { "epoch": 1.4887421332995847, "grad_norm": 1.7122118473052979, "learning_rate": 8.05068599679192e-05, "loss": 5.771, "step": 32350 }, { "epoch": 1.4910431791249124, "grad_norm": 1.6952186822891235, "learning_rate": 8.044902011833292e-05, "loss": 5.8197, "step": 32400 }, { "epoch": 1.4933442249502398, "grad_norm": 2.2066383361816406, "learning_rate": 8.039111543139399e-05, "loss": 5.7991, "step": 32450 }, { "epoch": 1.4956452707755674, "grad_norm": 1.7441210746765137, "learning_rate": 8.033314603040316e-05, "loss": 5.7646, "step": 32500 }, { "epoch": 1.497946316600895, "grad_norm": 1.9478139877319336, "learning_rate": 8.027511203879896e-05, "loss": 5.8002, "step": 32550 }, { "epoch": 1.5002473624262227, "grad_norm": 1.7180629968643188, "learning_rate": 8.02170135801575e-05, "loss": 5.7598, "step": 32600 }, { "epoch": 1.5025484082515503, "grad_norm": 1.8434754610061646, "learning_rate": 8.015885077819206e-05, "loss": 5.7933, "step": 32650 }, { "epoch": 1.504849454076878, "grad_norm": 1.6318494081497192, "learning_rate": 8.010062375675304e-05, "loss": 5.8359, "step": 32700 }, { "epoch": 1.5071504999022056, "grad_norm": 1.9151034355163574, "learning_rate": 8.004233263982757e-05, "loss": 5.7906, "step": 32750 }, { "epoch": 1.5094515457275333, "grad_norm": 1.8816766738891602, "learning_rate": 7.99839775515392e-05, "loss": 5.758, "step": 32800 }, { "epoch": 1.511752591552861, "grad_norm": 1.998426914215088, "learning_rate": 7.992555861614776e-05, "loss": 5.8639, "step": 32850 }, { "epoch": 1.5140536373781885, "grad_norm": 1.747751235961914, "learning_rate": 7.986707595804901e-05, "loss": 5.82, "step": 32900 }, { "epoch": 1.516354683203516, "grad_norm": 1.8607374429702759, "learning_rate": 7.980852970177439e-05, "loss": 5.7843, "step": 32950 }, { "epoch": 1.5186557290288436, "grad_norm": 2.160118818283081, "learning_rate": 7.97499199719908e-05, "loss": 5.76, "step": 33000 }, { "epoch": 1.5186557290288436, "eval_loss": 5.812412738800049, "eval_runtime": 33.9931, "eval_samples_per_second": 11.296, "eval_steps_per_second": 5.648, "eval_tts_loss": 8.298841324188611, "step": 33000 }, { "epoch": 1.5209567748541712, "grad_norm": 1.9013315439224243, "learning_rate": 7.969124689350027e-05, "loss": 5.7762, "step": 33050 }, { "epoch": 1.5232578206794989, "grad_norm": 2.26086163520813, "learning_rate": 7.963251059123971e-05, "loss": 5.7748, "step": 33100 }, { "epoch": 1.5255588665048263, "grad_norm": 1.5581964254379272, "learning_rate": 7.957371119028072e-05, "loss": 5.785, "step": 33150 }, { "epoch": 1.527859912330154, "grad_norm": 1.8346717357635498, "learning_rate": 7.951484881582917e-05, "loss": 5.8623, "step": 33200 }, { "epoch": 1.5301609581554816, "grad_norm": 1.89267897605896, "learning_rate": 7.945592359322508e-05, "loss": 5.8608, "step": 33250 }, { "epoch": 1.5324620039808092, "grad_norm": 1.7842216491699219, "learning_rate": 7.939693564794233e-05, "loss": 5.7873, "step": 33300 }, { "epoch": 1.5347630498061369, "grad_norm": 1.5716899633407593, "learning_rate": 7.933906672907357e-05, "loss": 5.7573, "step": 33350 }, { "epoch": 1.5370640956314645, "grad_norm": 2.020463466644287, "learning_rate": 7.92799549635824e-05, "loss": 5.7473, "step": 33400 }, { "epoch": 1.5393651414567922, "grad_norm": 1.8860468864440918, "learning_rate": 7.922078085011557e-05, "loss": 5.7842, "step": 33450 }, { "epoch": 1.5416661872821198, "grad_norm": 1.9550951719284058, "learning_rate": 7.916154451467692e-05, "loss": 5.7405, "step": 33500 }, { "epoch": 1.5439672331074474, "grad_norm": 1.8158951997756958, "learning_rate": 7.910224608340272e-05, "loss": 5.7826, "step": 33550 }, { "epoch": 1.546268278932775, "grad_norm": 1.9261797666549683, "learning_rate": 7.904288568256155e-05, "loss": 5.7983, "step": 33600 }, { "epoch": 1.5485693247581027, "grad_norm": 1.7144373655319214, "learning_rate": 7.898346343855388e-05, "loss": 5.7314, "step": 33650 }, { "epoch": 1.5508703705834301, "grad_norm": 1.9878826141357422, "learning_rate": 7.892397947791191e-05, "loss": 5.7692, "step": 33700 }, { "epoch": 1.5531714164087578, "grad_norm": 1.7931249141693115, "learning_rate": 7.886443392729923e-05, "loss": 5.7466, "step": 33750 }, { "epoch": 1.5554724622340854, "grad_norm": 2.2137207984924316, "learning_rate": 7.880482691351059e-05, "loss": 5.7228, "step": 33800 }, { "epoch": 1.5577735080594128, "grad_norm": 1.9212007522583008, "learning_rate": 7.874515856347162e-05, "loss": 5.8418, "step": 33850 }, { "epoch": 1.5600745538847405, "grad_norm": 1.9778438806533813, "learning_rate": 7.868542900423854e-05, "loss": 5.739, "step": 33900 }, { "epoch": 1.5623755997100681, "grad_norm": 1.924148678779602, "learning_rate": 7.862563836299796e-05, "loss": 5.8036, "step": 33950 }, { "epoch": 1.5646766455353958, "grad_norm": 1.5860093832015991, "learning_rate": 7.856578676706649e-05, "loss": 5.7642, "step": 34000 }, { "epoch": 1.5646766455353958, "eval_loss": 5.801419734954834, "eval_runtime": 35.559, "eval_samples_per_second": 10.799, "eval_steps_per_second": 5.399, "eval_tts_loss": 8.180454099509982, "step": 34000 }, { "epoch": 1.5669776913607234, "grad_norm": 1.8571122884750366, "learning_rate": 7.850587434389056e-05, "loss": 5.7714, "step": 34050 }, { "epoch": 1.569278737186051, "grad_norm": 2.0070884227752686, "learning_rate": 7.844590122104614e-05, "loss": 5.778, "step": 34100 }, { "epoch": 1.5715797830113787, "grad_norm": 1.6417818069458008, "learning_rate": 7.838586752623844e-05, "loss": 5.796, "step": 34150 }, { "epoch": 1.5738808288367063, "grad_norm": 2.259556531906128, "learning_rate": 7.83269758616054e-05, "loss": 5.7928, "step": 34200 }, { "epoch": 1.576181874662034, "grad_norm": 1.756239414215088, "learning_rate": 7.826682261157087e-05, "loss": 5.8136, "step": 34250 }, { "epoch": 1.5784829204873616, "grad_norm": 1.6941827535629272, "learning_rate": 7.820660917089842e-05, "loss": 5.7911, "step": 34300 }, { "epoch": 1.5807839663126892, "grad_norm": 1.8113830089569092, "learning_rate": 7.814633566780493e-05, "loss": 5.843, "step": 34350 }, { "epoch": 1.583085012138017, "grad_norm": 1.920233130455017, "learning_rate": 7.80860022306353e-05, "loss": 5.8077, "step": 34400 }, { "epoch": 1.5853860579633443, "grad_norm": 1.5396602153778076, "learning_rate": 7.802560898786191e-05, "loss": 5.8207, "step": 34450 }, { "epoch": 1.587687103788672, "grad_norm": 1.6108042001724243, "learning_rate": 7.796515606808463e-05, "loss": 5.7094, "step": 34500 }, { "epoch": 1.5899881496139996, "grad_norm": 1.8248738050460815, "learning_rate": 7.79046436000303e-05, "loss": 5.8258, "step": 34550 }, { "epoch": 1.592289195439327, "grad_norm": 2.126722812652588, "learning_rate": 7.784407171255261e-05, "loss": 5.7698, "step": 34600 }, { "epoch": 1.5945902412646547, "grad_norm": 1.9345743656158447, "learning_rate": 7.778344053463177e-05, "loss": 5.8193, "step": 34650 }, { "epoch": 1.5968912870899823, "grad_norm": 1.651314377784729, "learning_rate": 7.772275019537423e-05, "loss": 5.7444, "step": 34700 }, { "epoch": 1.59919233291531, "grad_norm": 1.9649804830551147, "learning_rate": 7.766200082401241e-05, "loss": 5.7637, "step": 34750 }, { "epoch": 1.6014933787406376, "grad_norm": 1.9861186742782593, "learning_rate": 7.760119254990448e-05, "loss": 5.8492, "step": 34800 }, { "epoch": 1.6037944245659652, "grad_norm": 1.8200690746307373, "learning_rate": 7.754032550253394e-05, "loss": 5.7746, "step": 34850 }, { "epoch": 1.6060954703912929, "grad_norm": 1.9275963306427002, "learning_rate": 7.747939981150955e-05, "loss": 5.7782, "step": 34900 }, { "epoch": 1.6083965162166205, "grad_norm": 2.0152270793914795, "learning_rate": 7.741841560656487e-05, "loss": 5.8126, "step": 34950 }, { "epoch": 1.6106975620419481, "grad_norm": 2.15605092048645, "learning_rate": 7.73573730175581e-05, "loss": 5.7019, "step": 35000 }, { "epoch": 1.6106975620419481, "eval_loss": 5.787054538726807, "eval_runtime": 33.4679, "eval_samples_per_second": 11.474, "eval_steps_per_second": 5.737, "eval_tts_loss": 8.289476949391128, "step": 35000 }, { "epoch": 1.6129986078672758, "grad_norm": 1.8516534566879272, "learning_rate": 7.729627217447174e-05, "loss": 5.7534, "step": 35050 }, { "epoch": 1.6152996536926034, "grad_norm": 1.9034392833709717, "learning_rate": 7.72351132074123e-05, "loss": 5.7853, "step": 35100 }, { "epoch": 1.617600699517931, "grad_norm": 4.265834808349609, "learning_rate": 7.717389624661014e-05, "loss": 5.7785, "step": 35150 }, { "epoch": 1.6199017453432585, "grad_norm": 1.859275460243225, "learning_rate": 7.711262142241906e-05, "loss": 5.7689, "step": 35200 }, { "epoch": 1.6222027911685861, "grad_norm": 1.9681501388549805, "learning_rate": 7.705128886531607e-05, "loss": 5.7189, "step": 35250 }, { "epoch": 1.6245038369939138, "grad_norm": 2.0118014812469482, "learning_rate": 7.698989870590114e-05, "loss": 5.8077, "step": 35300 }, { "epoch": 1.6268048828192412, "grad_norm": 1.8570785522460938, "learning_rate": 7.692845107489688e-05, "loss": 5.7515, "step": 35350 }, { "epoch": 1.6291059286445688, "grad_norm": 1.7671388387680054, "learning_rate": 7.686694610314824e-05, "loss": 5.7661, "step": 35400 }, { "epoch": 1.6314069744698965, "grad_norm": 1.5981802940368652, "learning_rate": 7.680538392162235e-05, "loss": 5.7891, "step": 35450 }, { "epoch": 1.633708020295224, "grad_norm": 1.7688723802566528, "learning_rate": 7.67437646614081e-05, "loss": 5.7618, "step": 35500 }, { "epoch": 1.6360090661205517, "grad_norm": 1.6916351318359375, "learning_rate": 7.668208845371596e-05, "loss": 5.7752, "step": 35550 }, { "epoch": 1.6383101119458794, "grad_norm": 1.6966232061386108, "learning_rate": 7.662035542987765e-05, "loss": 5.7361, "step": 35600 }, { "epoch": 1.640611157771207, "grad_norm": 1.9174034595489502, "learning_rate": 7.655856572134583e-05, "loss": 5.6983, "step": 35650 }, { "epoch": 1.6429122035965347, "grad_norm": 1.7266641855239868, "learning_rate": 7.649671945969392e-05, "loss": 5.7842, "step": 35700 }, { "epoch": 1.6452132494218623, "grad_norm": 2.1633241176605225, "learning_rate": 7.643481677661576e-05, "loss": 5.7123, "step": 35750 }, { "epoch": 1.64751429524719, "grad_norm": 1.7543927431106567, "learning_rate": 7.63728578039253e-05, "loss": 5.7678, "step": 35800 }, { "epoch": 1.6498153410725176, "grad_norm": 1.7160446643829346, "learning_rate": 7.631084267355637e-05, "loss": 5.7955, "step": 35850 }, { "epoch": 1.6521163868978452, "grad_norm": 1.841979742050171, "learning_rate": 7.624877151756238e-05, "loss": 5.7917, "step": 35900 }, { "epoch": 1.6544174327231727, "grad_norm": 1.9420545101165771, "learning_rate": 7.618664446811603e-05, "loss": 5.7249, "step": 35950 }, { "epoch": 1.6567184785485003, "grad_norm": 1.8593506813049316, "learning_rate": 7.612446165750906e-05, "loss": 5.8221, "step": 36000 }, { "epoch": 1.6567184785485003, "eval_loss": 5.778265476226807, "eval_runtime": 33.0355, "eval_samples_per_second": 11.624, "eval_steps_per_second": 5.812, "eval_tts_loss": 8.30205424102598, "step": 36000 }, { "epoch": 1.659019524373828, "grad_norm": 1.7334771156311035, "learning_rate": 7.606222321815193e-05, "loss": 5.7638, "step": 36050 }, { "epoch": 1.6613205701991554, "grad_norm": 2.0367279052734375, "learning_rate": 7.599992928257353e-05, "loss": 5.7243, "step": 36100 }, { "epoch": 1.663621616024483, "grad_norm": 2.257939577102661, "learning_rate": 7.593757998342096e-05, "loss": 5.8182, "step": 36150 }, { "epoch": 1.6659226618498106, "grad_norm": 1.722194790840149, "learning_rate": 7.587517545345923e-05, "loss": 5.8015, "step": 36200 }, { "epoch": 1.6682237076751383, "grad_norm": 1.7900599241256714, "learning_rate": 7.58127158255709e-05, "loss": 5.752, "step": 36250 }, { "epoch": 1.670524753500466, "grad_norm": 1.823088526725769, "learning_rate": 7.57502012327559e-05, "loss": 5.6804, "step": 36300 }, { "epoch": 1.6728257993257936, "grad_norm": 1.7726881504058838, "learning_rate": 7.568763180813112e-05, "loss": 5.7921, "step": 36350 }, { "epoch": 1.6751268451511212, "grad_norm": 1.8228272199630737, "learning_rate": 7.562500768493032e-05, "loss": 5.8114, "step": 36400 }, { "epoch": 1.6774278909764488, "grad_norm": 1.6336501836776733, "learning_rate": 7.556232899650371e-05, "loss": 5.8347, "step": 36450 }, { "epoch": 1.6797289368017765, "grad_norm": 1.5864665508270264, "learning_rate": 7.549959587631758e-05, "loss": 5.7587, "step": 36500 }, { "epoch": 1.6820299826271041, "grad_norm": 1.659406304359436, "learning_rate": 7.543680845795427e-05, "loss": 5.8137, "step": 36550 }, { "epoch": 1.6843310284524318, "grad_norm": 1.773979902267456, "learning_rate": 7.537396687511163e-05, "loss": 5.8097, "step": 36600 }, { "epoch": 1.6866320742777594, "grad_norm": 1.7780120372772217, "learning_rate": 7.531107126160295e-05, "loss": 5.7593, "step": 36650 }, { "epoch": 1.6889331201030868, "grad_norm": 1.6934456825256348, "learning_rate": 7.524812175135647e-05, "loss": 5.7162, "step": 36700 }, { "epoch": 1.6912341659284145, "grad_norm": 1.9389722347259521, "learning_rate": 7.518511847841528e-05, "loss": 5.7942, "step": 36750 }, { "epoch": 1.693535211753742, "grad_norm": 1.7949825525283813, "learning_rate": 7.512206157693686e-05, "loss": 5.6908, "step": 36800 }, { "epoch": 1.6958362575790695, "grad_norm": 1.7409274578094482, "learning_rate": 7.5058951181193e-05, "loss": 5.73, "step": 36850 }, { "epoch": 1.6981373034043972, "grad_norm": 1.825876235961914, "learning_rate": 7.49957874255693e-05, "loss": 5.7233, "step": 36900 }, { "epoch": 1.7004383492297248, "grad_norm": 1.8489861488342285, "learning_rate": 7.493257044456502e-05, "loss": 5.8163, "step": 36950 }, { "epoch": 1.7027393950550525, "grad_norm": 1.8161964416503906, "learning_rate": 7.486930037279278e-05, "loss": 5.7279, "step": 37000 }, { "epoch": 1.7027393950550525, "eval_loss": 5.769525051116943, "eval_runtime": 34.5593, "eval_samples_per_second": 11.111, "eval_steps_per_second": 5.556, "eval_tts_loss": 8.327279314398375, "step": 37000 }, { "epoch": 1.70504044088038, "grad_norm": 1.8013944625854492, "learning_rate": 7.480597734497821e-05, "loss": 5.7584, "step": 37050 }, { "epoch": 1.7073414867057077, "grad_norm": 2.0919711589813232, "learning_rate": 7.474260149595973e-05, "loss": 5.7107, "step": 37100 }, { "epoch": 1.7096425325310354, "grad_norm": 1.646823763847351, "learning_rate": 7.467917296068822e-05, "loss": 5.7999, "step": 37150 }, { "epoch": 1.711943578356363, "grad_norm": 1.934715986251831, "learning_rate": 7.461569187422675e-05, "loss": 5.7851, "step": 37200 }, { "epoch": 1.7142446241816907, "grad_norm": 1.8576382398605347, "learning_rate": 7.455215837175031e-05, "loss": 5.7287, "step": 37250 }, { "epoch": 1.7165456700070183, "grad_norm": 1.8678678274154663, "learning_rate": 7.448857258854549e-05, "loss": 5.7543, "step": 37300 }, { "epoch": 1.718846715832346, "grad_norm": 2.078073501586914, "learning_rate": 7.442493466001019e-05, "loss": 5.7417, "step": 37350 }, { "epoch": 1.7211477616576734, "grad_norm": 1.7547941207885742, "learning_rate": 7.436124472165338e-05, "loss": 5.7377, "step": 37400 }, { "epoch": 1.723448807483001, "grad_norm": 1.694883942604065, "learning_rate": 7.429750290909476e-05, "loss": 5.7243, "step": 37450 }, { "epoch": 1.7257498533083286, "grad_norm": 1.8052371740341187, "learning_rate": 7.423370935806447e-05, "loss": 5.7679, "step": 37500 }, { "epoch": 1.7280508991336563, "grad_norm": 1.943966269493103, "learning_rate": 7.416986420440288e-05, "loss": 5.748, "step": 37550 }, { "epoch": 1.7303519449589837, "grad_norm": 1.619152307510376, "learning_rate": 7.410596758406015e-05, "loss": 5.7693, "step": 37600 }, { "epoch": 1.7326529907843113, "grad_norm": 1.9282474517822266, "learning_rate": 7.404201963309611e-05, "loss": 5.7253, "step": 37650 }, { "epoch": 1.734954036609639, "grad_norm": 2.065870523452759, "learning_rate": 7.397802048767987e-05, "loss": 5.7773, "step": 37700 }, { "epoch": 1.7372550824349666, "grad_norm": 2.3351128101348877, "learning_rate": 7.391397028408956e-05, "loss": 5.7022, "step": 37750 }, { "epoch": 1.7395561282602943, "grad_norm": 1.6631519794464111, "learning_rate": 7.384986915871199e-05, "loss": 5.7305, "step": 37800 }, { "epoch": 1.741857174085622, "grad_norm": 1.9036861658096313, "learning_rate": 7.378571724804247e-05, "loss": 5.7526, "step": 37850 }, { "epoch": 1.7441582199109495, "grad_norm": 1.6608564853668213, "learning_rate": 7.372151468868434e-05, "loss": 5.7262, "step": 37900 }, { "epoch": 1.7464592657362772, "grad_norm": 2.1009271144866943, "learning_rate": 7.365726161734895e-05, "loss": 5.7258, "step": 37950 }, { "epoch": 1.7487603115616048, "grad_norm": 2.0036394596099854, "learning_rate": 7.359295817085507e-05, "loss": 5.7435, "step": 38000 }, { "epoch": 1.7487603115616048, "eval_loss": 5.767555236816406, "eval_runtime": 35.2579, "eval_samples_per_second": 10.891, "eval_steps_per_second": 5.446, "eval_tts_loss": 8.310344913623878, "step": 38000 }, { "epoch": 1.7510613573869325, "grad_norm": 1.9608486890792847, "learning_rate": 7.35286044861288e-05, "loss": 5.7333, "step": 38050 }, { "epoch": 1.7533624032122601, "grad_norm": 1.794594168663025, "learning_rate": 7.346420070020322e-05, "loss": 5.7214, "step": 38100 }, { "epoch": 1.7556634490375875, "grad_norm": 1.8481963872909546, "learning_rate": 7.339974695021805e-05, "loss": 5.707, "step": 38150 }, { "epoch": 1.7579644948629152, "grad_norm": 1.9794644117355347, "learning_rate": 7.333524337341946e-05, "loss": 5.6877, "step": 38200 }, { "epoch": 1.7602655406882428, "grad_norm": 1.830579161643982, "learning_rate": 7.327069010715966e-05, "loss": 5.784, "step": 38250 }, { "epoch": 1.7625665865135705, "grad_norm": 2.2297370433807373, "learning_rate": 7.320608728889674e-05, "loss": 5.7061, "step": 38300 }, { "epoch": 1.7648676323388979, "grad_norm": 1.972904920578003, "learning_rate": 7.314143505619419e-05, "loss": 5.7295, "step": 38350 }, { "epoch": 1.7671686781642255, "grad_norm": 1.9099104404449463, "learning_rate": 7.307673354672089e-05, "loss": 5.7593, "step": 38400 }, { "epoch": 1.7694697239895532, "grad_norm": 1.6860655546188354, "learning_rate": 7.30119828982505e-05, "loss": 5.798, "step": 38450 }, { "epoch": 1.7717707698148808, "grad_norm": 1.81305992603302, "learning_rate": 7.294718324866138e-05, "loss": 5.7393, "step": 38500 }, { "epoch": 1.7740718156402084, "grad_norm": 1.9312961101531982, "learning_rate": 7.28823347359362e-05, "loss": 5.721, "step": 38550 }, { "epoch": 1.776372861465536, "grad_norm": 1.8625140190124512, "learning_rate": 7.281743749816174e-05, "loss": 5.7543, "step": 38600 }, { "epoch": 1.7786739072908637, "grad_norm": 1.628096342086792, "learning_rate": 7.275249167352847e-05, "loss": 5.7525, "step": 38650 }, { "epoch": 1.7809749531161914, "grad_norm": 1.6411455869674683, "learning_rate": 7.268749740033035e-05, "loss": 5.7497, "step": 38700 }, { "epoch": 1.783275998941519, "grad_norm": 2.187689781188965, "learning_rate": 7.262245481696446e-05, "loss": 5.7272, "step": 38750 }, { "epoch": 1.7855770447668466, "grad_norm": 1.7856398820877075, "learning_rate": 7.255736406193083e-05, "loss": 5.7525, "step": 38800 }, { "epoch": 1.7878780905921743, "grad_norm": 1.9144530296325684, "learning_rate": 7.249222527383201e-05, "loss": 5.7671, "step": 38850 }, { "epoch": 1.7901791364175017, "grad_norm": 1.8890655040740967, "learning_rate": 7.242703859137286e-05, "loss": 5.7387, "step": 38900 }, { "epoch": 1.7924801822428293, "grad_norm": 1.8778235912322998, "learning_rate": 7.236180415336017e-05, "loss": 5.748, "step": 38950 }, { "epoch": 1.794781228068157, "grad_norm": 2.0300331115722656, "learning_rate": 7.229652209870249e-05, "loss": 5.7996, "step": 39000 }, { "epoch": 1.794781228068157, "eval_loss": 5.750644207000732, "eval_runtime": 34.215, "eval_samples_per_second": 11.223, "eval_steps_per_second": 5.612, "eval_tts_loss": 8.316885493755828, "step": 39000 }, { "epoch": 1.7970822738934846, "grad_norm": 1.9163450002670288, "learning_rate": 7.223119256640974e-05, "loss": 5.7187, "step": 39050 }, { "epoch": 1.799383319718812, "grad_norm": 1.8101003170013428, "learning_rate": 7.216581569559288e-05, "loss": 5.7177, "step": 39100 }, { "epoch": 1.8016843655441397, "grad_norm": 1.856192946434021, "learning_rate": 7.210039162546378e-05, "loss": 5.7414, "step": 39150 }, { "epoch": 1.8039854113694673, "grad_norm": 2.071753740310669, "learning_rate": 7.203492049533472e-05, "loss": 5.7616, "step": 39200 }, { "epoch": 1.806286457194795, "grad_norm": 2.085846185684204, "learning_rate": 7.196940244461824e-05, "loss": 5.6907, "step": 39250 }, { "epoch": 1.8085875030201226, "grad_norm": 1.751923680305481, "learning_rate": 7.190383761282677e-05, "loss": 5.7245, "step": 39300 }, { "epoch": 1.8108885488454503, "grad_norm": 2.07672119140625, "learning_rate": 7.183822613957234e-05, "loss": 5.7792, "step": 39350 }, { "epoch": 1.813189594670778, "grad_norm": 2.1125259399414062, "learning_rate": 7.177256816456637e-05, "loss": 5.7704, "step": 39400 }, { "epoch": 1.8154906404961055, "grad_norm": 1.7339011430740356, "learning_rate": 7.170686382761917e-05, "loss": 5.7657, "step": 39450 }, { "epoch": 1.8177916863214332, "grad_norm": 1.9891151189804077, "learning_rate": 7.164111326863991e-05, "loss": 5.7256, "step": 39500 }, { "epoch": 1.8200927321467608, "grad_norm": 1.7742396593093872, "learning_rate": 7.157531662763612e-05, "loss": 5.7252, "step": 39550 }, { "epoch": 1.8223937779720885, "grad_norm": 2.2981250286102295, "learning_rate": 7.150947404471342e-05, "loss": 5.8321, "step": 39600 }, { "epoch": 1.8246948237974159, "grad_norm": 1.7247025966644287, "learning_rate": 7.144358566007531e-05, "loss": 5.7159, "step": 39650 }, { "epoch": 1.8269958696227435, "grad_norm": 2.1468214988708496, "learning_rate": 7.137765161402285e-05, "loss": 5.6725, "step": 39700 }, { "epoch": 1.8292969154480712, "grad_norm": 1.9942362308502197, "learning_rate": 7.131167204695424e-05, "loss": 5.6895, "step": 39750 }, { "epoch": 1.8315979612733988, "grad_norm": 1.7421067953109741, "learning_rate": 7.124564709936465e-05, "loss": 5.7035, "step": 39800 }, { "epoch": 1.8338990070987262, "grad_norm": 1.6837613582611084, "learning_rate": 7.11795769118459e-05, "loss": 5.691, "step": 39850 }, { "epoch": 1.8362000529240539, "grad_norm": 1.8079731464385986, "learning_rate": 7.111346162508615e-05, "loss": 5.7645, "step": 39900 }, { "epoch": 1.8385010987493815, "grad_norm": 1.8383841514587402, "learning_rate": 7.104730137986956e-05, "loss": 5.8337, "step": 39950 }, { "epoch": 1.8408021445747091, "grad_norm": 1.6859551668167114, "learning_rate": 7.098109631707604e-05, "loss": 5.6273, "step": 40000 }, { "epoch": 1.8408021445747091, "eval_loss": 5.748322010040283, "eval_runtime": 34.6862, "eval_samples_per_second": 11.071, "eval_steps_per_second": 5.535, "eval_tts_loss": 8.36463535504836, "step": 40000 }, { "epoch": 1.8431031904000368, "grad_norm": 1.8230650424957275, "learning_rate": 7.091484657768092e-05, "loss": 5.6678, "step": 40050 }, { "epoch": 1.8454042362253644, "grad_norm": 2.0616815090179443, "learning_rate": 7.084855230275466e-05, "loss": 5.7542, "step": 40100 }, { "epoch": 1.847705282050692, "grad_norm": 1.8527641296386719, "learning_rate": 7.078221363346263e-05, "loss": 5.6728, "step": 40150 }, { "epoch": 1.8500063278760197, "grad_norm": 1.6423956155776978, "learning_rate": 7.071583071106462e-05, "loss": 5.7776, "step": 40200 }, { "epoch": 1.8523073737013473, "grad_norm": 2.137951374053955, "learning_rate": 7.064940367691471e-05, "loss": 5.7493, "step": 40250 }, { "epoch": 1.854608419526675, "grad_norm": 2.6266961097717285, "learning_rate": 7.05829326724609e-05, "loss": 5.7793, "step": 40300 }, { "epoch": 1.8569094653520026, "grad_norm": 2.0090668201446533, "learning_rate": 7.051641783924483e-05, "loss": 5.6379, "step": 40350 }, { "epoch": 1.85921051117733, "grad_norm": 1.942234754562378, "learning_rate": 7.044985931890144e-05, "loss": 5.805, "step": 40400 }, { "epoch": 1.8615115570026577, "grad_norm": 1.8053412437438965, "learning_rate": 7.038325725315874e-05, "loss": 5.7068, "step": 40450 }, { "epoch": 1.8638126028279853, "grad_norm": 2.0646331310272217, "learning_rate": 7.031661178383741e-05, "loss": 5.6697, "step": 40500 }, { "epoch": 1.866113648653313, "grad_norm": 1.813453197479248, "learning_rate": 7.025125725051633e-05, "loss": 5.6648, "step": 40550 }, { "epoch": 1.8684146944786404, "grad_norm": 1.700060248374939, "learning_rate": 7.018586130234475e-05, "loss": 5.7569, "step": 40600 }, { "epoch": 1.870715740303968, "grad_norm": 2.1247808933258057, "learning_rate": 7.011908819050767e-05, "loss": 5.6719, "step": 40650 }, { "epoch": 1.8730167861292957, "grad_norm": 2.1463191509246826, "learning_rate": 7.005227223760671e-05, "loss": 5.761, "step": 40700 }, { "epoch": 1.8753178319546233, "grad_norm": 1.92280912399292, "learning_rate": 6.998541358591804e-05, "loss": 5.633, "step": 40750 }, { "epoch": 1.877618877779951, "grad_norm": 1.797415018081665, "learning_rate": 6.991851237780871e-05, "loss": 5.7209, "step": 40800 }, { "epoch": 1.8799199236052786, "grad_norm": 1.9149625301361084, "learning_rate": 6.985156875573646e-05, "loss": 5.7149, "step": 40850 }, { "epoch": 1.8822209694306062, "grad_norm": 2.1698825359344482, "learning_rate": 6.978458286224925e-05, "loss": 5.7347, "step": 40900 }, { "epoch": 1.8845220152559339, "grad_norm": 2.2453010082244873, "learning_rate": 6.97188958123694e-05, "loss": 5.6696, "step": 40950 }, { "epoch": 1.8868230610812615, "grad_norm": 1.8774462938308716, "learning_rate": 6.965182664237777e-05, "loss": 5.6944, "step": 41000 }, { "epoch": 1.8868230610812615, "eval_loss": 5.7265706062316895, "eval_runtime": 35.1085, "eval_samples_per_second": 10.938, "eval_steps_per_second": 5.469, "eval_tts_loss": 8.331914464921848, "step": 41000 }, { "epoch": 1.8891241069065892, "grad_norm": 2.033698797225952, "learning_rate": 6.958471562629687e-05, "loss": 5.7328, "step": 41050 }, { "epoch": 1.8914251527319168, "grad_norm": 1.8300659656524658, "learning_rate": 6.951756290703115e-05, "loss": 5.7555, "step": 41100 }, { "epoch": 1.8937261985572442, "grad_norm": 1.8035775423049927, "learning_rate": 6.945036862757388e-05, "loss": 5.7462, "step": 41150 }, { "epoch": 1.8960272443825719, "grad_norm": 2.0775158405303955, "learning_rate": 6.938313293100682e-05, "loss": 5.6722, "step": 41200 }, { "epoch": 1.8983282902078995, "grad_norm": 1.7843129634857178, "learning_rate": 6.931585596049988e-05, "loss": 5.6753, "step": 41250 }, { "epoch": 1.9006293360332271, "grad_norm": 2.2548415660858154, "learning_rate": 6.924853785931092e-05, "loss": 5.7172, "step": 41300 }, { "epoch": 1.9029303818585546, "grad_norm": 1.7711260318756104, "learning_rate": 6.918117877078535e-05, "loss": 5.75, "step": 41350 }, { "epoch": 1.9052314276838822, "grad_norm": 1.7503825426101685, "learning_rate": 6.911377883835588e-05, "loss": 5.6779, "step": 41400 }, { "epoch": 1.9075324735092098, "grad_norm": 1.8140344619750977, "learning_rate": 6.904633820554216e-05, "loss": 5.6786, "step": 41450 }, { "epoch": 1.9098335193345375, "grad_norm": 1.9237339496612549, "learning_rate": 6.897885701595052e-05, "loss": 5.6965, "step": 41500 }, { "epoch": 1.9121345651598651, "grad_norm": 2.2851338386535645, "learning_rate": 6.891133541327363e-05, "loss": 5.6899, "step": 41550 }, { "epoch": 1.9144356109851928, "grad_norm": 1.6410493850708008, "learning_rate": 6.884377354129029e-05, "loss": 5.7291, "step": 41600 }, { "epoch": 1.9167366568105204, "grad_norm": 1.809895634651184, "learning_rate": 6.877617154386498e-05, "loss": 5.7246, "step": 41650 }, { "epoch": 1.919037702635848, "grad_norm": 2.089245557785034, "learning_rate": 6.87085295649476e-05, "loss": 5.7542, "step": 41700 }, { "epoch": 1.9213387484611757, "grad_norm": 2.0820837020874023, "learning_rate": 6.864084774857325e-05, "loss": 5.7559, "step": 41750 }, { "epoch": 1.9236397942865033, "grad_norm": 1.8039745092391968, "learning_rate": 6.857312623886185e-05, "loss": 5.7623, "step": 41800 }, { "epoch": 1.925940840111831, "grad_norm": 2.192228317260742, "learning_rate": 6.85053651800178e-05, "loss": 5.7104, "step": 41850 }, { "epoch": 1.9282418859371584, "grad_norm": 2.2128982543945312, "learning_rate": 6.843756471632974e-05, "loss": 5.7603, "step": 41900 }, { "epoch": 1.930542931762486, "grad_norm": 1.7217659950256348, "learning_rate": 6.836972499217023e-05, "loss": 5.7817, "step": 41950 }, { "epoch": 1.9328439775878137, "grad_norm": 1.9377886056900024, "learning_rate": 6.83018461519954e-05, "loss": 5.6959, "step": 42000 }, { "epoch": 1.9328439775878137, "eval_loss": 5.725087642669678, "eval_runtime": 34.6969, "eval_samples_per_second": 11.067, "eval_steps_per_second": 5.534, "eval_tts_loss": 8.384733506297332, "step": 42000 }, { "epoch": 1.9351450234131413, "grad_norm": 1.861262559890747, "learning_rate": 6.82339283403447e-05, "loss": 5.7196, "step": 42050 }, { "epoch": 1.9374460692384687, "grad_norm": 1.799297571182251, "learning_rate": 6.816597170184054e-05, "loss": 5.6952, "step": 42100 }, { "epoch": 1.9397471150637964, "grad_norm": 2.116269588470459, "learning_rate": 6.809797638118805e-05, "loss": 5.7379, "step": 42150 }, { "epoch": 1.942048160889124, "grad_norm": 1.9920415878295898, "learning_rate": 6.802994252317467e-05, "loss": 5.5964, "step": 42200 }, { "epoch": 1.9443492067144517, "grad_norm": 1.7967939376831055, "learning_rate": 6.796187027266995e-05, "loss": 5.6312, "step": 42250 }, { "epoch": 1.9466502525397793, "grad_norm": 2.3710741996765137, "learning_rate": 6.789375977462515e-05, "loss": 5.6975, "step": 42300 }, { "epoch": 1.948951298365107, "grad_norm": 2.0796024799346924, "learning_rate": 6.782561117407303e-05, "loss": 5.7446, "step": 42350 }, { "epoch": 1.9512523441904346, "grad_norm": 2.202712297439575, "learning_rate": 6.775742461612742e-05, "loss": 5.7564, "step": 42400 }, { "epoch": 1.9535533900157622, "grad_norm": 1.7688695192337036, "learning_rate": 6.768920024598303e-05, "loss": 5.7891, "step": 42450 }, { "epoch": 1.9558544358410899, "grad_norm": 2.2628657817840576, "learning_rate": 6.762093820891502e-05, "loss": 5.6503, "step": 42500 }, { "epoch": 1.9581554816664175, "grad_norm": 2.0447051525115967, "learning_rate": 6.755263865027885e-05, "loss": 5.6383, "step": 42550 }, { "epoch": 1.9604565274917451, "grad_norm": 1.6910145282745361, "learning_rate": 6.748430171550979e-05, "loss": 5.6793, "step": 42600 }, { "epoch": 1.9627575733170726, "grad_norm": 1.6950970888137817, "learning_rate": 6.741592755012276e-05, "loss": 5.7613, "step": 42650 }, { "epoch": 1.9650586191424002, "grad_norm": 2.0867919921875, "learning_rate": 6.734751629971192e-05, "loss": 5.7127, "step": 42700 }, { "epoch": 1.9673596649677279, "grad_norm": 2.344790458679199, "learning_rate": 6.727906810995041e-05, "loss": 5.7208, "step": 42750 }, { "epoch": 1.9696607107930555, "grad_norm": 2.1740739345550537, "learning_rate": 6.721058312659001e-05, "loss": 5.7091, "step": 42800 }, { "epoch": 1.971961756618383, "grad_norm": 2.326795816421509, "learning_rate": 6.714206149546088e-05, "loss": 5.7289, "step": 42850 }, { "epoch": 1.9742628024437106, "grad_norm": 1.5980671644210815, "learning_rate": 6.707350336247122e-05, "loss": 5.7832, "step": 42900 }, { "epoch": 1.9765638482690382, "grad_norm": 2.2809202671051025, "learning_rate": 6.700490887360692e-05, "loss": 5.7059, "step": 42950 }, { "epoch": 1.9788648940943658, "grad_norm": 1.9004167318344116, "learning_rate": 6.69362781749313e-05, "loss": 5.6628, "step": 43000 }, { "epoch": 1.9788648940943658, "eval_loss": 5.719640254974365, "eval_runtime": 34.4151, "eval_samples_per_second": 11.158, "eval_steps_per_second": 5.579, "eval_tts_loss": 8.411078965968017, "step": 43000 }, { "epoch": 1.9811659399196935, "grad_norm": 1.7471081018447876, "learning_rate": 6.68676114125848e-05, "loss": 5.6815, "step": 43050 }, { "epoch": 1.9834669857450211, "grad_norm": 1.7803364992141724, "learning_rate": 6.679890873278463e-05, "loss": 5.7429, "step": 43100 }, { "epoch": 1.9857680315703488, "grad_norm": 1.9929248094558716, "learning_rate": 6.67301702818245e-05, "loss": 5.6513, "step": 43150 }, { "epoch": 1.9880690773956764, "grad_norm": 2.0318520069122314, "learning_rate": 6.666139620607426e-05, "loss": 5.7224, "step": 43200 }, { "epoch": 1.990370123221004, "grad_norm": 1.9984978437423706, "learning_rate": 6.659258665197966e-05, "loss": 5.6904, "step": 43250 }, { "epoch": 1.9926711690463317, "grad_norm": 1.683851718902588, "learning_rate": 6.652374176606196e-05, "loss": 5.6555, "step": 43300 }, { "epoch": 1.9949722148716593, "grad_norm": 1.9689456224441528, "learning_rate": 6.64548616949177e-05, "loss": 5.6529, "step": 43350 }, { "epoch": 1.9972732606969867, "grad_norm": 2.1004714965820312, "learning_rate": 6.638594658521831e-05, "loss": 5.7275, "step": 43400 }, { "epoch": 1.9995743065223144, "grad_norm": 1.9687811136245728, "learning_rate": 6.631699658370982e-05, "loss": 5.6693, "step": 43450 }, { "epoch": 2.001840836660262, "grad_norm": 2.143148899078369, "learning_rate": 6.624801183721259e-05, "loss": 5.5271, "step": 43500 }, { "epoch": 2.0041418824855897, "grad_norm": 2.1372528076171875, "learning_rate": 6.617899249262091e-05, "loss": 5.536, "step": 43550 }, { "epoch": 2.0064429283109173, "grad_norm": 2.233785629272461, "learning_rate": 6.610993869690284e-05, "loss": 5.5525, "step": 43600 }, { "epoch": 2.008743974136245, "grad_norm": 1.8795326948165894, "learning_rate": 6.604085059709968e-05, "loss": 5.5732, "step": 43650 }, { "epoch": 2.0110450199615726, "grad_norm": 1.8018789291381836, "learning_rate": 6.597172834032587e-05, "loss": 5.5204, "step": 43700 }, { "epoch": 2.0133460657869002, "grad_norm": 1.9285874366760254, "learning_rate": 6.590257207376853e-05, "loss": 5.5367, "step": 43750 }, { "epoch": 2.015647111612228, "grad_norm": 1.70220148563385, "learning_rate": 6.583338194468724e-05, "loss": 5.5598, "step": 43800 }, { "epoch": 2.0179481574375555, "grad_norm": 1.989536166191101, "learning_rate": 6.576415810041365e-05, "loss": 5.5385, "step": 43850 }, { "epoch": 2.020249203262883, "grad_norm": 1.9194942712783813, "learning_rate": 6.569490068835121e-05, "loss": 5.5637, "step": 43900 }, { "epoch": 2.022550249088211, "grad_norm": 2.056601047515869, "learning_rate": 6.562560985597485e-05, "loss": 5.4488, "step": 43950 }, { "epoch": 2.024851294913538, "grad_norm": 1.9041422605514526, "learning_rate": 6.55562857508307e-05, "loss": 5.5098, "step": 44000 }, { "epoch": 2.024851294913538, "eval_loss": 5.726797580718994, "eval_runtime": 35.4492, "eval_samples_per_second": 10.832, "eval_steps_per_second": 5.416, "eval_tts_loss": 8.615723402425886, "step": 44000 }, { "epoch": 2.0271523407388656, "grad_norm": 1.8542007207870483, "learning_rate": 6.548692852053569e-05, "loss": 5.4843, "step": 44050 }, { "epoch": 2.0294533865641933, "grad_norm": 1.984620213508606, "learning_rate": 6.541753831277729e-05, "loss": 5.5621, "step": 44100 }, { "epoch": 2.031754432389521, "grad_norm": 1.7690690755844116, "learning_rate": 6.534811527531323e-05, "loss": 5.5137, "step": 44150 }, { "epoch": 2.0340554782148486, "grad_norm": 1.8489230871200562, "learning_rate": 6.527865955597113e-05, "loss": 5.5089, "step": 44200 }, { "epoch": 2.036356524040176, "grad_norm": 1.9018126726150513, "learning_rate": 6.520917130264814e-05, "loss": 5.5108, "step": 44250 }, { "epoch": 2.038657569865504, "grad_norm": 1.6569920778274536, "learning_rate": 6.51396506633108e-05, "loss": 5.4855, "step": 44300 }, { "epoch": 2.0409586156908315, "grad_norm": 1.9102799892425537, "learning_rate": 6.507009778599454e-05, "loss": 5.5066, "step": 44350 }, { "epoch": 2.043259661516159, "grad_norm": 1.9131476879119873, "learning_rate": 6.500051281880343e-05, "loss": 5.5734, "step": 44400 }, { "epoch": 2.0455607073414868, "grad_norm": 1.878501296043396, "learning_rate": 6.493089590990993e-05, "loss": 5.4392, "step": 44450 }, { "epoch": 2.0478617531668144, "grad_norm": 1.891527771949768, "learning_rate": 6.486124720755444e-05, "loss": 5.5329, "step": 44500 }, { "epoch": 2.050162798992142, "grad_norm": 2.036492109298706, "learning_rate": 6.47915668600451e-05, "loss": 5.5437, "step": 44550 }, { "epoch": 2.0524638448174697, "grad_norm": 2.31026029586792, "learning_rate": 6.472185501575746e-05, "loss": 5.5779, "step": 44600 }, { "epoch": 2.0547648906427973, "grad_norm": 2.0821564197540283, "learning_rate": 6.465211182313409e-05, "loss": 5.5374, "step": 44650 }, { "epoch": 2.057065936468125, "grad_norm": 1.7920081615447998, "learning_rate": 6.458233743068436e-05, "loss": 5.5222, "step": 44700 }, { "epoch": 2.059366982293452, "grad_norm": 1.7931830883026123, "learning_rate": 6.4512531986984e-05, "loss": 5.4864, "step": 44750 }, { "epoch": 2.06166802811878, "grad_norm": 1.9625083208084106, "learning_rate": 6.444269564067495e-05, "loss": 5.5036, "step": 44800 }, { "epoch": 2.0639690739441074, "grad_norm": 1.7811228036880493, "learning_rate": 6.437282854046489e-05, "loss": 5.5504, "step": 44850 }, { "epoch": 2.066270119769435, "grad_norm": 1.8563337326049805, "learning_rate": 6.430293083512702e-05, "loss": 5.5086, "step": 44900 }, { "epoch": 2.0685711655947627, "grad_norm": 1.6173336505889893, "learning_rate": 6.423300267349969e-05, "loss": 5.5239, "step": 44950 }, { "epoch": 2.0708722114200904, "grad_norm": 2.1647262573242188, "learning_rate": 6.416304420448613e-05, "loss": 5.5063, "step": 45000 }, { "epoch": 2.0708722114200904, "eval_loss": 5.722973346710205, "eval_runtime": 34.5048, "eval_samples_per_second": 11.129, "eval_steps_per_second": 5.564, "eval_tts_loss": 8.551408735813336, "step": 45000 }, { "epoch": 2.073173257245418, "grad_norm": 1.8066037893295288, "learning_rate": 6.409305557705408e-05, "loss": 5.572, "step": 45050 }, { "epoch": 2.0754743030707457, "grad_norm": 2.123093366622925, "learning_rate": 6.402303694023548e-05, "loss": 5.5142, "step": 45100 }, { "epoch": 2.0777753488960733, "grad_norm": 1.8673418760299683, "learning_rate": 6.395298844312623e-05, "loss": 5.4898, "step": 45150 }, { "epoch": 2.080076394721401, "grad_norm": 2.1663706302642822, "learning_rate": 6.388291023488575e-05, "loss": 5.4957, "step": 45200 }, { "epoch": 2.0823774405467286, "grad_norm": 1.9892181158065796, "learning_rate": 6.381280246473678e-05, "loss": 5.5585, "step": 45250 }, { "epoch": 2.084678486372056, "grad_norm": 2.284586191177368, "learning_rate": 6.3742665281965e-05, "loss": 5.5635, "step": 45300 }, { "epoch": 2.086979532197384, "grad_norm": 2.3510870933532715, "learning_rate": 6.367249883591865e-05, "loss": 5.5342, "step": 45350 }, { "epoch": 2.0892805780227115, "grad_norm": 2.212132215499878, "learning_rate": 6.36023032760084e-05, "loss": 5.5002, "step": 45400 }, { "epoch": 2.091581623848039, "grad_norm": 1.725682020187378, "learning_rate": 6.35320787517068e-05, "loss": 5.5957, "step": 45450 }, { "epoch": 2.0938826696733663, "grad_norm": 2.1455774307250977, "learning_rate": 6.346182541254817e-05, "loss": 5.4675, "step": 45500 }, { "epoch": 2.096183715498694, "grad_norm": 2.027888298034668, "learning_rate": 6.33929493281682e-05, "loss": 5.5287, "step": 45550 }, { "epoch": 2.0984847613240216, "grad_norm": 1.811853051185608, "learning_rate": 6.332263937698848e-05, "loss": 5.5309, "step": 45600 }, { "epoch": 2.1007858071493493, "grad_norm": 1.9663208723068237, "learning_rate": 6.325230105692648e-05, "loss": 5.4658, "step": 45650 }, { "epoch": 2.103086852974677, "grad_norm": 1.7255808115005493, "learning_rate": 6.31819345177588e-05, "loss": 5.5497, "step": 45700 }, { "epoch": 2.1053878988000045, "grad_norm": 1.5630170106887817, "learning_rate": 6.311153990932208e-05, "loss": 5.4754, "step": 45750 }, { "epoch": 2.107688944625332, "grad_norm": 1.9326938390731812, "learning_rate": 6.304111738151283e-05, "loss": 5.5607, "step": 45800 }, { "epoch": 2.10998999045066, "grad_norm": 1.7057856321334839, "learning_rate": 6.297066708428691e-05, "loss": 5.4917, "step": 45850 }, { "epoch": 2.1122910362759875, "grad_norm": 1.9797825813293457, "learning_rate": 6.290018916765936e-05, "loss": 5.4714, "step": 45900 }, { "epoch": 2.114592082101315, "grad_norm": 2.143786907196045, "learning_rate": 6.282968378170405e-05, "loss": 5.4879, "step": 45950 }, { "epoch": 2.1168931279266427, "grad_norm": 2.2895607948303223, "learning_rate": 6.275915107655329e-05, "loss": 5.5423, "step": 46000 }, { "epoch": 2.1168931279266427, "eval_loss": 5.717980861663818, "eval_runtime": 34.9486, "eval_samples_per_second": 10.988, "eval_steps_per_second": 5.494, "eval_tts_loss": 8.597503433606374, "step": 46000 }, { "epoch": 2.1191941737519704, "grad_norm": 1.834234356880188, "learning_rate": 6.268859120239758e-05, "loss": 5.4918, "step": 46050 }, { "epoch": 2.121495219577298, "grad_norm": 1.891732931137085, "learning_rate": 6.261800430948532e-05, "loss": 5.4975, "step": 46100 }, { "epoch": 2.1237962654026257, "grad_norm": 1.9583476781845093, "learning_rate": 6.25473905481224e-05, "loss": 5.5338, "step": 46150 }, { "epoch": 2.1260973112279533, "grad_norm": 2.160682439804077, "learning_rate": 6.24767500686719e-05, "loss": 5.534, "step": 46200 }, { "epoch": 2.1283983570532805, "grad_norm": 1.747267484664917, "learning_rate": 6.240608302155384e-05, "loss": 5.4919, "step": 46250 }, { "epoch": 2.130699402878608, "grad_norm": 2.0734097957611084, "learning_rate": 6.233538955724481e-05, "loss": 5.534, "step": 46300 }, { "epoch": 2.133000448703936, "grad_norm": 1.9301162958145142, "learning_rate": 6.226466982627763e-05, "loss": 5.49, "step": 46350 }, { "epoch": 2.1353014945292634, "grad_norm": 1.7851401567459106, "learning_rate": 6.219392397924104e-05, "loss": 5.3952, "step": 46400 }, { "epoch": 2.137602540354591, "grad_norm": 1.7382208108901978, "learning_rate": 6.212315216677942e-05, "loss": 5.4901, "step": 46450 }, { "epoch": 2.1399035861799187, "grad_norm": 2.1053762435913086, "learning_rate": 6.205235453959244e-05, "loss": 5.5327, "step": 46500 }, { "epoch": 2.1422046320052464, "grad_norm": 1.9295450448989868, "learning_rate": 6.19815312484347e-05, "loss": 5.4875, "step": 46550 }, { "epoch": 2.144505677830574, "grad_norm": 2.141199827194214, "learning_rate": 6.191068244411551e-05, "loss": 5.5757, "step": 46600 }, { "epoch": 2.1468067236559016, "grad_norm": 2.1582906246185303, "learning_rate": 6.183980827749844e-05, "loss": 5.5675, "step": 46650 }, { "epoch": 2.1491077694812293, "grad_norm": 1.7495667934417725, "learning_rate": 6.176890889950115e-05, "loss": 5.5199, "step": 46700 }, { "epoch": 2.151408815306557, "grad_norm": 1.7843763828277588, "learning_rate": 6.16979844610949e-05, "loss": 5.4777, "step": 46750 }, { "epoch": 2.1537098611318846, "grad_norm": 1.7702149152755737, "learning_rate": 6.162703511330434e-05, "loss": 5.4738, "step": 46800 }, { "epoch": 2.156010906957212, "grad_norm": 1.8765311241149902, "learning_rate": 6.155606100720717e-05, "loss": 5.5664, "step": 46850 }, { "epoch": 2.15831195278254, "grad_norm": 1.8246150016784668, "learning_rate": 6.148506229393383e-05, "loss": 5.5637, "step": 46900 }, { "epoch": 2.160612998607867, "grad_norm": 2.1555979251861572, "learning_rate": 6.141403912466712e-05, "loss": 5.5311, "step": 46950 }, { "epoch": 2.1629140444331947, "grad_norm": 1.9029009342193604, "learning_rate": 6.134299165064193e-05, "loss": 5.4818, "step": 47000 }, { "epoch": 2.1629140444331947, "eval_loss": 5.707584381103516, "eval_runtime": 33.3657, "eval_samples_per_second": 11.509, "eval_steps_per_second": 5.754, "eval_tts_loss": 8.569347690304697, "step": 47000 }, { "epoch": 2.1652150902585223, "grad_norm": 1.874457597732544, "learning_rate": 6.127192002314491e-05, "loss": 5.4641, "step": 47050 }, { "epoch": 2.16751613608385, "grad_norm": 1.8370598554611206, "learning_rate": 6.120082439351416e-05, "loss": 5.4618, "step": 47100 }, { "epoch": 2.1698171819091776, "grad_norm": 1.8363277912139893, "learning_rate": 6.112970491313882e-05, "loss": 5.4306, "step": 47150 }, { "epoch": 2.1721182277345052, "grad_norm": 2.3280389308929443, "learning_rate": 6.105856173345891e-05, "loss": 5.5492, "step": 47200 }, { "epoch": 2.174419273559833, "grad_norm": 2.1096155643463135, "learning_rate": 6.098739500596485e-05, "loss": 5.507, "step": 47250 }, { "epoch": 2.1767203193851605, "grad_norm": 1.8868945837020874, "learning_rate": 6.0916204882197204e-05, "loss": 5.5058, "step": 47300 }, { "epoch": 2.179021365210488, "grad_norm": 2.4336724281311035, "learning_rate": 6.084499151374641e-05, "loss": 5.4931, "step": 47350 }, { "epoch": 2.181322411035816, "grad_norm": 1.9747580289840698, "learning_rate": 6.077375505225233e-05, "loss": 5.482, "step": 47400 }, { "epoch": 2.1836234568611435, "grad_norm": 2.0846846103668213, "learning_rate": 6.070249564940407e-05, "loss": 5.5427, "step": 47450 }, { "epoch": 2.185924502686471, "grad_norm": 1.875829815864563, "learning_rate": 6.063121345693952e-05, "loss": 5.5404, "step": 47500 }, { "epoch": 2.1882255485117987, "grad_norm": 1.8179715871810913, "learning_rate": 6.055990862664514e-05, "loss": 5.5046, "step": 47550 }, { "epoch": 2.1905265943371264, "grad_norm": 2.135223150253296, "learning_rate": 6.0488581310355586e-05, "loss": 5.5869, "step": 47600 }, { "epoch": 2.192827640162454, "grad_norm": 1.8659194707870483, "learning_rate": 6.041723165995339e-05, "loss": 5.46, "step": 47650 }, { "epoch": 2.1951286859877817, "grad_norm": 1.751010775566101, "learning_rate": 6.0345859827368646e-05, "loss": 5.4801, "step": 47700 }, { "epoch": 2.197429731813109, "grad_norm": 1.752781629562378, "learning_rate": 6.027446596457869e-05, "loss": 5.4676, "step": 47750 }, { "epoch": 2.1997307776384365, "grad_norm": 1.8623894453048706, "learning_rate": 6.0203050223607736e-05, "loss": 5.4861, "step": 47800 }, { "epoch": 2.202031823463764, "grad_norm": 2.1729423999786377, "learning_rate": 6.0131612756526635e-05, "loss": 5.5679, "step": 47850 }, { "epoch": 2.204332869289092, "grad_norm": 1.8969271183013916, "learning_rate": 6.006015371545246e-05, "loss": 5.5307, "step": 47900 }, { "epoch": 2.2066339151144194, "grad_norm": 1.9392940998077393, "learning_rate": 5.998867325254822e-05, "loss": 5.5678, "step": 47950 }, { "epoch": 2.208934960939747, "grad_norm": 1.9349250793457031, "learning_rate": 5.991717152002261e-05, "loss": 5.5207, "step": 48000 }, { "epoch": 2.208934960939747, "eval_loss": 5.705930233001709, "eval_runtime": 33.794, "eval_samples_per_second": 11.363, "eval_steps_per_second": 5.681, "eval_tts_loss": 8.537661055543742, "step": 48000 }, { "epoch": 2.2112360067650747, "grad_norm": 2.1593496799468994, "learning_rate": 5.9845648670129506e-05, "loss": 5.5717, "step": 48050 }, { "epoch": 2.2135370525904023, "grad_norm": 1.7322163581848145, "learning_rate": 5.9774104855167837e-05, "loss": 5.4747, "step": 48100 }, { "epoch": 2.21583809841573, "grad_norm": 1.878806710243225, "learning_rate": 5.9702540227481116e-05, "loss": 5.5021, "step": 48150 }, { "epoch": 2.2181391442410576, "grad_norm": 1.9864357709884644, "learning_rate": 5.963095493945724e-05, "loss": 5.5111, "step": 48200 }, { "epoch": 2.2204401900663853, "grad_norm": 2.206557273864746, "learning_rate": 5.9559349143528044e-05, "loss": 5.5295, "step": 48250 }, { "epoch": 2.222741235891713, "grad_norm": 1.8144375085830688, "learning_rate": 5.948772299216905e-05, "loss": 5.5521, "step": 48300 }, { "epoch": 2.2250422817170405, "grad_norm": 2.5472915172576904, "learning_rate": 5.941607663789912e-05, "loss": 5.5341, "step": 48350 }, { "epoch": 2.227343327542368, "grad_norm": 2.2753031253814697, "learning_rate": 5.9344410233280146e-05, "loss": 5.4789, "step": 48400 }, { "epoch": 2.2296443733676954, "grad_norm": 2.0804171562194824, "learning_rate": 5.927272393091671e-05, "loss": 5.5196, "step": 48450 }, { "epoch": 2.231945419193023, "grad_norm": 2.0795419216156006, "learning_rate": 5.9201017883455755e-05, "loss": 5.5605, "step": 48500 }, { "epoch": 2.2342464650183507, "grad_norm": 1.8969382047653198, "learning_rate": 5.91292922435863e-05, "loss": 5.5587, "step": 48550 }, { "epoch": 2.2365475108436783, "grad_norm": 2.2194700241088867, "learning_rate": 5.905754716403902e-05, "loss": 5.5229, "step": 48600 }, { "epoch": 2.238848556669006, "grad_norm": 1.9214152097702026, "learning_rate": 5.8985782797586055e-05, "loss": 5.566, "step": 48650 }, { "epoch": 2.2411496024943336, "grad_norm": 1.9062443971633911, "learning_rate": 5.891399929704058e-05, "loss": 5.5084, "step": 48700 }, { "epoch": 2.2434506483196612, "grad_norm": 2.313995599746704, "learning_rate": 5.88421968152565e-05, "loss": 5.547, "step": 48750 }, { "epoch": 2.245751694144989, "grad_norm": 1.8967597484588623, "learning_rate": 5.877037550512816e-05, "loss": 5.559, "step": 48800 }, { "epoch": 2.2480527399703165, "grad_norm": 2.220062494277954, "learning_rate": 5.869853551959e-05, "loss": 5.5156, "step": 48850 }, { "epoch": 2.250353785795644, "grad_norm": 2.281755208969116, "learning_rate": 5.8626677011616226e-05, "loss": 5.4842, "step": 48900 }, { "epoch": 2.252654831620972, "grad_norm": 1.9974268674850464, "learning_rate": 5.8554800134220465e-05, "loss": 5.5098, "step": 48950 }, { "epoch": 2.2549558774462994, "grad_norm": 2.0634398460388184, "learning_rate": 5.848290504045548e-05, "loss": 5.5407, "step": 49000 }, { "epoch": 2.2549558774462994, "eval_loss": 5.699641704559326, "eval_runtime": 33.4391, "eval_samples_per_second": 11.484, "eval_steps_per_second": 5.742, "eval_tts_loss": 8.564900359093189, "step": 49000 }, { "epoch": 2.257256923271627, "grad_norm": 2.1197509765625, "learning_rate": 5.8410991883412805e-05, "loss": 5.5355, "step": 49050 }, { "epoch": 2.2595579690969547, "grad_norm": 2.0381312370300293, "learning_rate": 5.833906081622248e-05, "loss": 5.4852, "step": 49100 }, { "epoch": 2.2618590149222824, "grad_norm": 2.316908359527588, "learning_rate": 5.826711199205261e-05, "loss": 5.4951, "step": 49150 }, { "epoch": 2.26416006074761, "grad_norm": 2.0368828773498535, "learning_rate": 5.819514556410919e-05, "loss": 5.5427, "step": 49200 }, { "epoch": 2.2664611065729376, "grad_norm": 1.6029542684555054, "learning_rate": 5.812316168563561e-05, "loss": 5.4828, "step": 49250 }, { "epoch": 2.268762152398265, "grad_norm": 1.8522335290908813, "learning_rate": 5.805116050991253e-05, "loss": 5.418, "step": 49300 }, { "epoch": 2.2710631982235925, "grad_norm": 1.811919093132019, "learning_rate": 5.7979142190257366e-05, "loss": 5.5058, "step": 49350 }, { "epoch": 2.27336424404892, "grad_norm": 1.9893652200698853, "learning_rate": 5.790710688002403e-05, "loss": 5.4568, "step": 49400 }, { "epoch": 2.2756652898742478, "grad_norm": 2.328045606613159, "learning_rate": 5.7835054732602664e-05, "loss": 5.4936, "step": 49450 }, { "epoch": 2.2779663356995754, "grad_norm": 1.901153326034546, "learning_rate": 5.7762985901419254e-05, "loss": 5.5469, "step": 49500 }, { "epoch": 2.280267381524903, "grad_norm": 2.4046385288238525, "learning_rate": 5.769090053993528e-05, "loss": 5.4819, "step": 49550 }, { "epoch": 2.2825684273502307, "grad_norm": 2.1178832054138184, "learning_rate": 5.7618798801647424e-05, "loss": 5.4857, "step": 49600 }, { "epoch": 2.2848694731755583, "grad_norm": 1.8963590860366821, "learning_rate": 5.7546680840087285e-05, "loss": 5.5016, "step": 49650 }, { "epoch": 2.287170519000886, "grad_norm": 1.8767808675765991, "learning_rate": 5.747454680882095e-05, "loss": 5.52, "step": 49700 }, { "epoch": 2.2894715648262136, "grad_norm": 2.342343330383301, "learning_rate": 5.740239686144877e-05, "loss": 5.505, "step": 49750 }, { "epoch": 2.2917726106515413, "grad_norm": 2.1045210361480713, "learning_rate": 5.7330231151604966e-05, "loss": 5.5173, "step": 49800 }, { "epoch": 2.294073656476869, "grad_norm": 1.9127671718597412, "learning_rate": 5.725804983295733e-05, "loss": 5.477, "step": 49850 }, { "epoch": 2.2963747023021965, "grad_norm": 2.1847991943359375, "learning_rate": 5.718729714514758e-05, "loss": 5.5034, "step": 49900 }, { "epoch": 2.2986757481275237, "grad_norm": 1.8255267143249512, "learning_rate": 5.711508537454882e-05, "loss": 5.4878, "step": 49950 }, { "epoch": 2.3009767939528514, "grad_norm": 1.9621268510818481, "learning_rate": 5.7042858453272044e-05, "loss": 5.5685, "step": 50000 }, { "epoch": 2.3009767939528514, "eval_loss": 5.693967342376709, "eval_runtime": 33.0645, "eval_samples_per_second": 11.614, "eval_steps_per_second": 5.807, "eval_tts_loss": 8.575001189177728, "step": 50000 }, { "epoch": 2.303277839778179, "grad_norm": 2.2509212493896484, "learning_rate": 5.697061653511539e-05, "loss": 5.4954, "step": 50050 }, { "epoch": 2.3055788856035067, "grad_norm": 1.7002668380737305, "learning_rate": 5.689835977390893e-05, "loss": 5.4782, "step": 50100 }, { "epoch": 2.3078799314288343, "grad_norm": 2.2835309505462646, "learning_rate": 5.682608832351433e-05, "loss": 5.495, "step": 50150 }, { "epoch": 2.310180977254162, "grad_norm": 1.9559133052825928, "learning_rate": 5.6753802337824526e-05, "loss": 5.4967, "step": 50200 }, { "epoch": 2.3124820230794896, "grad_norm": 1.6724474430084229, "learning_rate": 5.6681501970763426e-05, "loss": 5.487, "step": 50250 }, { "epoch": 2.314783068904817, "grad_norm": 1.6113662719726562, "learning_rate": 5.660918737628557e-05, "loss": 5.4981, "step": 50300 }, { "epoch": 2.317084114730145, "grad_norm": 2.1076881885528564, "learning_rate": 5.653685870837575e-05, "loss": 5.5033, "step": 50350 }, { "epoch": 2.3193851605554725, "grad_norm": 1.9511075019836426, "learning_rate": 5.646451612104877e-05, "loss": 5.5259, "step": 50400 }, { "epoch": 2.3216862063808, "grad_norm": 1.9726592302322388, "learning_rate": 5.639215976834906e-05, "loss": 5.5563, "step": 50450 }, { "epoch": 2.323987252206128, "grad_norm": 1.6485875844955444, "learning_rate": 5.631978980435035e-05, "loss": 5.4276, "step": 50500 }, { "epoch": 2.3262882980314554, "grad_norm": 1.8284225463867188, "learning_rate": 5.624740638315538e-05, "loss": 5.5438, "step": 50550 }, { "epoch": 2.328589343856783, "grad_norm": 1.7105023860931396, "learning_rate": 5.6175009658895486e-05, "loss": 5.5755, "step": 50600 }, { "epoch": 2.3308903896821107, "grad_norm": 2.0239145755767822, "learning_rate": 5.6102599785730406e-05, "loss": 5.5332, "step": 50650 }, { "epoch": 2.3331914355074383, "grad_norm": 2.122833490371704, "learning_rate": 5.6030176917847834e-05, "loss": 5.5843, "step": 50700 }, { "epoch": 2.335492481332766, "grad_norm": 1.7509838342666626, "learning_rate": 5.595774120946316e-05, "loss": 5.5071, "step": 50750 }, { "epoch": 2.337793527158093, "grad_norm": 1.953904628753662, "learning_rate": 5.5885292814819054e-05, "loss": 5.5062, "step": 50800 }, { "epoch": 2.340094572983421, "grad_norm": 2.0808749198913574, "learning_rate": 5.5812831888185305e-05, "loss": 5.5189, "step": 50850 }, { "epoch": 2.3423956188087485, "grad_norm": 1.8985790014266968, "learning_rate": 5.574035858385829e-05, "loss": 5.5271, "step": 50900 }, { "epoch": 2.344696664634076, "grad_norm": 2.062648057937622, "learning_rate": 5.5667873056160794e-05, "loss": 5.5491, "step": 50950 }, { "epoch": 2.3469977104594038, "grad_norm": 1.934853434562683, "learning_rate": 5.559537545944161e-05, "loss": 5.5204, "step": 51000 }, { "epoch": 2.3469977104594038, "eval_loss": 5.6846184730529785, "eval_runtime": 32.7588, "eval_samples_per_second": 11.722, "eval_steps_per_second": 5.861, "eval_tts_loss": 8.581908826910935, "step": 51000 }, { "epoch": 2.3492987562847314, "grad_norm": 1.8930600881576538, "learning_rate": 5.552286594807523e-05, "loss": 5.4913, "step": 51050 }, { "epoch": 2.351599802110059, "grad_norm": 1.7008662223815918, "learning_rate": 5.545034467646155e-05, "loss": 5.5381, "step": 51100 }, { "epoch": 2.3539008479353867, "grad_norm": 2.0752475261688232, "learning_rate": 5.537781179902548e-05, "loss": 5.4563, "step": 51150 }, { "epoch": 2.3562018937607143, "grad_norm": 1.9323570728302002, "learning_rate": 5.530526747021663e-05, "loss": 5.513, "step": 51200 }, { "epoch": 2.358502939586042, "grad_norm": 2.2523136138916016, "learning_rate": 5.5232711844509e-05, "loss": 5.4431, "step": 51250 }, { "epoch": 2.3608039854113696, "grad_norm": 1.8251011371612549, "learning_rate": 5.516014507640067e-05, "loss": 5.5666, "step": 51300 }, { "epoch": 2.3631050312366972, "grad_norm": 2.220978260040283, "learning_rate": 5.5087567320413435e-05, "loss": 5.4692, "step": 51350 }, { "epoch": 2.365406077062025, "grad_norm": 2.1732263565063477, "learning_rate": 5.501497873109248e-05, "loss": 5.5466, "step": 51400 }, { "epoch": 2.367707122887352, "grad_norm": 1.958206057548523, "learning_rate": 5.494237946300606e-05, "loss": 5.5023, "step": 51450 }, { "epoch": 2.3700081687126797, "grad_norm": 1.8300808668136597, "learning_rate": 5.486976967074517e-05, "loss": 5.5164, "step": 51500 }, { "epoch": 2.3723092145380074, "grad_norm": 2.154264211654663, "learning_rate": 5.479714950892322e-05, "loss": 5.5505, "step": 51550 }, { "epoch": 2.374610260363335, "grad_norm": 2.135356903076172, "learning_rate": 5.472451913217569e-05, "loss": 5.5268, "step": 51600 }, { "epoch": 2.3769113061886626, "grad_norm": 1.6626617908477783, "learning_rate": 5.465187869515983e-05, "loss": 5.4863, "step": 51650 }, { "epoch": 2.3792123520139903, "grad_norm": 2.074730157852173, "learning_rate": 5.457922835255428e-05, "loss": 5.5061, "step": 51700 }, { "epoch": 2.381513397839318, "grad_norm": 1.85642409324646, "learning_rate": 5.450656825905881e-05, "loss": 5.5448, "step": 51750 }, { "epoch": 2.3838144436646456, "grad_norm": 2.1079165935516357, "learning_rate": 5.4433898569393926e-05, "loss": 5.5477, "step": 51800 }, { "epoch": 2.386115489489973, "grad_norm": 2.0697200298309326, "learning_rate": 5.43612194383006e-05, "loss": 5.5246, "step": 51850 }, { "epoch": 2.388416535315301, "grad_norm": 1.7789839506149292, "learning_rate": 5.4288531020539856e-05, "loss": 5.413, "step": 51900 }, { "epoch": 2.3907175811406285, "grad_norm": 1.9652003049850464, "learning_rate": 5.421583347089253e-05, "loss": 5.5111, "step": 51950 }, { "epoch": 2.393018626965956, "grad_norm": 2.0850377082824707, "learning_rate": 5.414312694415892e-05, "loss": 5.5198, "step": 52000 }, { "epoch": 2.393018626965956, "eval_loss": 5.685731410980225, "eval_runtime": 33.6338, "eval_samples_per_second": 11.417, "eval_steps_per_second": 5.709, "eval_tts_loss": 8.644896936048733, "step": 52000 }, { "epoch": 2.3953196727912838, "grad_norm": 1.8597215414047241, "learning_rate": 5.407186598759514e-05, "loss": 5.5228, "step": 52050 }, { "epoch": 2.3976207186166114, "grad_norm": 2.0384023189544678, "learning_rate": 5.399914214299692e-05, "loss": 5.5392, "step": 52100 }, { "epoch": 2.399921764441939, "grad_norm": 2.2333080768585205, "learning_rate": 5.392640978272927e-05, "loss": 5.548, "step": 52150 }, { "epoch": 2.4022228102672667, "grad_norm": 1.875346064567566, "learning_rate": 5.38536690616666e-05, "loss": 5.5701, "step": 52200 }, { "epoch": 2.404523856092594, "grad_norm": 1.9534131288528442, "learning_rate": 5.3780920134701115e-05, "loss": 5.5083, "step": 52250 }, { "epoch": 2.4068249019179215, "grad_norm": 1.6405394077301025, "learning_rate": 5.37081631567425e-05, "loss": 5.5181, "step": 52300 }, { "epoch": 2.409125947743249, "grad_norm": 2.1161701679229736, "learning_rate": 5.363539828271756e-05, "loss": 5.5107, "step": 52350 }, { "epoch": 2.411426993568577, "grad_norm": 1.9055460691452026, "learning_rate": 5.3562625667569945e-05, "loss": 5.5516, "step": 52400 }, { "epoch": 2.4137280393939045, "grad_norm": 1.7940001487731934, "learning_rate": 5.348984546625975e-05, "loss": 5.4008, "step": 52450 }, { "epoch": 2.416029085219232, "grad_norm": 2.0796279907226562, "learning_rate": 5.341705783376325e-05, "loss": 5.4998, "step": 52500 }, { "epoch": 2.4183301310445597, "grad_norm": 1.901845932006836, "learning_rate": 5.334426292507255e-05, "loss": 5.5447, "step": 52550 }, { "epoch": 2.4206311768698874, "grad_norm": 2.040921926498413, "learning_rate": 5.3271460895195224e-05, "loss": 5.5044, "step": 52600 }, { "epoch": 2.422932222695215, "grad_norm": 1.8908004760742188, "learning_rate": 5.319865189915404e-05, "loss": 5.4536, "step": 52650 }, { "epoch": 2.4252332685205427, "grad_norm": 2.002741813659668, "learning_rate": 5.312583609198658e-05, "loss": 5.4478, "step": 52700 }, { "epoch": 2.4275343143458703, "grad_norm": 1.7424709796905518, "learning_rate": 5.305301362874492e-05, "loss": 5.4934, "step": 52750 }, { "epoch": 2.429835360171198, "grad_norm": 1.6303448677062988, "learning_rate": 5.2980184664495336e-05, "loss": 5.5592, "step": 52800 }, { "epoch": 2.4321364059965256, "grad_norm": 2.0222649574279785, "learning_rate": 5.29073493543179e-05, "loss": 5.5131, "step": 52850 }, { "epoch": 2.434437451821853, "grad_norm": 1.923041582107544, "learning_rate": 5.283450785330629e-05, "loss": 5.4102, "step": 52900 }, { "epoch": 2.4367384976471804, "grad_norm": 1.6201173067092896, "learning_rate": 5.276166031656727e-05, "loss": 5.5084, "step": 52950 }, { "epoch": 2.439039543472508, "grad_norm": 1.916977047920227, "learning_rate": 5.268880689922049e-05, "loss": 5.507, "step": 53000 }, { "epoch": 2.439039543472508, "eval_loss": 5.673426151275635, "eval_runtime": 32.5382, "eval_samples_per_second": 11.801, "eval_steps_per_second": 5.901, "eval_tts_loss": 8.599758905226478, "step": 53000 }, { "epoch": 2.4413405892978357, "grad_norm": 1.7312027215957642, "learning_rate": 5.261594775639812e-05, "loss": 5.4897, "step": 53050 }, { "epoch": 2.4436416351231633, "grad_norm": 1.809523582458496, "learning_rate": 5.2543083043244544e-05, "loss": 5.4711, "step": 53100 }, { "epoch": 2.445942680948491, "grad_norm": 1.8643138408660889, "learning_rate": 5.2470212914915995e-05, "loss": 5.5297, "step": 53150 }, { "epoch": 2.4482437267738186, "grad_norm": 2.256096124649048, "learning_rate": 5.2397337526580205e-05, "loss": 5.4853, "step": 53200 }, { "epoch": 2.4505447725991463, "grad_norm": 1.8998197317123413, "learning_rate": 5.2324457033416154e-05, "loss": 5.486, "step": 53250 }, { "epoch": 2.452845818424474, "grad_norm": 2.2109556198120117, "learning_rate": 5.225157159061366e-05, "loss": 5.5047, "step": 53300 }, { "epoch": 2.4551468642498016, "grad_norm": 2.0667428970336914, "learning_rate": 5.2178681353373096e-05, "loss": 5.4553, "step": 53350 }, { "epoch": 2.457447910075129, "grad_norm": 1.889162540435791, "learning_rate": 5.210578647690505e-05, "loss": 5.4784, "step": 53400 }, { "epoch": 2.459748955900457, "grad_norm": 2.207839012145996, "learning_rate": 5.203288711642996e-05, "loss": 5.5107, "step": 53450 }, { "epoch": 2.4620500017257845, "grad_norm": 1.8210887908935547, "learning_rate": 5.195998342717785e-05, "loss": 5.5447, "step": 53500 }, { "epoch": 2.464351047551112, "grad_norm": 1.9892346858978271, "learning_rate": 5.188707556438793e-05, "loss": 5.5459, "step": 53550 }, { "epoch": 2.4666520933764398, "grad_norm": 2.049320697784424, "learning_rate": 5.1814163683308324e-05, "loss": 5.5597, "step": 53600 }, { "epoch": 2.4689531392017674, "grad_norm": 2.0622878074645996, "learning_rate": 5.174124793919568e-05, "loss": 5.5092, "step": 53650 }, { "epoch": 2.471254185027095, "grad_norm": 2.1357452869415283, "learning_rate": 5.166832848731488e-05, "loss": 5.5497, "step": 53700 }, { "epoch": 2.4735552308524222, "grad_norm": 1.9745289087295532, "learning_rate": 5.159540548293874e-05, "loss": 5.4755, "step": 53750 }, { "epoch": 2.47585627667775, "grad_norm": 1.9489434957504272, "learning_rate": 5.152247908134761e-05, "loss": 5.5102, "step": 53800 }, { "epoch": 2.4781573225030775, "grad_norm": 2.2445592880249023, "learning_rate": 5.144954943782905e-05, "loss": 5.499, "step": 53850 }, { "epoch": 2.480458368328405, "grad_norm": 2.048757791519165, "learning_rate": 5.137661670767757e-05, "loss": 5.516, "step": 53900 }, { "epoch": 2.482759414153733, "grad_norm": 1.8567497730255127, "learning_rate": 5.130368104619422e-05, "loss": 5.487, "step": 53950 }, { "epoch": 2.4850604599790604, "grad_norm": 2.4062752723693848, "learning_rate": 5.123074260868631e-05, "loss": 5.5128, "step": 54000 }, { "epoch": 2.4850604599790604, "eval_loss": 5.6724090576171875, "eval_runtime": 34.2786, "eval_samples_per_second": 11.202, "eval_steps_per_second": 5.601, "eval_tts_loss": 8.662209894940489, "step": 54000 }, { "epoch": 2.487361505804388, "grad_norm": 1.895520806312561, "learning_rate": 5.1157801550467064e-05, "loss": 5.5928, "step": 54050 }, { "epoch": 2.4896625516297157, "grad_norm": 2.2730302810668945, "learning_rate": 5.108485802685526e-05, "loss": 5.4582, "step": 54100 }, { "epoch": 2.4919635974550434, "grad_norm": 1.9927568435668945, "learning_rate": 5.101191219317495e-05, "loss": 5.5229, "step": 54150 }, { "epoch": 2.494264643280371, "grad_norm": 1.9329229593276978, "learning_rate": 5.0940423184635246e-05, "loss": 5.549, "step": 54200 }, { "epoch": 2.4965656891056986, "grad_norm": 1.611726999282837, "learning_rate": 5.086747323527522e-05, "loss": 5.5071, "step": 54250 }, { "epoch": 2.4988667349310263, "grad_norm": 1.863103985786438, "learning_rate": 5.079452143874022e-05, "loss": 5.4997, "step": 54300 }, { "epoch": 2.5011677807563535, "grad_norm": 2.2052907943725586, "learning_rate": 5.07215679503719e-05, "loss": 5.5179, "step": 54350 }, { "epoch": 2.503468826581681, "grad_norm": 1.9319581985473633, "learning_rate": 5.064861292551552e-05, "loss": 5.5353, "step": 54400 }, { "epoch": 2.5057698724070088, "grad_norm": 2.0617754459381104, "learning_rate": 5.057565651951963e-05, "loss": 5.4807, "step": 54450 }, { "epoch": 2.5080709182323364, "grad_norm": 2.021653652191162, "learning_rate": 5.05026988877357e-05, "loss": 5.5129, "step": 54500 }, { "epoch": 2.510371964057664, "grad_norm": 2.167088031768799, "learning_rate": 5.0429740185517805e-05, "loss": 5.5161, "step": 54550 }, { "epoch": 2.5126730098829917, "grad_norm": 1.8927797079086304, "learning_rate": 5.0356780568222326e-05, "loss": 5.4857, "step": 54600 }, { "epoch": 2.5149740557083193, "grad_norm": 1.97638738155365, "learning_rate": 5.0283820191207574e-05, "loss": 5.4169, "step": 54650 }, { "epoch": 2.517275101533647, "grad_norm": 1.7517626285552979, "learning_rate": 5.0210859209833485e-05, "loss": 5.4838, "step": 54700 }, { "epoch": 2.5195761473589746, "grad_norm": 1.9045029878616333, "learning_rate": 5.013789777946125e-05, "loss": 5.474, "step": 54750 }, { "epoch": 2.5218771931843023, "grad_norm": 2.1002869606018066, "learning_rate": 5.006493605545308e-05, "loss": 5.4747, "step": 54800 }, { "epoch": 2.52417823900963, "grad_norm": 1.861536979675293, "learning_rate": 4.9991974193171746e-05, "loss": 5.4659, "step": 54850 }, { "epoch": 2.5264792848349575, "grad_norm": 2.3293421268463135, "learning_rate": 4.991901234798037e-05, "loss": 5.4849, "step": 54900 }, { "epoch": 2.528780330660285, "grad_norm": 1.9434945583343506, "learning_rate": 4.9846050675241994e-05, "loss": 5.5403, "step": 54950 }, { "epoch": 2.531081376485613, "grad_norm": 1.791527509689331, "learning_rate": 4.977308933031931e-05, "loss": 5.5135, "step": 55000 }, { "epoch": 2.531081376485613, "eval_loss": 5.663011074066162, "eval_runtime": 33.2046, "eval_samples_per_second": 11.565, "eval_steps_per_second": 5.782, "eval_tts_loss": 8.593881697905651, "step": 55000 }, { "epoch": 2.5333824223109405, "grad_norm": 2.0611414909362793, "learning_rate": 4.970012846857432e-05, "loss": 5.5242, "step": 55050 }, { "epoch": 2.535683468136268, "grad_norm": 1.8130731582641602, "learning_rate": 4.962716824536796e-05, "loss": 5.4987, "step": 55100 }, { "epoch": 2.5379845139615957, "grad_norm": 2.204890489578247, "learning_rate": 4.9554208816059876e-05, "loss": 5.5122, "step": 55150 }, { "epoch": 2.5402855597869234, "grad_norm": 1.6537048816680908, "learning_rate": 4.9481250336007955e-05, "loss": 5.5315, "step": 55200 }, { "epoch": 2.542586605612251, "grad_norm": 2.3211939334869385, "learning_rate": 4.9408292960568096e-05, "loss": 5.5083, "step": 55250 }, { "epoch": 2.5448876514375782, "grad_norm": 2.3750381469726562, "learning_rate": 4.933533684509383e-05, "loss": 5.4613, "step": 55300 }, { "epoch": 2.547188697262906, "grad_norm": 1.7778195142745972, "learning_rate": 4.926238214493604e-05, "loss": 5.5137, "step": 55350 }, { "epoch": 2.5494897430882335, "grad_norm": 2.055008888244629, "learning_rate": 4.918942901544257e-05, "loss": 5.4564, "step": 55400 }, { "epoch": 2.551790788913561, "grad_norm": 2.363105297088623, "learning_rate": 4.9116477611957865e-05, "loss": 5.5054, "step": 55450 }, { "epoch": 2.554091834738889, "grad_norm": 2.1759395599365234, "learning_rate": 4.904352808982281e-05, "loss": 5.4812, "step": 55500 }, { "epoch": 2.5563928805642164, "grad_norm": 1.9954317808151245, "learning_rate": 4.897058060437419e-05, "loss": 5.5206, "step": 55550 }, { "epoch": 2.558693926389544, "grad_norm": 2.1544883251190186, "learning_rate": 4.88976353109445e-05, "loss": 5.5394, "step": 55600 }, { "epoch": 2.5609949722148717, "grad_norm": 2.238513231277466, "learning_rate": 4.882469236486155e-05, "loss": 5.4833, "step": 55650 }, { "epoch": 2.5632960180401994, "grad_norm": 2.2825424671173096, "learning_rate": 4.875175192144814e-05, "loss": 5.4567, "step": 55700 }, { "epoch": 2.565597063865527, "grad_norm": 1.8230923414230347, "learning_rate": 4.867881413602178e-05, "loss": 5.5013, "step": 55750 }, { "epoch": 2.5678981096908546, "grad_norm": 2.1068201065063477, "learning_rate": 4.8605879163894286e-05, "loss": 5.5096, "step": 55800 }, { "epoch": 2.570199155516182, "grad_norm": 2.3456852436065674, "learning_rate": 4.853294716037149e-05, "loss": 5.4725, "step": 55850 }, { "epoch": 2.5725002013415095, "grad_norm": 2.4963150024414062, "learning_rate": 4.846001828075292e-05, "loss": 5.4823, "step": 55900 }, { "epoch": 2.574801247166837, "grad_norm": 1.9656461477279663, "learning_rate": 4.838709268033141e-05, "loss": 5.4807, "step": 55950 }, { "epoch": 2.5771022929921648, "grad_norm": 2.1608853340148926, "learning_rate": 4.8314170514392874e-05, "loss": 5.5417, "step": 56000 }, { "epoch": 2.5771022929921648, "eval_loss": 5.6606011390686035, "eval_runtime": 32.8942, "eval_samples_per_second": 11.674, "eval_steps_per_second": 5.837, "eval_tts_loss": 8.585531058636887, "step": 56000 }, { "epoch": 2.5794033388174924, "grad_norm": 2.1026828289031982, "learning_rate": 4.8241251938215855e-05, "loss": 5.4739, "step": 56050 }, { "epoch": 2.58170438464282, "grad_norm": 2.2356417179107666, "learning_rate": 4.816833710707128e-05, "loss": 5.4498, "step": 56100 }, { "epoch": 2.5840054304681477, "grad_norm": 2.1985368728637695, "learning_rate": 4.80954261762221e-05, "loss": 5.4379, "step": 56150 }, { "epoch": 2.5863064762934753, "grad_norm": 1.9118517637252808, "learning_rate": 4.8022519300922944e-05, "loss": 5.5355, "step": 56200 }, { "epoch": 2.588607522118803, "grad_norm": 2.080960512161255, "learning_rate": 4.794961663641985e-05, "loss": 5.4729, "step": 56250 }, { "epoch": 2.5909085679441306, "grad_norm": 1.9839308261871338, "learning_rate": 4.787671833794983e-05, "loss": 5.4719, "step": 56300 }, { "epoch": 2.5932096137694582, "grad_norm": Infinity, "learning_rate": 4.78052823909725e-05, "loss": 5.4361, "step": 56350 }, { "epoch": 2.595510659594786, "grad_norm": 1.9188679456710815, "learning_rate": 4.77323931951916e-05, "loss": 5.5138, "step": 56400 }, { "epoch": 2.5978117054201135, "grad_norm": 2.0594546794891357, "learning_rate": 4.765950882799373e-05, "loss": 5.4922, "step": 56450 }, { "epoch": 2.600112751245441, "grad_norm": 2.096745729446411, "learning_rate": 4.7586629444577e-05, "loss": 5.5055, "step": 56500 }, { "epoch": 2.602413797070769, "grad_norm": 2.441319227218628, "learning_rate": 4.751375520012884e-05, "loss": 5.4756, "step": 56550 }, { "epoch": 2.6047148428960964, "grad_norm": 2.135594606399536, "learning_rate": 4.744088624982577e-05, "loss": 5.4237, "step": 56600 }, { "epoch": 2.607015888721424, "grad_norm": 2.0528063774108887, "learning_rate": 4.736802274883307e-05, "loss": 5.4687, "step": 56650 }, { "epoch": 2.6093169345467517, "grad_norm": 2.743708372116089, "learning_rate": 4.7295164852304374e-05, "loss": 5.5248, "step": 56700 }, { "epoch": 2.6116179803720794, "grad_norm": 2.06074857711792, "learning_rate": 4.722231271538139e-05, "loss": 5.4548, "step": 56750 }, { "epoch": 2.6139190261974066, "grad_norm": 2.297532558441162, "learning_rate": 4.71509233586696e-05, "loss": 5.4986, "step": 56800 }, { "epoch": 2.616220072022734, "grad_norm": 2.359525203704834, "learning_rate": 4.70780830834167e-05, "loss": 5.5232, "step": 56850 }, { "epoch": 2.618521117848062, "grad_norm": 2.092503786087036, "learning_rate": 4.700524903001777e-05, "loss": 5.4316, "step": 56900 }, { "epoch": 2.6208221636733895, "grad_norm": 1.9667569398880005, "learning_rate": 4.6932421353563806e-05, "loss": 5.422, "step": 56950 }, { "epoch": 2.623123209498717, "grad_norm": 2.1023378372192383, "learning_rate": 4.685960020913213e-05, "loss": 5.4845, "step": 57000 }, { "epoch": 2.623123209498717, "eval_loss": 5.65239953994751, "eval_runtime": 34.0261, "eval_samples_per_second": 11.285, "eval_steps_per_second": 5.643, "eval_tts_loss": 8.61307597893829, "step": 57000 }, { "epoch": 2.6254242553240448, "grad_norm": 2.099184036254883, "learning_rate": 4.6786785751786216e-05, "loss": 5.522, "step": 57050 }, { "epoch": 2.6277253011493724, "grad_norm": 2.0189764499664307, "learning_rate": 4.671397813657527e-05, "loss": 5.4654, "step": 57100 }, { "epoch": 2.6300263469747, "grad_norm": 1.772560954093933, "learning_rate": 4.6641177518533926e-05, "loss": 5.4866, "step": 57150 }, { "epoch": 2.6323273928000277, "grad_norm": 2.440157890319824, "learning_rate": 4.656838405268195e-05, "loss": 5.5017, "step": 57200 }, { "epoch": 2.6346284386253553, "grad_norm": 1.6734415292739868, "learning_rate": 4.649559789402385e-05, "loss": 5.5278, "step": 57250 }, { "epoch": 2.636929484450683, "grad_norm": 1.8009145259857178, "learning_rate": 4.6422819197548586e-05, "loss": 5.4958, "step": 57300 }, { "epoch": 2.63923053027601, "grad_norm": 2.123201370239258, "learning_rate": 4.63500481182292e-05, "loss": 5.5163, "step": 57350 }, { "epoch": 2.641531576101338, "grad_norm": 1.80511474609375, "learning_rate": 4.627728481102257e-05, "loss": 5.431, "step": 57400 }, { "epoch": 2.6438326219266655, "grad_norm": 2.240438222885132, "learning_rate": 4.620452943086898e-05, "loss": 5.4903, "step": 57450 }, { "epoch": 2.646133667751993, "grad_norm": 2.5990798473358154, "learning_rate": 4.613178213269184e-05, "loss": 5.4867, "step": 57500 }, { "epoch": 2.6484347135773207, "grad_norm": 2.257694959640503, "learning_rate": 4.6059043071397374e-05, "loss": 5.4922, "step": 57550 }, { "epoch": 2.6507357594026484, "grad_norm": 2.099280595779419, "learning_rate": 4.598631240187424e-05, "loss": 5.5309, "step": 57600 }, { "epoch": 2.653036805227976, "grad_norm": 2.5522916316986084, "learning_rate": 4.591359027899323e-05, "loss": 5.4671, "step": 57650 }, { "epoch": 2.6553378510533037, "grad_norm": 1.9019410610198975, "learning_rate": 4.5840876857606966e-05, "loss": 5.4628, "step": 57700 }, { "epoch": 2.6576388968786313, "grad_norm": 1.938440203666687, "learning_rate": 4.576817229254947e-05, "loss": 5.4546, "step": 57750 }, { "epoch": 2.659939942703959, "grad_norm": 1.9857474565505981, "learning_rate": 4.569547673863601e-05, "loss": 5.4707, "step": 57800 }, { "epoch": 2.6622409885292866, "grad_norm": 2.172809600830078, "learning_rate": 4.562279035066259e-05, "loss": 5.5027, "step": 57850 }, { "epoch": 2.6645420343546142, "grad_norm": 2.026803731918335, "learning_rate": 4.5550113283405716e-05, "loss": 5.5109, "step": 57900 }, { "epoch": 2.666843080179942, "grad_norm": 2.1904194355010986, "learning_rate": 4.547744569162204e-05, "loss": 5.4626, "step": 57950 }, { "epoch": 2.6691441260052695, "grad_norm": 2.199883460998535, "learning_rate": 4.540478773004804e-05, "loss": 5.5057, "step": 58000 }, { "epoch": 2.6691441260052695, "eval_loss": 5.646541118621826, "eval_runtime": 32.9768, "eval_samples_per_second": 11.645, "eval_steps_per_second": 5.822, "eval_tts_loss": 8.600281850515511, "step": 58000 }, { "epoch": 2.671445171830597, "grad_norm": 1.8649810552597046, "learning_rate": 4.533213955339972e-05, "loss": 5.4598, "step": 58050 }, { "epoch": 2.673746217655925, "grad_norm": 1.9523557424545288, "learning_rate": 4.5259501316372205e-05, "loss": 5.4585, "step": 58100 }, { "epoch": 2.6760472634812524, "grad_norm": 2.0407845973968506, "learning_rate": 4.518687317363947e-05, "loss": 5.4856, "step": 58150 }, { "epoch": 2.67834830930658, "grad_norm": 1.945887804031372, "learning_rate": 4.5114255279853987e-05, "loss": 5.4216, "step": 58200 }, { "epoch": 2.6806493551319077, "grad_norm": 1.912412166595459, "learning_rate": 4.5041647789646426e-05, "loss": 5.4695, "step": 58250 }, { "epoch": 2.682950400957235, "grad_norm": 2.170290946960449, "learning_rate": 4.496905085762529e-05, "loss": 5.4304, "step": 58300 }, { "epoch": 2.6852514467825626, "grad_norm": 2.0460116863250732, "learning_rate": 4.4896464638376594e-05, "loss": 5.5117, "step": 58350 }, { "epoch": 2.68755249260789, "grad_norm": 1.9400469064712524, "learning_rate": 4.4823889286463554e-05, "loss": 5.4373, "step": 58400 }, { "epoch": 2.689853538433218, "grad_norm": 2.4364376068115234, "learning_rate": 4.4751324956426235e-05, "loss": 5.4845, "step": 58450 }, { "epoch": 2.6921545842585455, "grad_norm": 1.7145034074783325, "learning_rate": 4.467877180278124e-05, "loss": 5.4182, "step": 58500 }, { "epoch": 2.694455630083873, "grad_norm": 2.023035764694214, "learning_rate": 4.460622998002136e-05, "loss": 5.4882, "step": 58550 }, { "epoch": 2.6967566759092008, "grad_norm": 2.2280147075653076, "learning_rate": 4.4533699642615295e-05, "loss": 5.4417, "step": 58600 }, { "epoch": 2.6990577217345284, "grad_norm": 1.7972197532653809, "learning_rate": 4.4461180945007235e-05, "loss": 5.5224, "step": 58650 }, { "epoch": 2.701358767559856, "grad_norm": 2.305210828781128, "learning_rate": 4.438867404161663e-05, "loss": 5.4077, "step": 58700 }, { "epoch": 2.7036598133851837, "grad_norm": 2.05381178855896, "learning_rate": 4.431617908683778e-05, "loss": 5.5049, "step": 58750 }, { "epoch": 2.7059608592105113, "grad_norm": 1.8015038967132568, "learning_rate": 4.424369623503957e-05, "loss": 5.4476, "step": 58800 }, { "epoch": 2.7082619050358385, "grad_norm": 1.750535249710083, "learning_rate": 4.417122564056508e-05, "loss": 5.5655, "step": 58850 }, { "epoch": 2.710562950861166, "grad_norm": 2.220609188079834, "learning_rate": 4.409876745773134e-05, "loss": 5.5094, "step": 58900 }, { "epoch": 2.712863996686494, "grad_norm": 1.8255053758621216, "learning_rate": 4.402632184082892e-05, "loss": 5.5111, "step": 58950 }, { "epoch": 2.7151650425118214, "grad_norm": 1.8717236518859863, "learning_rate": 4.3953888944121625e-05, "loss": 5.4799, "step": 59000 }, { "epoch": 2.7151650425118214, "eval_loss": 5.639944553375244, "eval_runtime": 33.1091, "eval_samples_per_second": 11.598, "eval_steps_per_second": 5.799, "eval_tts_loss": 8.586594152952083, "step": 59000 }, { "epoch": 2.717466088337149, "grad_norm": 1.870437741279602, "learning_rate": 4.3881468921846186e-05, "loss": 5.4899, "step": 59050 }, { "epoch": 2.7197671341624767, "grad_norm": 2.0649781227111816, "learning_rate": 4.38090619282119e-05, "loss": 5.4897, "step": 59100 }, { "epoch": 2.7220681799878044, "grad_norm": 2.2909843921661377, "learning_rate": 4.373666811740038e-05, "loss": 5.4443, "step": 59150 }, { "epoch": 2.724369225813132, "grad_norm": 1.9959808588027954, "learning_rate": 4.36642876435651e-05, "loss": 5.421, "step": 59200 }, { "epoch": 2.7266702716384597, "grad_norm": 1.8193045854568481, "learning_rate": 4.3591920660831165e-05, "loss": 5.5136, "step": 59250 }, { "epoch": 2.7289713174637873, "grad_norm": 2.53044056892395, "learning_rate": 4.351956732329494e-05, "loss": 5.4864, "step": 59300 }, { "epoch": 2.731272363289115, "grad_norm": 2.0809009075164795, "learning_rate": 4.344722778502376e-05, "loss": 5.5361, "step": 59350 }, { "epoch": 2.7335734091144426, "grad_norm": 1.832938313484192, "learning_rate": 4.3374902200055544e-05, "loss": 5.4392, "step": 59400 }, { "epoch": 2.73587445493977, "grad_norm": 2.035078763961792, "learning_rate": 4.330259072239853e-05, "loss": 5.5084, "step": 59450 }, { "epoch": 2.738175500765098, "grad_norm": 2.0738892555236816, "learning_rate": 4.3230293506030885e-05, "loss": 5.5201, "step": 59500 }, { "epoch": 2.7404765465904255, "grad_norm": 2.5500268936157227, "learning_rate": 4.315801070490042e-05, "loss": 5.4403, "step": 59550 }, { "epoch": 2.742777592415753, "grad_norm": 2.205916404724121, "learning_rate": 4.308574247292428e-05, "loss": 5.4321, "step": 59600 }, { "epoch": 2.745078638241081, "grad_norm": 1.8907051086425781, "learning_rate": 4.3013488963988544e-05, "loss": 5.5129, "step": 59650 }, { "epoch": 2.7473796840664084, "grad_norm": 2.2338571548461914, "learning_rate": 4.2941250331947955e-05, "loss": 5.4867, "step": 59700 }, { "epoch": 2.749680729891736, "grad_norm": 1.948060393333435, "learning_rate": 4.2869026730625586e-05, "loss": 5.4172, "step": 59750 }, { "epoch": 2.7519817757170633, "grad_norm": 1.7842371463775635, "learning_rate": 4.279681831381251e-05, "loss": 5.4382, "step": 59800 }, { "epoch": 2.754282821542391, "grad_norm": 1.6758701801300049, "learning_rate": 4.272462523526743e-05, "loss": 5.5079, "step": 59850 }, { "epoch": 2.7565838673677185, "grad_norm": 2.385894775390625, "learning_rate": 4.265244764871642e-05, "loss": 5.4478, "step": 59900 }, { "epoch": 2.758884913193046, "grad_norm": 1.8913649320602417, "learning_rate": 4.2580285707852554e-05, "loss": 5.4042, "step": 59950 }, { "epoch": 2.761185959018374, "grad_norm": 2.1212165355682373, "learning_rate": 4.250813956633561e-05, "loss": 5.4479, "step": 60000 }, { "epoch": 2.761185959018374, "eval_loss": 5.63777494430542, "eval_runtime": 33.443, "eval_samples_per_second": 11.482, "eval_steps_per_second": 5.741, "eval_tts_loss": 8.638248861690514, "step": 60000 }, { "epoch": 2.7634870048437015, "grad_norm": 2.0569007396698, "learning_rate": 4.24360093777917e-05, "loss": 5.4278, "step": 60050 }, { "epoch": 2.765788050669029, "grad_norm": 1.98093843460083, "learning_rate": 4.236389529581297e-05, "loss": 5.4418, "step": 60100 }, { "epoch": 2.7680890964943567, "grad_norm": 1.9621886014938354, "learning_rate": 4.229179747395727e-05, "loss": 5.4371, "step": 60150 }, { "epoch": 2.7703901423196844, "grad_norm": 2.040923595428467, "learning_rate": 4.221971606574785e-05, "loss": 5.4686, "step": 60200 }, { "epoch": 2.772691188145012, "grad_norm": 2.557758331298828, "learning_rate": 4.214765122467297e-05, "loss": 5.4559, "step": 60250 }, { "epoch": 2.7749922339703397, "grad_norm": 1.9892888069152832, "learning_rate": 4.207560310418564e-05, "loss": 5.4328, "step": 60300 }, { "epoch": 2.777293279795667, "grad_norm": 1.8860628604888916, "learning_rate": 4.200357185770326e-05, "loss": 5.4393, "step": 60350 }, { "epoch": 2.7795943256209945, "grad_norm": 2.2031514644622803, "learning_rate": 4.193155763860727e-05, "loss": 5.4776, "step": 60400 }, { "epoch": 2.781895371446322, "grad_norm": 2.3934335708618164, "learning_rate": 4.1859560600242904e-05, "loss": 5.4665, "step": 60450 }, { "epoch": 2.78419641727165, "grad_norm": 2.1561715602874756, "learning_rate": 4.1787580895918774e-05, "loss": 5.408, "step": 60500 }, { "epoch": 2.7864974630969774, "grad_norm": 2.138029098510742, "learning_rate": 4.171561867890661e-05, "loss": 5.5624, "step": 60550 }, { "epoch": 2.788798508922305, "grad_norm": 1.9366000890731812, "learning_rate": 4.164367410244087e-05, "loss": 5.4615, "step": 60600 }, { "epoch": 2.7910995547476327, "grad_norm": 2.0961766242980957, "learning_rate": 4.1571747319718457e-05, "loss": 5.4132, "step": 60650 }, { "epoch": 2.7934006005729604, "grad_norm": 2.026878833770752, "learning_rate": 4.1499838483898426e-05, "loss": 5.4426, "step": 60700 }, { "epoch": 2.795701646398288, "grad_norm": 2.110774278640747, "learning_rate": 4.142794774810156e-05, "loss": 5.4435, "step": 60750 }, { "epoch": 2.7980026922236156, "grad_norm": 1.7830619812011719, "learning_rate": 4.135607526541013e-05, "loss": 5.4528, "step": 60800 }, { "epoch": 2.8003037380489433, "grad_norm": 1.9309149980545044, "learning_rate": 4.128422118886754e-05, "loss": 5.4216, "step": 60850 }, { "epoch": 2.802604783874271, "grad_norm": 2.1686344146728516, "learning_rate": 4.121238567147801e-05, "loss": 5.4811, "step": 60900 }, { "epoch": 2.8049058296995986, "grad_norm": 1.9331154823303223, "learning_rate": 4.114056886620618e-05, "loss": 5.4188, "step": 60950 }, { "epoch": 2.807206875524926, "grad_norm": 2.110539674758911, "learning_rate": 4.106877092597692e-05, "loss": 5.4931, "step": 61000 }, { "epoch": 2.807206875524926, "eval_loss": 5.6315765380859375, "eval_runtime": 33.9234, "eval_samples_per_second": 11.32, "eval_steps_per_second": 5.66, "eval_tts_loss": 8.616176136626422, "step": 61000 }, { "epoch": 2.809507921350254, "grad_norm": 1.987831711769104, "learning_rate": 4.099699200367488e-05, "loss": 5.482, "step": 61050 }, { "epoch": 2.8118089671755815, "grad_norm": 2.204674243927002, "learning_rate": 4.0925232252144266e-05, "loss": 5.4937, "step": 61100 }, { "epoch": 2.814110013000909, "grad_norm": 2.2550504207611084, "learning_rate": 4.085349182418841e-05, "loss": 5.4465, "step": 61150 }, { "epoch": 2.8164110588262368, "grad_norm": 2.125481367111206, "learning_rate": 4.0781770872569514e-05, "loss": 5.3974, "step": 61200 }, { "epoch": 2.8187121046515644, "grad_norm": 1.787434458732605, "learning_rate": 4.07100695500083e-05, "loss": 5.395, "step": 61250 }, { "epoch": 2.8210131504768916, "grad_norm": 2.028134346008301, "learning_rate": 4.063982144515187e-05, "loss": 5.4681, "step": 61300 }, { "epoch": 2.8233141963022192, "grad_norm": 2.4000136852264404, "learning_rate": 4.056815943851771e-05, "loss": 5.4792, "step": 61350 }, { "epoch": 2.825615242127547, "grad_norm": 1.750549077987671, "learning_rate": 4.049651751579987e-05, "loss": 5.5025, "step": 61400 }, { "epoch": 2.8279162879528745, "grad_norm": 2.1570894718170166, "learning_rate": 4.042489582955082e-05, "loss": 5.4873, "step": 61450 }, { "epoch": 2.830217333778202, "grad_norm": 1.8570895195007324, "learning_rate": 4.0353294532279904e-05, "loss": 5.4529, "step": 61500 }, { "epoch": 2.83251837960353, "grad_norm": 2.0952110290527344, "learning_rate": 4.028171377645307e-05, "loss": 5.4145, "step": 61550 }, { "epoch": 2.8348194254288575, "grad_norm": 1.6754381656646729, "learning_rate": 4.021015371449254e-05, "loss": 5.4637, "step": 61600 }, { "epoch": 2.837120471254185, "grad_norm": 1.8659425973892212, "learning_rate": 4.013861449877643e-05, "loss": 5.4274, "step": 61650 }, { "epoch": 2.8394215170795127, "grad_norm": 1.94052255153656, "learning_rate": 4.006709628163849e-05, "loss": 5.4469, "step": 61700 }, { "epoch": 2.8417225629048404, "grad_norm": 2.1029059886932373, "learning_rate": 3.999559921536776e-05, "loss": 5.4863, "step": 61750 }, { "epoch": 2.844023608730168, "grad_norm": 2.0127499103546143, "learning_rate": 3.9924123452208226e-05, "loss": 5.4694, "step": 61800 }, { "epoch": 2.846324654555495, "grad_norm": 1.892794132232666, "learning_rate": 3.985266914435853e-05, "loss": 5.4628, "step": 61850 }, { "epoch": 2.848625700380823, "grad_norm": 1.9065146446228027, "learning_rate": 3.9781236443971624e-05, "loss": 5.4541, "step": 61900 }, { "epoch": 2.8509267462061505, "grad_norm": 1.7953850030899048, "learning_rate": 3.970982550315445e-05, "loss": 5.5099, "step": 61950 }, { "epoch": 2.853227792031478, "grad_norm": 2.3671352863311768, "learning_rate": 3.96384364739676e-05, "loss": 5.4451, "step": 62000 }, { "epoch": 2.853227792031478, "eval_loss": 5.629026889801025, "eval_runtime": 33.1809, "eval_samples_per_second": 11.573, "eval_steps_per_second": 5.786, "eval_tts_loss": 8.676706925015079, "step": 62000 }, { "epoch": 2.855528837856806, "grad_norm": 2.3126471042633057, "learning_rate": 3.9567069508425006e-05, "loss": 5.4794, "step": 62050 }, { "epoch": 2.8578298836821334, "grad_norm": 2.4819459915161133, "learning_rate": 3.9495724758493645e-05, "loss": 5.3972, "step": 62100 }, { "epoch": 2.860130929507461, "grad_norm": 2.0319995880126953, "learning_rate": 3.9424402376093166e-05, "loss": 5.5088, "step": 62150 }, { "epoch": 2.8624319753327887, "grad_norm": 2.0810019969940186, "learning_rate": 3.9353102513095615e-05, "loss": 5.4903, "step": 62200 }, { "epoch": 2.8647330211581163, "grad_norm": 2.255082607269287, "learning_rate": 3.928182532132506e-05, "loss": 5.4857, "step": 62250 }, { "epoch": 2.867034066983444, "grad_norm": 1.9748070240020752, "learning_rate": 3.92105709525573e-05, "loss": 5.4819, "step": 62300 }, { "epoch": 2.8693351128087716, "grad_norm": 2.1835029125213623, "learning_rate": 3.913933955851953e-05, "loss": 5.4376, "step": 62350 }, { "epoch": 2.8716361586340993, "grad_norm": 2.0077571868896484, "learning_rate": 3.906813129089004e-05, "loss": 5.4494, "step": 62400 }, { "epoch": 2.873937204459427, "grad_norm": 1.9124616384506226, "learning_rate": 3.8996946301297864e-05, "loss": 5.5265, "step": 62450 }, { "epoch": 2.8762382502847545, "grad_norm": 1.8583190441131592, "learning_rate": 3.892578474132248e-05, "loss": 5.4479, "step": 62500 }, { "epoch": 2.878539296110082, "grad_norm": 2.0919296741485596, "learning_rate": 3.885464676249345e-05, "loss": 5.5268, "step": 62550 }, { "epoch": 2.88084034193541, "grad_norm": 2.1244289875030518, "learning_rate": 3.878353251629014e-05, "loss": 5.4922, "step": 62600 }, { "epoch": 2.8831413877607375, "grad_norm": 2.20938777923584, "learning_rate": 3.871244215414138e-05, "loss": 5.4315, "step": 62650 }, { "epoch": 2.885442433586065, "grad_norm": 1.9381208419799805, "learning_rate": 3.8641375827425155e-05, "loss": 5.4294, "step": 62700 }, { "epoch": 2.8877434794113928, "grad_norm": 2.0562620162963867, "learning_rate": 3.8570333687468246e-05, "loss": 5.5141, "step": 62750 }, { "epoch": 2.89004452523672, "grad_norm": 2.103546619415283, "learning_rate": 3.8499315885545936e-05, "loss": 5.4004, "step": 62800 }, { "epoch": 2.8923455710620476, "grad_norm": 2.0385918617248535, "learning_rate": 3.8428322572881694e-05, "loss": 5.4527, "step": 62850 }, { "epoch": 2.8946466168873752, "grad_norm": 1.977550983428955, "learning_rate": 3.835735390064682e-05, "loss": 5.463, "step": 62900 }, { "epoch": 2.896947662712703, "grad_norm": 2.7615981101989746, "learning_rate": 3.828641001996018e-05, "loss": 5.3981, "step": 62950 }, { "epoch": 2.8992487085380305, "grad_norm": 1.8082493543624878, "learning_rate": 3.821549108188784e-05, "loss": 5.5473, "step": 63000 }, { "epoch": 2.8992487085380305, "eval_loss": 5.622348308563232, "eval_runtime": 34.1856, "eval_samples_per_second": 11.233, "eval_steps_per_second": 5.616, "eval_tts_loss": 8.661594926281072, "step": 63000 }, { "epoch": 2.901549754363358, "grad_norm": 1.863789677619934, "learning_rate": 3.814459723744272e-05, "loss": 5.4345, "step": 63050 }, { "epoch": 2.903850800188686, "grad_norm": 1.918397068977356, "learning_rate": 3.8073728637584364e-05, "loss": 5.468, "step": 63100 }, { "epoch": 2.9061518460140134, "grad_norm": 2.23038387298584, "learning_rate": 3.8002885433218484e-05, "loss": 5.481, "step": 63150 }, { "epoch": 2.908452891839341, "grad_norm": 1.9242464303970337, "learning_rate": 3.7932067775196786e-05, "loss": 5.3977, "step": 63200 }, { "epoch": 2.9107539376646687, "grad_norm": 2.0655956268310547, "learning_rate": 3.786127581431653e-05, "loss": 5.469, "step": 63250 }, { "epoch": 2.9130549834899964, "grad_norm": 2.5878562927246094, "learning_rate": 3.779050970132032e-05, "loss": 5.4816, "step": 63300 }, { "epoch": 2.9153560293153236, "grad_norm": 2.0930867195129395, "learning_rate": 3.7719769586895644e-05, "loss": 5.4662, "step": 63350 }, { "epoch": 2.917657075140651, "grad_norm": 2.034477710723877, "learning_rate": 3.764905562167468e-05, "loss": 5.4143, "step": 63400 }, { "epoch": 2.919958120965979, "grad_norm": 2.3117949962615967, "learning_rate": 3.757836795623391e-05, "loss": 5.4487, "step": 63450 }, { "epoch": 2.9222591667913065, "grad_norm": 2.2001752853393555, "learning_rate": 3.750770674109379e-05, "loss": 5.4595, "step": 63500 }, { "epoch": 2.924560212616634, "grad_norm": 2.0831422805786133, "learning_rate": 3.743707212671851e-05, "loss": 5.5518, "step": 63550 }, { "epoch": 2.9268612584419618, "grad_norm": 2.453908681869507, "learning_rate": 3.736646426351556e-05, "loss": 5.442, "step": 63600 }, { "epoch": 2.9291623042672894, "grad_norm": 2.2254319190979004, "learning_rate": 3.72958833018355e-05, "loss": 5.4166, "step": 63650 }, { "epoch": 2.931463350092617, "grad_norm": 2.0115628242492676, "learning_rate": 3.7225329391971595e-05, "loss": 5.4603, "step": 63700 }, { "epoch": 2.9337643959179447, "grad_norm": 2.1176035404205322, "learning_rate": 3.715480268415951e-05, "loss": 5.4859, "step": 63750 }, { "epoch": 2.9360654417432723, "grad_norm": 2.4771876335144043, "learning_rate": 3.708430332857698e-05, "loss": 5.4187, "step": 63800 }, { "epoch": 2.9383664875686, "grad_norm": 2.307610273361206, "learning_rate": 3.7013831475343505e-05, "loss": 5.4417, "step": 63850 }, { "epoch": 2.9406675333939276, "grad_norm": 2.129941463470459, "learning_rate": 3.694338727452001e-05, "loss": 5.4669, "step": 63900 }, { "epoch": 2.9429685792192553, "grad_norm": 2.152264356613159, "learning_rate": 3.687297087610857e-05, "loss": 5.4354, "step": 63950 }, { "epoch": 2.945269625044583, "grad_norm": 1.7948946952819824, "learning_rate": 3.680258243005201e-05, "loss": 5.4489, "step": 64000 }, { "epoch": 2.945269625044583, "eval_loss": 5.6132378578186035, "eval_runtime": 34.2421, "eval_samples_per_second": 11.214, "eval_steps_per_second": 5.607, "eval_tts_loss": 8.660041312188223, "step": 64000 }, { "epoch": 2.9475706708699105, "grad_norm": 2.310068130493164, "learning_rate": 3.673222208623367e-05, "loss": 5.4306, "step": 64050 }, { "epoch": 2.949871716695238, "grad_norm": 1.8985226154327393, "learning_rate": 3.666188999447704e-05, "loss": 5.4886, "step": 64100 }, { "epoch": 2.952172762520566, "grad_norm": 1.9499763250350952, "learning_rate": 3.659158630454546e-05, "loss": 5.4375, "step": 64150 }, { "epoch": 2.9544738083458935, "grad_norm": 2.236942768096924, "learning_rate": 3.652131116614176e-05, "loss": 5.4618, "step": 64200 }, { "epoch": 2.9567748541712207, "grad_norm": 1.9788742065429688, "learning_rate": 3.6451064728908005e-05, "loss": 5.4773, "step": 64250 }, { "epoch": 2.9590758999965483, "grad_norm": 2.213575601577759, "learning_rate": 3.638084714242513e-05, "loss": 5.5173, "step": 64300 }, { "epoch": 2.961376945821876, "grad_norm": 2.0651750564575195, "learning_rate": 3.631065855621263e-05, "loss": 5.4846, "step": 64350 }, { "epoch": 2.9636779916472036, "grad_norm": 2.2218382358551025, "learning_rate": 3.624190202182424e-05, "loss": 5.4418, "step": 64400 }, { "epoch": 2.965979037472531, "grad_norm": 1.9353138208389282, "learning_rate": 3.6171771297017563e-05, "loss": 5.4387, "step": 64450 }, { "epoch": 2.968280083297859, "grad_norm": 2.242788791656494, "learning_rate": 3.6101670017681946e-05, "loss": 5.4293, "step": 64500 }, { "epoch": 2.9705811291231865, "grad_norm": 1.8731865882873535, "learning_rate": 3.603159833308924e-05, "loss": 5.5088, "step": 64550 }, { "epoch": 2.972882174948514, "grad_norm": 2.418881893157959, "learning_rate": 3.5961556392448265e-05, "loss": 5.4744, "step": 64600 }, { "epoch": 2.975183220773842, "grad_norm": 1.8968603610992432, "learning_rate": 3.58915443449045e-05, "loss": 5.4569, "step": 64650 }, { "epoch": 2.9774842665991694, "grad_norm": 2.053711175918579, "learning_rate": 3.58215623395398e-05, "loss": 5.4719, "step": 64700 }, { "epoch": 2.979785312424497, "grad_norm": 2.1504952907562256, "learning_rate": 3.575161052537203e-05, "loss": 5.4488, "step": 64750 }, { "epoch": 2.9820863582498247, "grad_norm": 1.9735080003738403, "learning_rate": 3.568168905135475e-05, "loss": 5.3378, "step": 64800 }, { "epoch": 2.984387404075152, "grad_norm": 2.453549861907959, "learning_rate": 3.5611798066376935e-05, "loss": 5.4512, "step": 64850 }, { "epoch": 2.9866884499004795, "grad_norm": 2.467222213745117, "learning_rate": 3.554193771926263e-05, "loss": 5.4896, "step": 64900 }, { "epoch": 2.988989495725807, "grad_norm": 1.9303815364837646, "learning_rate": 3.5472108158770665e-05, "loss": 5.4282, "step": 64950 }, { "epoch": 2.991290541551135, "grad_norm": 2.030698776245117, "learning_rate": 3.5402309533594276e-05, "loss": 5.5008, "step": 65000 }, { "epoch": 2.991290541551135, "eval_loss": 5.609447479248047, "eval_runtime": 32.7857, "eval_samples_per_second": 11.712, "eval_steps_per_second": 5.856, "eval_tts_loss": 8.679980597459995, "step": 65000 }, { "epoch": 2.9935915873764625, "grad_norm": 2.023101568222046, "learning_rate": 3.533254199236084e-05, "loss": 5.4982, "step": 65050 }, { "epoch": 2.99589263320179, "grad_norm": 2.273653745651245, "learning_rate": 3.526280568363155e-05, "loss": 5.4369, "step": 65100 }, { "epoch": 2.9981936790271178, "grad_norm": 1.9517712593078613, "learning_rate": 3.51931007559011e-05, "loss": 5.447, "step": 65150 }, { "epoch": 3.0004602091650656, "grad_norm": 2.0129427909851074, "learning_rate": 3.5123427357597344e-05, "loss": 5.4359, "step": 65200 }, { "epoch": 3.0027612549903933, "grad_norm": 2.1433475017547607, "learning_rate": 3.505378563708101e-05, "loss": 5.1849, "step": 65250 }, { "epoch": 3.005062300815721, "grad_norm": 2.248459577560425, "learning_rate": 3.4984175742645374e-05, "loss": 5.1406, "step": 65300 }, { "epoch": 3.0073633466410485, "grad_norm": 1.8070954084396362, "learning_rate": 3.491459782251593e-05, "loss": 5.1737, "step": 65350 }, { "epoch": 3.009664392466376, "grad_norm": 2.1215107440948486, "learning_rate": 3.484505202485009e-05, "loss": 5.2554, "step": 65400 }, { "epoch": 3.0119654382917034, "grad_norm": 1.9235641956329346, "learning_rate": 3.477553849773687e-05, "loss": 5.2332, "step": 65450 }, { "epoch": 3.014266484117031, "grad_norm": 2.419490098953247, "learning_rate": 3.470605738919657e-05, "loss": 5.1585, "step": 65500 }, { "epoch": 3.0165675299423587, "grad_norm": 2.0951311588287354, "learning_rate": 3.463660884718046e-05, "loss": 5.2394, "step": 65550 }, { "epoch": 3.0188685757676863, "grad_norm": 2.0080130100250244, "learning_rate": 3.456719301957048e-05, "loss": 5.1791, "step": 65600 }, { "epoch": 3.021169621593014, "grad_norm": 2.4206161499023438, "learning_rate": 3.4497810054178864e-05, "loss": 5.2573, "step": 65650 }, { "epoch": 3.0234706674183416, "grad_norm": 2.1025230884552, "learning_rate": 3.442846009874791e-05, "loss": 5.2453, "step": 65700 }, { "epoch": 3.0257717132436692, "grad_norm": 2.4409613609313965, "learning_rate": 3.435914330094959e-05, "loss": 5.1971, "step": 65750 }, { "epoch": 3.028072759068997, "grad_norm": 1.6904783248901367, "learning_rate": 3.428985980838533e-05, "loss": 5.1878, "step": 65800 }, { "epoch": 3.0303738048943245, "grad_norm": 2.728496551513672, "learning_rate": 3.4220609768585574e-05, "loss": 5.2017, "step": 65850 }, { "epoch": 3.032674850719652, "grad_norm": 2.500746250152588, "learning_rate": 3.415139332900957e-05, "loss": 5.1878, "step": 65900 }, { "epoch": 3.03497589654498, "grad_norm": 2.43241024017334, "learning_rate": 3.4082210637045e-05, "loss": 5.1689, "step": 65950 }, { "epoch": 3.0372769423703074, "grad_norm": 2.33240008354187, "learning_rate": 3.4013061840007684e-05, "loss": 5.226, "step": 66000 }, { "epoch": 3.0372769423703074, "eval_loss": 5.651639938354492, "eval_runtime": 33.64, "eval_samples_per_second": 11.415, "eval_steps_per_second": 5.707, "eval_tts_loss": 8.971735107629264, "step": 66000 }, { "epoch": 3.039577988195635, "grad_norm": 2.2935338020324707, "learning_rate": 3.3943947085141306e-05, "loss": 5.217, "step": 66050 }, { "epoch": 3.0418790340209627, "grad_norm": 2.198589324951172, "learning_rate": 3.3874866519616986e-05, "loss": 5.237, "step": 66100 }, { "epoch": 3.0441800798462904, "grad_norm": 1.9840471744537354, "learning_rate": 3.380582029053312e-05, "loss": 5.2196, "step": 66150 }, { "epoch": 3.0464811256716176, "grad_norm": 2.086278200149536, "learning_rate": 3.373680854491493e-05, "loss": 5.1712, "step": 66200 }, { "epoch": 3.048782171496945, "grad_norm": 2.9629907608032227, "learning_rate": 3.366783142971427e-05, "loss": 5.2284, "step": 66250 }, { "epoch": 3.051083217322273, "grad_norm": 2.235788345336914, "learning_rate": 3.359888909180918e-05, "loss": 5.2337, "step": 66300 }, { "epoch": 3.0533842631476005, "grad_norm": 2.0074462890625, "learning_rate": 3.352998167800371e-05, "loss": 5.1967, "step": 66350 }, { "epoch": 3.055685308972928, "grad_norm": 1.99937105178833, "learning_rate": 3.3461109335027506e-05, "loss": 5.2483, "step": 66400 }, { "epoch": 3.0579863547982558, "grad_norm": 2.276930093765259, "learning_rate": 3.339227220953555e-05, "loss": 5.2434, "step": 66450 }, { "epoch": 3.0602874006235834, "grad_norm": 2.292680501937866, "learning_rate": 3.332347044810782e-05, "loss": 5.1313, "step": 66500 }, { "epoch": 3.062588446448911, "grad_norm": 2.2346673011779785, "learning_rate": 3.3254704197249006e-05, "loss": 5.2221, "step": 66550 }, { "epoch": 3.0648894922742387, "grad_norm": 2.0256075859069824, "learning_rate": 3.3185973603388165e-05, "loss": 5.2076, "step": 66600 }, { "epoch": 3.0671905380995663, "grad_norm": 2.891855478286743, "learning_rate": 3.311727881287846e-05, "loss": 5.2096, "step": 66650 }, { "epoch": 3.069491583924894, "grad_norm": 2.2586610317230225, "learning_rate": 3.304861997199679e-05, "loss": 5.166, "step": 66700 }, { "epoch": 3.0717926297502216, "grad_norm": 1.9196617603302002, "learning_rate": 3.2979997226943496e-05, "loss": 5.171, "step": 66750 }, { "epoch": 3.0740936755755492, "grad_norm": 2.212149143218994, "learning_rate": 3.291141072384207e-05, "loss": 5.2251, "step": 66800 }, { "epoch": 3.076394721400877, "grad_norm": 2.19577693939209, "learning_rate": 3.284286060873884e-05, "loss": 5.1944, "step": 66850 }, { "epoch": 3.078695767226204, "grad_norm": 2.422126054763794, "learning_rate": 3.277434702760264e-05, "loss": 5.2308, "step": 66900 }, { "epoch": 3.0809968130515317, "grad_norm": 2.0679380893707275, "learning_rate": 3.27058701263245e-05, "loss": 5.25, "step": 66950 }, { "epoch": 3.0832978588768594, "grad_norm": 1.9856886863708496, "learning_rate": 3.2637430050717374e-05, "loss": 5.2306, "step": 67000 }, { "epoch": 3.0832978588768594, "eval_loss": 5.655904293060303, "eval_runtime": 33.6757, "eval_samples_per_second": 11.403, "eval_steps_per_second": 5.701, "eval_tts_loss": 8.988588783504623, "step": 67000 }, { "epoch": 3.085598904702187, "grad_norm": 2.128262996673584, "learning_rate": 3.2569026946515744e-05, "loss": 5.2399, "step": 67050 }, { "epoch": 3.0878999505275146, "grad_norm": 2.2318902015686035, "learning_rate": 3.250066095937545e-05, "loss": 5.2388, "step": 67100 }, { "epoch": 3.0902009963528423, "grad_norm": 2.4614462852478027, "learning_rate": 3.243233223487323e-05, "loss": 5.2573, "step": 67150 }, { "epoch": 3.09250204217817, "grad_norm": 2.1323800086975098, "learning_rate": 3.236404091850649e-05, "loss": 5.1872, "step": 67200 }, { "epoch": 3.0948030880034976, "grad_norm": 2.0967583656311035, "learning_rate": 3.2295787155692985e-05, "loss": 5.198, "step": 67250 }, { "epoch": 3.097104133828825, "grad_norm": 2.0951972007751465, "learning_rate": 3.22275710917705e-05, "loss": 5.1414, "step": 67300 }, { "epoch": 3.099405179654153, "grad_norm": 2.2159359455108643, "learning_rate": 3.2159392871996574e-05, "loss": 5.2265, "step": 67350 }, { "epoch": 3.1017062254794805, "grad_norm": 2.2702057361602783, "learning_rate": 3.209125264154812e-05, "loss": 5.1164, "step": 67400 }, { "epoch": 3.104007271304808, "grad_norm": 2.1611013412475586, "learning_rate": 3.202315054552117e-05, "loss": 5.2414, "step": 67450 }, { "epoch": 3.1063083171301358, "grad_norm": 2.3985390663146973, "learning_rate": 3.1955086728930584e-05, "loss": 5.228, "step": 67500 }, { "epoch": 3.1086093629554634, "grad_norm": 2.3189857006073, "learning_rate": 3.188842146705822e-05, "loss": 5.228, "step": 67550 }, { "epoch": 3.110910408780791, "grad_norm": 2.1470682621002197, "learning_rate": 3.182043387125502e-05, "loss": 5.211, "step": 67600 }, { "epoch": 3.1132114546061187, "grad_norm": 2.4096908569335938, "learning_rate": 3.175248498654779e-05, "loss": 5.2127, "step": 67650 }, { "epoch": 3.115512500431446, "grad_norm": 2.0130157470703125, "learning_rate": 3.168457495762513e-05, "loss": 5.2123, "step": 67700 }, { "epoch": 3.1178135462567735, "grad_norm": 2.5356950759887695, "learning_rate": 3.161670392909286e-05, "loss": 5.1882, "step": 67750 }, { "epoch": 3.120114592082101, "grad_norm": 2.213794708251953, "learning_rate": 3.1548872045473806e-05, "loss": 5.1411, "step": 67800 }, { "epoch": 3.122415637907429, "grad_norm": 2.2162258625030518, "learning_rate": 3.148107945120743e-05, "loss": 5.1809, "step": 67850 }, { "epoch": 3.1247166837327565, "grad_norm": 2.2594213485717773, "learning_rate": 3.1413326290649514e-05, "loss": 5.1974, "step": 67900 }, { "epoch": 3.127017729558084, "grad_norm": 2.3324978351593018, "learning_rate": 3.134561270807186e-05, "loss": 5.2048, "step": 67950 }, { "epoch": 3.1293187753834117, "grad_norm": 2.50360107421875, "learning_rate": 3.127793884766203e-05, "loss": 5.1518, "step": 68000 }, { "epoch": 3.1293187753834117, "eval_loss": 5.650763988494873, "eval_runtime": 34.0787, "eval_samples_per_second": 11.268, "eval_steps_per_second": 5.634, "eval_tts_loss": 8.96684687412939, "step": 68000 }, { "epoch": 3.1316198212087394, "grad_norm": 2.1465678215026855, "learning_rate": 3.121030485352299e-05, "loss": 5.1121, "step": 68050 }, { "epoch": 3.133920867034067, "grad_norm": 2.2086095809936523, "learning_rate": 3.114271086967279e-05, "loss": 5.2291, "step": 68100 }, { "epoch": 3.1362219128593947, "grad_norm": 2.5457003116607666, "learning_rate": 3.107515704004432e-05, "loss": 5.2126, "step": 68150 }, { "epoch": 3.1385229586847223, "grad_norm": 2.3758766651153564, "learning_rate": 3.100764350848494e-05, "loss": 5.2588, "step": 68200 }, { "epoch": 3.14082400451005, "grad_norm": 2.52122163772583, "learning_rate": 3.0940170418756233e-05, "loss": 5.2151, "step": 68250 }, { "epoch": 3.1431250503353776, "grad_norm": 2.119819164276123, "learning_rate": 3.0872737914533614e-05, "loss": 5.1885, "step": 68300 }, { "epoch": 3.1454260961607052, "grad_norm": 2.662466049194336, "learning_rate": 3.0805346139406126e-05, "loss": 5.2543, "step": 68350 }, { "epoch": 3.1477271419860324, "grad_norm": 2.391465902328491, "learning_rate": 3.073799523687606e-05, "loss": 5.1837, "step": 68400 }, { "epoch": 3.15002818781136, "grad_norm": 2.0778284072875977, "learning_rate": 3.067068535035866e-05, "loss": 5.1969, "step": 68450 }, { "epoch": 3.1523292336366877, "grad_norm": 2.5482311248779297, "learning_rate": 3.060341662318188e-05, "loss": 5.2041, "step": 68500 }, { "epoch": 3.1546302794620154, "grad_norm": 2.060176372528076, "learning_rate": 3.053618919858599e-05, "loss": 5.2056, "step": 68550 }, { "epoch": 3.156931325287343, "grad_norm": 2.577627658843994, "learning_rate": 3.046900321972333e-05, "loss": 5.1266, "step": 68600 }, { "epoch": 3.1592323711126706, "grad_norm": 2.43023943901062, "learning_rate": 3.040185882965796e-05, "loss": 5.1717, "step": 68650 }, { "epoch": 3.1615334169379983, "grad_norm": 2.572831630706787, "learning_rate": 3.0334756171365403e-05, "loss": 5.2253, "step": 68700 }, { "epoch": 3.163834462763326, "grad_norm": 2.3290741443634033, "learning_rate": 3.026769538773232e-05, "loss": 5.2018, "step": 68750 }, { "epoch": 3.1661355085886536, "grad_norm": 2.153881072998047, "learning_rate": 3.0200676621556214e-05, "loss": 5.2132, "step": 68800 }, { "epoch": 3.168436554413981, "grad_norm": 1.900630235671997, "learning_rate": 3.013370001554508e-05, "loss": 5.1798, "step": 68850 }, { "epoch": 3.170737600239309, "grad_norm": 2.3022022247314453, "learning_rate": 3.006676571231719e-05, "loss": 5.2217, "step": 68900 }, { "epoch": 3.1730386460646365, "grad_norm": 2.353461742401123, "learning_rate": 2.9999873854400694e-05, "loss": 5.1732, "step": 68950 }, { "epoch": 3.175339691889964, "grad_norm": 2.6645305156707764, "learning_rate": 2.9933024584233395e-05, "loss": 5.2012, "step": 69000 }, { "epoch": 3.175339691889964, "eval_loss": 5.656210422515869, "eval_runtime": 33.5674, "eval_samples_per_second": 11.44, "eval_steps_per_second": 5.72, "eval_tts_loss": 8.98287127706553, "step": 69000 }, { "epoch": 3.1776407377152918, "grad_norm": 2.240955114364624, "learning_rate": 2.9866218044162358e-05, "loss": 5.2563, "step": 69050 }, { "epoch": 3.1799417835406194, "grad_norm": 1.9387973546981812, "learning_rate": 2.9799454376443735e-05, "loss": 5.2699, "step": 69100 }, { "epoch": 3.182242829365947, "grad_norm": 2.1009185314178467, "learning_rate": 2.9732733723242322e-05, "loss": 5.1805, "step": 69150 }, { "epoch": 3.1845438751912742, "grad_norm": 2.2868292331695557, "learning_rate": 2.9666056226631356e-05, "loss": 5.1707, "step": 69200 }, { "epoch": 3.186844921016602, "grad_norm": 2.269939661026001, "learning_rate": 2.959942202859216e-05, "loss": 5.2238, "step": 69250 }, { "epoch": 3.1891459668419295, "grad_norm": 2.072275400161743, "learning_rate": 2.953283127101386e-05, "loss": 5.1747, "step": 69300 }, { "epoch": 3.191447012667257, "grad_norm": 2.3713185787200928, "learning_rate": 2.94662840956931e-05, "loss": 5.2188, "step": 69350 }, { "epoch": 3.193748058492585, "grad_norm": 2.2737011909484863, "learning_rate": 2.9399780644333695e-05, "loss": 5.2647, "step": 69400 }, { "epoch": 3.1960491043179124, "grad_norm": 2.3678393363952637, "learning_rate": 2.9333321058546372e-05, "loss": 5.2281, "step": 69450 }, { "epoch": 3.19835015014324, "grad_norm": 2.1628551483154297, "learning_rate": 2.926690547984845e-05, "loss": 5.2008, "step": 69500 }, { "epoch": 3.2006511959685677, "grad_norm": 2.6043412685394287, "learning_rate": 2.920053404966352e-05, "loss": 5.2402, "step": 69550 }, { "epoch": 3.2029522417938954, "grad_norm": 2.384507179260254, "learning_rate": 2.9134206909321215e-05, "loss": 5.205, "step": 69600 }, { "epoch": 3.205253287619223, "grad_norm": 2.3745009899139404, "learning_rate": 2.9067924200056774e-05, "loss": 5.2248, "step": 69650 }, { "epoch": 3.2075543334445507, "grad_norm": 1.8591980934143066, "learning_rate": 2.9001686063010953e-05, "loss": 5.1882, "step": 69700 }, { "epoch": 3.2098553792698783, "grad_norm": 2.359219789505005, "learning_rate": 2.893549263922945e-05, "loss": 5.1539, "step": 69750 }, { "epoch": 3.212156425095206, "grad_norm": 2.183582305908203, "learning_rate": 2.886934406966285e-05, "loss": 5.232, "step": 69800 }, { "epoch": 3.2144574709205336, "grad_norm": 2.331078052520752, "learning_rate": 2.88032404951662e-05, "loss": 5.1943, "step": 69850 }, { "epoch": 3.2167585167458608, "grad_norm": 1.92499577999115, "learning_rate": 2.8737182056498728e-05, "loss": 5.1943, "step": 69900 }, { "epoch": 3.2190595625711884, "grad_norm": 2.7040982246398926, "learning_rate": 2.8671168894323562e-05, "loss": 5.212, "step": 69950 }, { "epoch": 3.221360608396516, "grad_norm": 2.5483896732330322, "learning_rate": 2.8605201149207416e-05, "loss": 5.1936, "step": 70000 }, { "epoch": 3.221360608396516, "eval_loss": 5.647762298583984, "eval_runtime": 34.2747, "eval_samples_per_second": 11.204, "eval_steps_per_second": 5.602, "eval_tts_loss": 8.978084567315824, "step": 70000 }, { "epoch": 3.2236616542218437, "grad_norm": 2.2737507820129395, "learning_rate": 2.8539278961620275e-05, "loss": 5.1953, "step": 70050 }, { "epoch": 3.2259627000471713, "grad_norm": 2.2077796459198, "learning_rate": 2.8473402471935164e-05, "loss": 5.1825, "step": 70100 }, { "epoch": 3.228263745872499, "grad_norm": 1.9657273292541504, "learning_rate": 2.8407571820427757e-05, "loss": 5.2468, "step": 70150 }, { "epoch": 3.2305647916978266, "grad_norm": 2.0409305095672607, "learning_rate": 2.8341787147276134e-05, "loss": 5.1942, "step": 70200 }, { "epoch": 3.2328658375231543, "grad_norm": 2.118298053741455, "learning_rate": 2.827604859256046e-05, "loss": 5.1739, "step": 70250 }, { "epoch": 3.235166883348482, "grad_norm": 2.3376784324645996, "learning_rate": 2.821035629626272e-05, "loss": 5.2008, "step": 70300 }, { "epoch": 3.2374679291738095, "grad_norm": 2.356640100479126, "learning_rate": 2.8144710398266373e-05, "loss": 5.2111, "step": 70350 }, { "epoch": 3.239768974999137, "grad_norm": 1.9011934995651245, "learning_rate": 2.80791110383561e-05, "loss": 5.2511, "step": 70400 }, { "epoch": 3.242070020824465, "grad_norm": 2.2594473361968994, "learning_rate": 2.801355835621746e-05, "loss": 5.2381, "step": 70450 }, { "epoch": 3.2443710666497925, "grad_norm": 1.9861502647399902, "learning_rate": 2.794805249143662e-05, "loss": 5.2064, "step": 70500 }, { "epoch": 3.24667211247512, "grad_norm": 2.493450880050659, "learning_rate": 2.788259358350011e-05, "loss": 5.2375, "step": 70550 }, { "epoch": 3.2489731583004477, "grad_norm": 2.1547906398773193, "learning_rate": 2.7817181771794405e-05, "loss": 5.2282, "step": 70600 }, { "epoch": 3.2512742041257754, "grad_norm": 2.2504146099090576, "learning_rate": 2.7751817195605716e-05, "loss": 5.2275, "step": 70650 }, { "epoch": 3.2535752499511026, "grad_norm": 2.092921257019043, "learning_rate": 2.7686499994119674e-05, "loss": 5.2166, "step": 70700 }, { "epoch": 3.2558762957764302, "grad_norm": 2.1429293155670166, "learning_rate": 2.762123030642104e-05, "loss": 5.1866, "step": 70750 }, { "epoch": 3.258177341601758, "grad_norm": 2.242461681365967, "learning_rate": 2.7556008271493406e-05, "loss": 5.2111, "step": 70800 }, { "epoch": 3.2604783874270855, "grad_norm": 2.464775323867798, "learning_rate": 2.7490834028218832e-05, "loss": 5.2219, "step": 70850 }, { "epoch": 3.262779433252413, "grad_norm": 2.140676736831665, "learning_rate": 2.7425707715377667e-05, "loss": 5.2287, "step": 70900 }, { "epoch": 3.265080479077741, "grad_norm": 2.3947184085845947, "learning_rate": 2.7361930564549054e-05, "loss": 5.1812, "step": 70950 }, { "epoch": 3.2673815249030684, "grad_norm": 2.0517752170562744, "learning_rate": 2.7296899562996115e-05, "loss": 5.1908, "step": 71000 }, { "epoch": 3.2673815249030684, "eval_loss": 5.633924961090088, "eval_runtime": 35.1466, "eval_samples_per_second": 10.926, "eval_steps_per_second": 5.463, "eval_tts_loss": 8.900792966671826, "step": 71000 }, { "epoch": 3.269682570728396, "grad_norm": 2.8329784870147705, "learning_rate": 2.7231916904835607e-05, "loss": 5.2412, "step": 71050 }, { "epoch": 3.2719836165537237, "grad_norm": 2.969275712966919, "learning_rate": 2.716698272843994e-05, "loss": 5.2071, "step": 71100 }, { "epoch": 3.2742846623790514, "grad_norm": 2.4925220012664795, "learning_rate": 2.7102097172078234e-05, "loss": 5.2071, "step": 71150 }, { "epoch": 3.276585708204379, "grad_norm": 2.5884768962860107, "learning_rate": 2.7037260373916108e-05, "loss": 5.1944, "step": 71200 }, { "epoch": 3.2788867540297066, "grad_norm": 2.3826942443847656, "learning_rate": 2.6972472472015388e-05, "loss": 5.2356, "step": 71250 }, { "epoch": 3.2811877998550343, "grad_norm": 2.155606269836426, "learning_rate": 2.690773360433373e-05, "loss": 5.2248, "step": 71300 }, { "epoch": 3.2834888456803615, "grad_norm": 2.455932855606079, "learning_rate": 2.6843043908724398e-05, "loss": 5.2249, "step": 71350 }, { "epoch": 3.285789891505689, "grad_norm": 1.9317978620529175, "learning_rate": 2.6778403522935952e-05, "loss": 5.1775, "step": 71400 }, { "epoch": 3.2880909373310168, "grad_norm": 2.2273571491241455, "learning_rate": 2.6713812584611953e-05, "loss": 5.2158, "step": 71450 }, { "epoch": 3.2903919831563444, "grad_norm": 2.5110819339752197, "learning_rate": 2.664927123129065e-05, "loss": 5.2597, "step": 71500 }, { "epoch": 3.292693028981672, "grad_norm": 2.180302619934082, "learning_rate": 2.6584779600404752e-05, "loss": 5.2747, "step": 71550 }, { "epoch": 3.2949940748069997, "grad_norm": 2.3025472164154053, "learning_rate": 2.652033782928104e-05, "loss": 5.206, "step": 71600 }, { "epoch": 3.2972951206323273, "grad_norm": 2.1360127925872803, "learning_rate": 2.6455946055140142e-05, "loss": 5.2514, "step": 71650 }, { "epoch": 3.299596166457655, "grad_norm": 2.336792230606079, "learning_rate": 2.639160441509626e-05, "loss": 5.187, "step": 71700 }, { "epoch": 3.3018972122829826, "grad_norm": 2.4843273162841797, "learning_rate": 2.6327313046156797e-05, "loss": 5.1613, "step": 71750 }, { "epoch": 3.3041982581083102, "grad_norm": 2.6479268074035645, "learning_rate": 2.626307208522212e-05, "loss": 5.1823, "step": 71800 }, { "epoch": 3.306499303933638, "grad_norm": 2.1863152980804443, "learning_rate": 2.619888166908527e-05, "loss": 5.2027, "step": 71850 }, { "epoch": 3.3088003497589655, "grad_norm": 2.1743619441986084, "learning_rate": 2.613474193443166e-05, "loss": 5.1915, "step": 71900 }, { "epoch": 3.311101395584293, "grad_norm": 2.2786686420440674, "learning_rate": 2.6070653017838753e-05, "loss": 5.2057, "step": 71950 }, { "epoch": 3.313402441409621, "grad_norm": 2.755693197250366, "learning_rate": 2.6006615055775822e-05, "loss": 5.1556, "step": 72000 }, { "epoch": 3.313402441409621, "eval_loss": 5.637906551361084, "eval_runtime": 34.4223, "eval_samples_per_second": 11.156, "eval_steps_per_second": 5.578, "eval_tts_loss": 8.920327802164545, "step": 72000 }, { "epoch": 3.3157034872349485, "grad_norm": 2.590268135070801, "learning_rate": 2.594262818460364e-05, "loss": 5.2122, "step": 72050 }, { "epoch": 3.318004533060276, "grad_norm": 2.144188404083252, "learning_rate": 2.5878692540574216e-05, "loss": 5.2054, "step": 72100 }, { "epoch": 3.3203055788856037, "grad_norm": 2.2111215591430664, "learning_rate": 2.5814808259830437e-05, "loss": 5.2366, "step": 72150 }, { "epoch": 3.322606624710931, "grad_norm": 2.2800168991088867, "learning_rate": 2.5750975478405824e-05, "loss": 5.1807, "step": 72200 }, { "epoch": 3.3249076705362586, "grad_norm": 2.367213010787964, "learning_rate": 2.5687194332224252e-05, "loss": 5.2799, "step": 72250 }, { "epoch": 3.327208716361586, "grad_norm": 2.457364559173584, "learning_rate": 2.5623464957099628e-05, "loss": 5.206, "step": 72300 }, { "epoch": 3.329509762186914, "grad_norm": 2.512826442718506, "learning_rate": 2.5559787488735642e-05, "loss": 5.1831, "step": 72350 }, { "epoch": 3.3318108080122415, "grad_norm": 2.583421230316162, "learning_rate": 2.5496162062725427e-05, "loss": 5.2227, "step": 72400 }, { "epoch": 3.334111853837569, "grad_norm": 2.224989891052246, "learning_rate": 2.5432588814551316e-05, "loss": 5.1983, "step": 72450 }, { "epoch": 3.336412899662897, "grad_norm": 2.566269874572754, "learning_rate": 2.536906787958452e-05, "loss": 5.1781, "step": 72500 }, { "epoch": 3.3387139454882244, "grad_norm": 2.3426711559295654, "learning_rate": 2.53055993930849e-05, "loss": 5.1926, "step": 72550 }, { "epoch": 3.341014991313552, "grad_norm": 2.5896639823913574, "learning_rate": 2.524218349020059e-05, "loss": 5.1832, "step": 72600 }, { "epoch": 3.3433160371388797, "grad_norm": 2.0832769870758057, "learning_rate": 2.5178820305967764e-05, "loss": 5.1218, "step": 72650 }, { "epoch": 3.3456170829642073, "grad_norm": 2.4810876846313477, "learning_rate": 2.5115509975310336e-05, "loss": 5.2014, "step": 72700 }, { "epoch": 3.347918128789535, "grad_norm": 2.3670923709869385, "learning_rate": 2.5052252633039685e-05, "loss": 5.1575, "step": 72750 }, { "epoch": 3.3502191746148626, "grad_norm": 2.1796634197235107, "learning_rate": 2.498904841385435e-05, "loss": 5.2025, "step": 72800 }, { "epoch": 3.35252022044019, "grad_norm": 2.492060899734497, "learning_rate": 2.4925897452339758e-05, "loss": 5.1873, "step": 72850 }, { "epoch": 3.3548212662655175, "grad_norm": 2.886307716369629, "learning_rate": 2.486279988296792e-05, "loss": 5.2142, "step": 72900 }, { "epoch": 3.357122312090845, "grad_norm": 2.1251938343048096, "learning_rate": 2.4799755840097167e-05, "loss": 5.1861, "step": 72950 }, { "epoch": 3.3594233579161727, "grad_norm": 2.2697665691375732, "learning_rate": 2.4736765457971844e-05, "loss": 5.2057, "step": 73000 }, { "epoch": 3.3594233579161727, "eval_loss": 5.634683132171631, "eval_runtime": 33.3076, "eval_samples_per_second": 11.529, "eval_steps_per_second": 5.764, "eval_tts_loss": 8.970257855765373, "step": 73000 }, { "epoch": 3.3617244037415004, "grad_norm": 2.1380538940429688, "learning_rate": 2.4673828870722034e-05, "loss": 5.219, "step": 73050 }, { "epoch": 3.364025449566828, "grad_norm": 2.223712205886841, "learning_rate": 2.4610946212363273e-05, "loss": 5.2005, "step": 73100 }, { "epoch": 3.3663264953921557, "grad_norm": 2.7483530044555664, "learning_rate": 2.4548117616796263e-05, "loss": 5.2475, "step": 73150 }, { "epoch": 3.3686275412174833, "grad_norm": 2.250915765762329, "learning_rate": 2.448534321780658e-05, "loss": 5.2745, "step": 73200 }, { "epoch": 3.370928587042811, "grad_norm": 2.6626806259155273, "learning_rate": 2.4422623149064394e-05, "loss": 5.2455, "step": 73250 }, { "epoch": 3.3732296328681386, "grad_norm": 2.2990479469299316, "learning_rate": 2.43599575441242e-05, "loss": 5.1909, "step": 73300 }, { "epoch": 3.3755306786934662, "grad_norm": 2.3052256107330322, "learning_rate": 2.4297346536424486e-05, "loss": 5.1939, "step": 73350 }, { "epoch": 3.377831724518794, "grad_norm": 2.1850342750549316, "learning_rate": 2.423479025928755e-05, "loss": 5.2347, "step": 73400 }, { "epoch": 3.3801327703441215, "grad_norm": 2.586782455444336, "learning_rate": 2.417228884591907e-05, "loss": 5.1781, "step": 73450 }, { "epoch": 3.382433816169449, "grad_norm": 2.611459732055664, "learning_rate": 2.4109842429407946e-05, "loss": 5.2054, "step": 73500 }, { "epoch": 3.384734861994777, "grad_norm": 2.115934133529663, "learning_rate": 2.404745114272596e-05, "loss": 5.194, "step": 73550 }, { "epoch": 3.3870359078201044, "grad_norm": 2.2402496337890625, "learning_rate": 2.3985115118727476e-05, "loss": 5.1598, "step": 73600 }, { "epoch": 3.389336953645432, "grad_norm": 2.0667028427124023, "learning_rate": 2.3922834490149222e-05, "loss": 5.2018, "step": 73650 }, { "epoch": 3.3916379994707593, "grad_norm": 2.5982606410980225, "learning_rate": 2.386060938960995e-05, "loss": 5.1964, "step": 73700 }, { "epoch": 3.393939045296087, "grad_norm": 2.2558162212371826, "learning_rate": 2.3798439949610164e-05, "loss": 5.2415, "step": 73750 }, { "epoch": 3.3962400911214146, "grad_norm": 2.270799160003662, "learning_rate": 2.3736326302531864e-05, "loss": 5.2191, "step": 73800 }, { "epoch": 3.398541136946742, "grad_norm": 2.231250762939453, "learning_rate": 2.367426858063821e-05, "loss": 5.1996, "step": 73850 }, { "epoch": 3.40084218277207, "grad_norm": 2.2880544662475586, "learning_rate": 2.3612266916073344e-05, "loss": 5.1638, "step": 73900 }, { "epoch": 3.4031432285973975, "grad_norm": 1.8324086666107178, "learning_rate": 2.355032144086199e-05, "loss": 5.1729, "step": 73950 }, { "epoch": 3.405444274422725, "grad_norm": 2.1723499298095703, "learning_rate": 2.3488432286909224e-05, "loss": 5.1841, "step": 74000 }, { "epoch": 3.405444274422725, "eval_loss": 5.629993915557861, "eval_runtime": 34.5361, "eval_samples_per_second": 11.119, "eval_steps_per_second": 5.559, "eval_tts_loss": 8.953473981543894, "step": 74000 }, { "epoch": 3.4077453202480528, "grad_norm": 2.0692930221557617, "learning_rate": 2.3426599586000203e-05, "loss": 5.1673, "step": 74050 }, { "epoch": 3.4100463660733804, "grad_norm": 2.055471181869507, "learning_rate": 2.3364823469799895e-05, "loss": 5.1185, "step": 74100 }, { "epoch": 3.412347411898708, "grad_norm": 2.4703516960144043, "learning_rate": 2.330310406985273e-05, "loss": 5.2096, "step": 74150 }, { "epoch": 3.4146484577240357, "grad_norm": 2.161728620529175, "learning_rate": 2.324144151758238e-05, "loss": 5.1937, "step": 74200 }, { "epoch": 3.4169495035493633, "grad_norm": 2.2135536670684814, "learning_rate": 2.3179835944291524e-05, "loss": 5.2268, "step": 74250 }, { "epoch": 3.419250549374691, "grad_norm": 2.2817091941833496, "learning_rate": 2.3118287481161456e-05, "loss": 5.1812, "step": 74300 }, { "epoch": 3.421551595200018, "grad_norm": 2.546320915222168, "learning_rate": 2.3056796259251874e-05, "loss": 5.1771, "step": 74350 }, { "epoch": 3.423852641025346, "grad_norm": 1.982683539390564, "learning_rate": 2.299536240950058e-05, "loss": 5.2161, "step": 74400 }, { "epoch": 3.4261536868506735, "grad_norm": 1.9768283367156982, "learning_rate": 2.293398606272323e-05, "loss": 5.2182, "step": 74450 }, { "epoch": 3.428454732676001, "grad_norm": 2.2743465900421143, "learning_rate": 2.287266734961302e-05, "loss": 5.2402, "step": 74500 }, { "epoch": 3.4307557785013287, "grad_norm": 2.384610176086426, "learning_rate": 2.281140640074042e-05, "loss": 5.2114, "step": 74550 }, { "epoch": 3.4330568243266564, "grad_norm": 2.158867120742798, "learning_rate": 2.2750203346552908e-05, "loss": 5.1912, "step": 74600 }, { "epoch": 3.435357870151984, "grad_norm": 2.0050137042999268, "learning_rate": 2.2689058317374674e-05, "loss": 5.2088, "step": 74650 }, { "epoch": 3.4376589159773117, "grad_norm": 2.704558849334717, "learning_rate": 2.2627971443406344e-05, "loss": 5.1843, "step": 74700 }, { "epoch": 3.4399599618026393, "grad_norm": 2.4487814903259277, "learning_rate": 2.256694285472475e-05, "loss": 5.2682, "step": 74750 }, { "epoch": 3.442261007627967, "grad_norm": 2.0930614471435547, "learning_rate": 2.2505972681282566e-05, "loss": 5.1559, "step": 74800 }, { "epoch": 3.4445620534532946, "grad_norm": 2.091146945953369, "learning_rate": 2.24450610529081e-05, "loss": 5.2461, "step": 74850 }, { "epoch": 3.4468630992786222, "grad_norm": 2.681933641433716, "learning_rate": 2.2384208099304994e-05, "loss": 5.2385, "step": 74900 }, { "epoch": 3.44916414510395, "grad_norm": 2.2790756225585938, "learning_rate": 2.232341395005194e-05, "loss": 5.2283, "step": 74950 }, { "epoch": 3.4514651909292775, "grad_norm": 2.6264398097991943, "learning_rate": 2.226267873460242e-05, "loss": 5.222, "step": 75000 }, { "epoch": 3.4514651909292775, "eval_loss": 5.62860631942749, "eval_runtime": 33.0203, "eval_samples_per_second": 11.629, "eval_steps_per_second": 5.815, "eval_tts_loss": 8.956448180894176, "step": 75000 }, { "epoch": 3.453766236754605, "grad_norm": 2.6697559356689453, "learning_rate": 2.220200258228444e-05, "loss": 5.183, "step": 75050 }, { "epoch": 3.456067282579933, "grad_norm": 2.3682973384857178, "learning_rate": 2.21413856223002e-05, "loss": 5.2183, "step": 75100 }, { "epoch": 3.4583683284052604, "grad_norm": 2.0403831005096436, "learning_rate": 2.208082798372589e-05, "loss": 5.2039, "step": 75150 }, { "epoch": 3.4606693742305876, "grad_norm": 2.470094919204712, "learning_rate": 2.2022748579976672e-05, "loss": 5.1217, "step": 75200 }, { "epoch": 3.4629704200559153, "grad_norm": 2.267977476119995, "learning_rate": 2.196230758530608e-05, "loss": 5.2224, "step": 75250 }, { "epoch": 3.465271465881243, "grad_norm": 2.673049211502075, "learning_rate": 2.1901926293369545e-05, "loss": 5.1764, "step": 75300 }, { "epoch": 3.4675725117065705, "grad_norm": 2.249359607696533, "learning_rate": 2.184160483274142e-05, "loss": 5.1799, "step": 75350 }, { "epoch": 3.469873557531898, "grad_norm": 2.004697561264038, "learning_rate": 2.178134333186865e-05, "loss": 5.1787, "step": 75400 }, { "epoch": 3.472174603357226, "grad_norm": 2.118774890899658, "learning_rate": 2.1721141919070507e-05, "loss": 5.1152, "step": 75450 }, { "epoch": 3.4744756491825535, "grad_norm": 2.2793705463409424, "learning_rate": 2.1661000722538326e-05, "loss": 5.1887, "step": 75500 }, { "epoch": 3.476776695007881, "grad_norm": 2.111333131790161, "learning_rate": 2.1600919870335214e-05, "loss": 5.2267, "step": 75550 }, { "epoch": 3.4790777408332088, "grad_norm": 1.9498274326324463, "learning_rate": 2.1540899490395755e-05, "loss": 5.1931, "step": 75600 }, { "epoch": 3.4813787866585364, "grad_norm": 2.229842185974121, "learning_rate": 2.1480939710525795e-05, "loss": 5.1317, "step": 75650 }, { "epoch": 3.483679832483864, "grad_norm": 2.236582040786743, "learning_rate": 2.142104065840215e-05, "loss": 5.202, "step": 75700 }, { "epoch": 3.4859808783091917, "grad_norm": 2.623948812484741, "learning_rate": 2.136120246157229e-05, "loss": 5.1442, "step": 75750 }, { "epoch": 3.4882819241345193, "grad_norm": 2.466200113296509, "learning_rate": 2.1301425247454127e-05, "loss": 5.1478, "step": 75800 }, { "epoch": 3.4905829699598465, "grad_norm": 2.081251382827759, "learning_rate": 2.1241709143335696e-05, "loss": 5.2129, "step": 75850 }, { "epoch": 3.492884015785174, "grad_norm": 2.99832820892334, "learning_rate": 2.118205427637493e-05, "loss": 5.2158, "step": 75900 }, { "epoch": 3.495185061610502, "grad_norm": 1.9218469858169556, "learning_rate": 2.1122460773599345e-05, "loss": 5.2348, "step": 75950 }, { "epoch": 3.4974861074358294, "grad_norm": 2.370565176010132, "learning_rate": 2.1062928761905805e-05, "loss": 5.2643, "step": 76000 }, { "epoch": 3.4974861074358294, "eval_loss": 5.631509780883789, "eval_runtime": 33.4738, "eval_samples_per_second": 11.472, "eval_steps_per_second": 5.736, "eval_tts_loss": 8.98740807562086, "step": 76000 }, { "epoch": 3.499787153261157, "grad_norm": 2.2371163368225098, "learning_rate": 2.100345836806022e-05, "loss": 5.2018, "step": 76050 }, { "epoch": 3.5020881990864847, "grad_norm": 2.599576950073242, "learning_rate": 2.094404971869731e-05, "loss": 5.1568, "step": 76100 }, { "epoch": 3.5043892449118124, "grad_norm": 2.2047317028045654, "learning_rate": 2.088470294032029e-05, "loss": 5.2178, "step": 76150 }, { "epoch": 3.50669029073714, "grad_norm": 2.1263697147369385, "learning_rate": 2.0826603246530247e-05, "loss": 5.2013, "step": 76200 }, { "epoch": 3.5089913365624676, "grad_norm": 2.0238592624664307, "learning_rate": 2.076737934539928e-05, "loss": 5.1675, "step": 76250 }, { "epoch": 3.5112923823877953, "grad_norm": 2.288980722427368, "learning_rate": 2.0708217691451527e-05, "loss": 5.2044, "step": 76300 }, { "epoch": 3.513593428213123, "grad_norm": 2.3009746074676514, "learning_rate": 2.0649118410664274e-05, "loss": 5.2159, "step": 76350 }, { "epoch": 3.5158944740384506, "grad_norm": 2.0506417751312256, "learning_rate": 2.0590081628881996e-05, "loss": 5.1579, "step": 76400 }, { "epoch": 3.518195519863778, "grad_norm": 2.861734628677368, "learning_rate": 2.0531107471816084e-05, "loss": 5.1541, "step": 76450 }, { "epoch": 3.520496565689106, "grad_norm": 2.453308343887329, "learning_rate": 2.047219606504458e-05, "loss": 5.1878, "step": 76500 }, { "epoch": 3.5227976115144335, "grad_norm": 2.1938374042510986, "learning_rate": 2.0413347534011896e-05, "loss": 5.2198, "step": 76550 }, { "epoch": 3.525098657339761, "grad_norm": 2.3495683670043945, "learning_rate": 2.0354562004028566e-05, "loss": 5.2011, "step": 76600 }, { "epoch": 3.5273997031650888, "grad_norm": 2.0548179149627686, "learning_rate": 2.0295839600270972e-05, "loss": 5.2454, "step": 76650 }, { "epoch": 3.5297007489904164, "grad_norm": 2.2966115474700928, "learning_rate": 2.0237180447781083e-05, "loss": 5.2069, "step": 76700 }, { "epoch": 3.5320017948157436, "grad_norm": 2.0031216144561768, "learning_rate": 2.0178584671466165e-05, "loss": 5.2371, "step": 76750 }, { "epoch": 3.5343028406410713, "grad_norm": 2.0526063442230225, "learning_rate": 2.0120052396098528e-05, "loss": 5.2393, "step": 76800 }, { "epoch": 3.536603886466399, "grad_norm": 2.4385509490966797, "learning_rate": 2.0061583746315317e-05, "loss": 5.1446, "step": 76850 }, { "epoch": 3.5389049322917265, "grad_norm": 2.629765510559082, "learning_rate": 2.0003178846618137e-05, "loss": 5.2291, "step": 76900 }, { "epoch": 3.541205978117054, "grad_norm": 2.386939525604248, "learning_rate": 1.9944837821372857e-05, "loss": 5.2247, "step": 76950 }, { "epoch": 3.543507023942382, "grad_norm": 2.1514835357666016, "learning_rate": 1.9886560794809355e-05, "loss": 5.1999, "step": 77000 }, { "epoch": 3.543507023942382, "eval_loss": 5.626565456390381, "eval_runtime": 32.058, "eval_samples_per_second": 11.978, "eval_steps_per_second": 5.989, "eval_tts_loss": 8.992654431853728, "step": 77000 }, { "epoch": 3.5458080697677095, "grad_norm": 2.399937868118286, "learning_rate": 1.982834789102121e-05, "loss": 5.1952, "step": 77050 }, { "epoch": 3.548109115593037, "grad_norm": 2.338700532913208, "learning_rate": 1.977019923396547e-05, "loss": 5.1768, "step": 77100 }, { "epoch": 3.5504101614183647, "grad_norm": 2.1986091136932373, "learning_rate": 1.9712114947462366e-05, "loss": 5.1892, "step": 77150 }, { "epoch": 3.5527112072436924, "grad_norm": 2.2410266399383545, "learning_rate": 1.9654095155195086e-05, "loss": 5.1718, "step": 77200 }, { "epoch": 3.55501225306902, "grad_norm": 2.3475921154022217, "learning_rate": 1.9596139980709448e-05, "loss": 5.2131, "step": 77250 }, { "epoch": 3.557313298894347, "grad_norm": 2.4052350521087646, "learning_rate": 1.95382495474137e-05, "loss": 5.2332, "step": 77300 }, { "epoch": 3.559614344719675, "grad_norm": 2.216869831085205, "learning_rate": 1.9480423978578245e-05, "loss": 5.1678, "step": 77350 }, { "epoch": 3.5619153905450025, "grad_norm": 2.421912908554077, "learning_rate": 1.9422663397335354e-05, "loss": 5.1614, "step": 77400 }, { "epoch": 3.56421643637033, "grad_norm": 2.5841050148010254, "learning_rate": 1.9364967926678877e-05, "loss": 5.2593, "step": 77450 }, { "epoch": 3.566517482195658, "grad_norm": 1.9473857879638672, "learning_rate": 1.9307337689464065e-05, "loss": 5.2039, "step": 77500 }, { "epoch": 3.5688185280209854, "grad_norm": 2.1319632530212402, "learning_rate": 1.9249772808407245e-05, "loss": 5.1229, "step": 77550 }, { "epoch": 3.571119573846313, "grad_norm": 2.149664878845215, "learning_rate": 1.9192273406085588e-05, "loss": 5.2108, "step": 77600 }, { "epoch": 3.5734206196716407, "grad_norm": 2.748605966567993, "learning_rate": 1.91348396049368e-05, "loss": 5.1874, "step": 77650 }, { "epoch": 3.5757216654969683, "grad_norm": 2.3713808059692383, "learning_rate": 1.9077471527258967e-05, "loss": 5.1959, "step": 77700 }, { "epoch": 3.578022711322296, "grad_norm": 2.1292314529418945, "learning_rate": 1.9020169295210162e-05, "loss": 5.2464, "step": 77750 }, { "epoch": 3.5803237571476236, "grad_norm": 2.0167505741119385, "learning_rate": 1.8962933030808277e-05, "loss": 5.1975, "step": 77800 }, { "epoch": 3.5826248029729513, "grad_norm": 2.4200239181518555, "learning_rate": 1.890576285593072e-05, "loss": 5.1826, "step": 77850 }, { "epoch": 3.584925848798279, "grad_norm": 2.4658398628234863, "learning_rate": 1.884865889231418e-05, "loss": 5.1882, "step": 77900 }, { "epoch": 3.5872268946236066, "grad_norm": 3.0140624046325684, "learning_rate": 1.8791621261554354e-05, "loss": 5.197, "step": 77950 }, { "epoch": 3.589527940448934, "grad_norm": 2.1732044219970703, "learning_rate": 1.873465008510569e-05, "loss": 5.215, "step": 78000 }, { "epoch": 3.589527940448934, "eval_loss": 5.619250774383545, "eval_runtime": 34.2668, "eval_samples_per_second": 11.206, "eval_steps_per_second": 5.603, "eval_tts_loss": 8.996872534198715, "step": 78000 }, { "epoch": 3.591828986274262, "grad_norm": 2.3192954063415527, "learning_rate": 1.8677745484281135e-05, "loss": 5.243, "step": 78050 }, { "epoch": 3.5941300320995895, "grad_norm": 2.143805742263794, "learning_rate": 1.8620907580251863e-05, "loss": 5.1755, "step": 78100 }, { "epoch": 3.596431077924917, "grad_norm": 2.305589437484741, "learning_rate": 1.856413649404702e-05, "loss": 5.2402, "step": 78150 }, { "epoch": 3.5987321237502448, "grad_norm": 2.2067654132843018, "learning_rate": 1.8507432346553504e-05, "loss": 5.1785, "step": 78200 }, { "epoch": 3.601033169575572, "grad_norm": 2.503913640975952, "learning_rate": 1.845079525851565e-05, "loss": 5.2286, "step": 78250 }, { "epoch": 3.6033342154008996, "grad_norm": 2.5515477657318115, "learning_rate": 1.8394225350535006e-05, "loss": 5.1196, "step": 78300 }, { "epoch": 3.6056352612262272, "grad_norm": 2.8195738792419434, "learning_rate": 1.8337722743070047e-05, "loss": 5.2123, "step": 78350 }, { "epoch": 3.607936307051555, "grad_norm": 2.128828525543213, "learning_rate": 1.8281287556435972e-05, "loss": 5.2335, "step": 78400 }, { "epoch": 3.6102373528768825, "grad_norm": 2.656996726989746, "learning_rate": 1.8224919910804395e-05, "loss": 5.204, "step": 78450 }, { "epoch": 3.61253839870221, "grad_norm": 2.313681125640869, "learning_rate": 1.8168619926203124e-05, "loss": 5.193, "step": 78500 }, { "epoch": 3.614839444527538, "grad_norm": 2.0406699180603027, "learning_rate": 1.8112387722515872e-05, "loss": 5.2394, "step": 78550 }, { "epoch": 3.6171404903528654, "grad_norm": 2.2279775142669678, "learning_rate": 1.8056223419482043e-05, "loss": 5.179, "step": 78600 }, { "epoch": 3.619441536178193, "grad_norm": 2.2751965522766113, "learning_rate": 1.8000127136696448e-05, "loss": 5.231, "step": 78650 }, { "epoch": 3.6217425820035207, "grad_norm": 2.374161958694458, "learning_rate": 1.7944098993609042e-05, "loss": 5.1862, "step": 78700 }, { "epoch": 3.6240436278288484, "grad_norm": 2.4539670944213867, "learning_rate": 1.788813910952471e-05, "loss": 5.2105, "step": 78750 }, { "epoch": 3.6263446736541756, "grad_norm": 2.1793766021728516, "learning_rate": 1.783224760360297e-05, "loss": 5.1908, "step": 78800 }, { "epoch": 3.628645719479503, "grad_norm": 2.398728132247925, "learning_rate": 1.7776424594857738e-05, "loss": 5.2523, "step": 78850 }, { "epoch": 3.630946765304831, "grad_norm": 2.1088905334472656, "learning_rate": 1.772067020215709e-05, "loss": 5.1547, "step": 78900 }, { "epoch": 3.6332478111301585, "grad_norm": 1.9953407049179077, "learning_rate": 1.766498454422298e-05, "loss": 5.1972, "step": 78950 }, { "epoch": 3.635548856955486, "grad_norm": 2.152448892593384, "learning_rate": 1.7609367739630976e-05, "loss": 5.2446, "step": 79000 }, { "epoch": 3.635548856955486, "eval_loss": 5.617215633392334, "eval_runtime": 32.3456, "eval_samples_per_second": 11.872, "eval_steps_per_second": 5.936, "eval_tts_loss": 8.97757097060575, "step": 79000 }, { "epoch": 3.6378499027808138, "grad_norm": 2.023789644241333, "learning_rate": 1.75538199068101e-05, "loss": 5.19, "step": 79050 }, { "epoch": 3.6401509486061414, "grad_norm": 2.7019007205963135, "learning_rate": 1.7498341164042442e-05, "loss": 5.2325, "step": 79100 }, { "epoch": 3.642451994431469, "grad_norm": 1.858165979385376, "learning_rate": 1.7442931629463e-05, "loss": 5.1992, "step": 79150 }, { "epoch": 3.6447530402567967, "grad_norm": 2.1814475059509277, "learning_rate": 1.7387591421059395e-05, "loss": 5.2014, "step": 79200 }, { "epoch": 3.6470540860821243, "grad_norm": 2.373379707336426, "learning_rate": 1.733232065667163e-05, "loss": 5.1136, "step": 79250 }, { "epoch": 3.649355131907452, "grad_norm": 2.069084644317627, "learning_rate": 1.7277119453991834e-05, "loss": 5.1485, "step": 79300 }, { "epoch": 3.6516561777327796, "grad_norm": 2.114830255508423, "learning_rate": 1.722198793056402e-05, "loss": 5.19, "step": 79350 }, { "epoch": 3.6539572235581073, "grad_norm": 1.8629753589630127, "learning_rate": 1.7166926203783822e-05, "loss": 5.1707, "step": 79400 }, { "epoch": 3.656258269383435, "grad_norm": 2.268953323364258, "learning_rate": 1.7111934390898225e-05, "loss": 5.2116, "step": 79450 }, { "epoch": 3.6585593152087625, "grad_norm": 2.427870988845825, "learning_rate": 1.7057012609005406e-05, "loss": 5.1697, "step": 79500 }, { "epoch": 3.66086036103409, "grad_norm": 2.392437696456909, "learning_rate": 1.7002160975054366e-05, "loss": 5.1588, "step": 79550 }, { "epoch": 3.663161406859418, "grad_norm": 2.81585955619812, "learning_rate": 1.6947379605844766e-05, "loss": 5.1922, "step": 79600 }, { "epoch": 3.6654624526847455, "grad_norm": 2.525327205657959, "learning_rate": 1.6892668618026596e-05, "loss": 5.2282, "step": 79650 }, { "epoch": 3.6677634985100727, "grad_norm": 2.584698438644409, "learning_rate": 1.6838028128100032e-05, "loss": 5.1869, "step": 79700 }, { "epoch": 3.6700645443354003, "grad_norm": 2.473557710647583, "learning_rate": 1.6783458252415108e-05, "loss": 5.2196, "step": 79750 }, { "epoch": 3.672365590160728, "grad_norm": 2.360694408416748, "learning_rate": 1.6728959107171504e-05, "loss": 5.1844, "step": 79800 }, { "epoch": 3.6746666359860556, "grad_norm": 2.617858648300171, "learning_rate": 1.667453080841827e-05, "loss": 5.1233, "step": 79850 }, { "epoch": 3.6769676818113832, "grad_norm": 2.8224070072174072, "learning_rate": 1.662017347205364e-05, "loss": 5.2381, "step": 79900 }, { "epoch": 3.679268727636711, "grad_norm": 1.9658337831497192, "learning_rate": 1.6565887213824704e-05, "loss": 5.1793, "step": 79950 }, { "epoch": 3.6815697734620385, "grad_norm": 2.1769447326660156, "learning_rate": 1.6511672149327206e-05, "loss": 5.1578, "step": 80000 }, { "epoch": 3.6815697734620385, "eval_loss": 5.613487243652344, "eval_runtime": 33.8101, "eval_samples_per_second": 11.358, "eval_steps_per_second": 5.679, "eval_tts_loss": 8.99393554956893, "step": 80000 }, { "epoch": 3.683870819287366, "grad_norm": 2.6718637943267822, "learning_rate": 1.6457528394005296e-05, "loss": 5.2155, "step": 80050 }, { "epoch": 3.686171865112694, "grad_norm": 2.3300375938415527, "learning_rate": 1.640345606315128e-05, "loss": 5.2027, "step": 80100 }, { "epoch": 3.6884729109380214, "grad_norm": 2.3993618488311768, "learning_rate": 1.634945527190538e-05, "loss": 5.2122, "step": 80150 }, { "epoch": 3.690773956763349, "grad_norm": 2.0944197177886963, "learning_rate": 1.6295526135255472e-05, "loss": 5.2506, "step": 80200 }, { "epoch": 3.6930750025886767, "grad_norm": 2.1264808177948, "learning_rate": 1.6241668768036862e-05, "loss": 5.1621, "step": 80250 }, { "epoch": 3.695376048414004, "grad_norm": 2.0797226428985596, "learning_rate": 1.6187883284932015e-05, "loss": 5.1699, "step": 80300 }, { "epoch": 3.6976770942393316, "grad_norm": 2.431781768798828, "learning_rate": 1.613416980047037e-05, "loss": 5.2216, "step": 80350 }, { "epoch": 3.699978140064659, "grad_norm": 2.323768377304077, "learning_rate": 1.6080528429028015e-05, "loss": 5.2144, "step": 80400 }, { "epoch": 3.702279185889987, "grad_norm": 2.5759477615356445, "learning_rate": 1.6026959284827485e-05, "loss": 5.2031, "step": 80450 }, { "epoch": 3.7045802317153145, "grad_norm": 2.581083297729492, "learning_rate": 1.597346248193753e-05, "loss": 5.181, "step": 80500 }, { "epoch": 3.706881277540642, "grad_norm": 1.9495331048965454, "learning_rate": 1.5920038134272853e-05, "loss": 5.1314, "step": 80550 }, { "epoch": 3.7091823233659698, "grad_norm": 2.2382566928863525, "learning_rate": 1.5866686355593874e-05, "loss": 5.2137, "step": 80600 }, { "epoch": 3.7114833691912974, "grad_norm": 2.1297311782836914, "learning_rate": 1.581340725950648e-05, "loss": 5.2077, "step": 80650 }, { "epoch": 3.713784415016625, "grad_norm": 2.0254764556884766, "learning_rate": 1.5760200959461797e-05, "loss": 5.2052, "step": 80700 }, { "epoch": 3.7160854608419527, "grad_norm": 2.093926191329956, "learning_rate": 1.5707067568755933e-05, "loss": 5.1405, "step": 80750 }, { "epoch": 3.7183865066672803, "grad_norm": 2.399249792098999, "learning_rate": 1.5654007200529763e-05, "loss": 5.1569, "step": 80800 }, { "epoch": 3.720687552492608, "grad_norm": 2.097546100616455, "learning_rate": 1.5601019967768648e-05, "loss": 5.2015, "step": 80850 }, { "epoch": 3.7229885983179356, "grad_norm": 2.278393268585205, "learning_rate": 1.5548105983302224e-05, "loss": 5.1624, "step": 80900 }, { "epoch": 3.7252896441432632, "grad_norm": 2.1321325302124023, "learning_rate": 1.549526535980416e-05, "loss": 5.2292, "step": 80950 }, { "epoch": 3.727590689968591, "grad_norm": 2.526561737060547, "learning_rate": 1.5442498209791905e-05, "loss": 5.2001, "step": 81000 }, { "epoch": 3.727590689968591, "eval_loss": 5.611248016357422, "eval_runtime": 33.0436, "eval_samples_per_second": 11.621, "eval_steps_per_second": 5.811, "eval_tts_loss": 8.977350676607491, "step": 81000 }, { "epoch": 3.7298917357939185, "grad_norm": 2.070404052734375, "learning_rate": 1.5389804645626453e-05, "loss": 5.1921, "step": 81050 }, { "epoch": 3.732192781619246, "grad_norm": 2.317744255065918, "learning_rate": 1.5337184779512125e-05, "loss": 5.1866, "step": 81100 }, { "epoch": 3.734493827444574, "grad_norm": 1.896338701248169, "learning_rate": 1.5284638723496268e-05, "loss": 5.2012, "step": 81150 }, { "epoch": 3.736794873269901, "grad_norm": 2.6844286918640137, "learning_rate": 1.5232166589469127e-05, "loss": 5.1391, "step": 81200 }, { "epoch": 3.7390959190952286, "grad_norm": 2.197406768798828, "learning_rate": 1.5179768489163481e-05, "loss": 5.1941, "step": 81250 }, { "epoch": 3.7413969649205563, "grad_norm": 2.0176520347595215, "learning_rate": 1.512744453415449e-05, "loss": 5.1764, "step": 81300 }, { "epoch": 3.743698010745884, "grad_norm": 1.9290133714675903, "learning_rate": 1.5075194835859418e-05, "loss": 5.1211, "step": 81350 }, { "epoch": 3.7459990565712116, "grad_norm": 1.9766486883163452, "learning_rate": 1.5023019505537422e-05, "loss": 5.1829, "step": 81400 }, { "epoch": 3.748300102396539, "grad_norm": 2.4024932384490967, "learning_rate": 1.4970918654289296e-05, "loss": 5.1121, "step": 81450 }, { "epoch": 3.750601148221867, "grad_norm": 2.7011728286743164, "learning_rate": 1.4918892393057227e-05, "loss": 5.1788, "step": 81500 }, { "epoch": 3.7529021940471945, "grad_norm": 2.2868247032165527, "learning_rate": 1.4866940832624598e-05, "loss": 5.1545, "step": 81550 }, { "epoch": 3.755203239872522, "grad_norm": 2.1724374294281006, "learning_rate": 1.481506408361571e-05, "loss": 5.1113, "step": 81600 }, { "epoch": 3.7575042856978498, "grad_norm": 2.5170412063598633, "learning_rate": 1.4763262256495547e-05, "loss": 5.1083, "step": 81650 }, { "epoch": 3.7598053315231774, "grad_norm": 2.5945959091186523, "learning_rate": 1.4711535461569609e-05, "loss": 5.0641, "step": 81700 }, { "epoch": 3.762106377348505, "grad_norm": 2.386784791946411, "learning_rate": 1.4659883808983565e-05, "loss": 5.1631, "step": 81750 }, { "epoch": 3.7644074231738323, "grad_norm": 2.390890598297119, "learning_rate": 1.4608307408723126e-05, "loss": 5.1748, "step": 81800 }, { "epoch": 3.76670846899916, "grad_norm": 2.576148271560669, "learning_rate": 1.4556806370613707e-05, "loss": 5.1687, "step": 81850 }, { "epoch": 3.7690095148244875, "grad_norm": 2.564168691635132, "learning_rate": 1.4505380804320295e-05, "loss": 5.133, "step": 81900 }, { "epoch": 3.771310560649815, "grad_norm": 2.1351702213287354, "learning_rate": 1.4454030819347158e-05, "loss": 5.2315, "step": 81950 }, { "epoch": 3.773611606475143, "grad_norm": 2.4622039794921875, "learning_rate": 1.440275652503762e-05, "loss": 5.1477, "step": 82000 }, { "epoch": 3.773611606475143, "eval_loss": 5.609696865081787, "eval_runtime": 34.3084, "eval_samples_per_second": 11.193, "eval_steps_per_second": 5.596, "eval_tts_loss": 8.997026660141712, "step": 82000 }, { "epoch": 3.7759126523004705, "grad_norm": 2.6888697147369385, "learning_rate": 1.4351558030573814e-05, "loss": 5.168, "step": 82050 }, { "epoch": 3.778213698125798, "grad_norm": 2.8415167331695557, "learning_rate": 1.4300435444976513e-05, "loss": 5.2276, "step": 82100 }, { "epoch": 3.7805147439511257, "grad_norm": 2.3651936054229736, "learning_rate": 1.4249388877104814e-05, "loss": 5.185, "step": 82150 }, { "epoch": 3.7828157897764534, "grad_norm": 2.442162275314331, "learning_rate": 1.4198418435655947e-05, "loss": 5.1614, "step": 82200 }, { "epoch": 3.785116835601781, "grad_norm": 1.8948955535888672, "learning_rate": 1.4147524229165049e-05, "loss": 5.1829, "step": 82250 }, { "epoch": 3.7874178814271087, "grad_norm": 2.331746816635132, "learning_rate": 1.409670636600492e-05, "loss": 5.1549, "step": 82300 }, { "epoch": 3.7897189272524363, "grad_norm": 2.317063093185425, "learning_rate": 1.4045964954385804e-05, "loss": 5.1202, "step": 82350 }, { "epoch": 3.792019973077764, "grad_norm": 2.8078041076660156, "learning_rate": 1.3995300102355136e-05, "loss": 5.1692, "step": 82400 }, { "epoch": 3.7943210189030916, "grad_norm": 2.278687000274658, "learning_rate": 1.3945722929450245e-05, "loss": 5.1753, "step": 82450 }, { "epoch": 3.7966220647284192, "grad_norm": 2.2202608585357666, "learning_rate": 1.3895209983528002e-05, "loss": 5.2014, "step": 82500 }, { "epoch": 3.798923110553747, "grad_norm": 2.7063558101654053, "learning_rate": 1.3844773918207926e-05, "loss": 5.1813, "step": 82550 }, { "epoch": 3.8012241563790745, "grad_norm": 2.3519999980926514, "learning_rate": 1.379441484088727e-05, "loss": 5.2055, "step": 82600 }, { "epoch": 3.803525202204402, "grad_norm": 2.1331920623779297, "learning_rate": 1.3744132858799346e-05, "loss": 5.1639, "step": 82650 }, { "epoch": 3.8058262480297294, "grad_norm": 2.179452657699585, "learning_rate": 1.3693928079013301e-05, "loss": 5.1549, "step": 82700 }, { "epoch": 3.808127293855057, "grad_norm": 2.223172903060913, "learning_rate": 1.3643800608433888e-05, "loss": 5.2003, "step": 82750 }, { "epoch": 3.8104283396803846, "grad_norm": 2.3072891235351562, "learning_rate": 1.3593750553801227e-05, "loss": 5.2101, "step": 82800 }, { "epoch": 3.8127293855057123, "grad_norm": 2.3883447647094727, "learning_rate": 1.3543778021690629e-05, "loss": 5.1857, "step": 82850 }, { "epoch": 3.81503043133104, "grad_norm": 2.0408358573913574, "learning_rate": 1.3493883118512313e-05, "loss": 5.2079, "step": 82900 }, { "epoch": 3.8173314771563676, "grad_norm": 2.2231454849243164, "learning_rate": 1.3444065950511159e-05, "loss": 5.0707, "step": 82950 }, { "epoch": 3.819632522981695, "grad_norm": 2.178635358810425, "learning_rate": 1.3394326623766563e-05, "loss": 5.179, "step": 83000 }, { "epoch": 3.819632522981695, "eval_loss": 5.609046459197998, "eval_runtime": 35.4306, "eval_samples_per_second": 10.838, "eval_steps_per_second": 5.419, "eval_tts_loss": 9.01051702197809, "step": 83000 }, { "epoch": 3.821933568807023, "grad_norm": 2.3109848499298096, "learning_rate": 1.3344665244192155e-05, "loss": 5.2137, "step": 83050 }, { "epoch": 3.8242346146323505, "grad_norm": 2.1816065311431885, "learning_rate": 1.329508191753559e-05, "loss": 5.136, "step": 83100 }, { "epoch": 3.826535660457678, "grad_norm": 2.9564876556396484, "learning_rate": 1.3245576749378297e-05, "loss": 5.1775, "step": 83150 }, { "epoch": 3.8288367062830058, "grad_norm": 2.3853161334991455, "learning_rate": 1.319614984513533e-05, "loss": 5.2273, "step": 83200 }, { "epoch": 3.831137752108333, "grad_norm": 2.4375455379486084, "learning_rate": 1.314680131005504e-05, "loss": 5.0597, "step": 83250 }, { "epoch": 3.8334387979336606, "grad_norm": 2.3408586978912354, "learning_rate": 1.3097531249218914e-05, "loss": 5.1273, "step": 83300 }, { "epoch": 3.8357398437589882, "grad_norm": 2.303234815597534, "learning_rate": 1.304833976754134e-05, "loss": 5.2115, "step": 83350 }, { "epoch": 3.838040889584316, "grad_norm": 2.0342304706573486, "learning_rate": 1.2999226969769379e-05, "loss": 5.1762, "step": 83400 }, { "epoch": 3.8403419354096435, "grad_norm": 2.351905584335327, "learning_rate": 1.295019296048255e-05, "loss": 5.18, "step": 83450 }, { "epoch": 3.842642981234971, "grad_norm": 2.341549873352051, "learning_rate": 1.29012378440926e-05, "loss": 5.2408, "step": 83500 }, { "epoch": 3.844944027060299, "grad_norm": 2.472702741622925, "learning_rate": 1.2852361724843282e-05, "loss": 5.1864, "step": 83550 }, { "epoch": 3.8472450728856264, "grad_norm": 2.308812141418457, "learning_rate": 1.2803564706810135e-05, "loss": 5.2345, "step": 83600 }, { "epoch": 3.849546118710954, "grad_norm": 2.3374664783477783, "learning_rate": 1.275484689390024e-05, "loss": 5.2227, "step": 83650 }, { "epoch": 3.8518471645362817, "grad_norm": 2.4744300842285156, "learning_rate": 1.2706208389852087e-05, "loss": 5.2145, "step": 83700 }, { "epoch": 3.8541482103616094, "grad_norm": 2.243946075439453, "learning_rate": 1.2657649298235213e-05, "loss": 5.1909, "step": 83750 }, { "epoch": 3.856449256186937, "grad_norm": 2.732558012008667, "learning_rate": 1.2609169722450093e-05, "loss": 5.1886, "step": 83800 }, { "epoch": 3.8587503020122647, "grad_norm": 2.3516011238098145, "learning_rate": 1.2560769765727881e-05, "loss": 5.146, "step": 83850 }, { "epoch": 3.8610513478375923, "grad_norm": 2.6143431663513184, "learning_rate": 1.2512449531130183e-05, "loss": 5.1711, "step": 83900 }, { "epoch": 3.86335239366292, "grad_norm": 2.5239765644073486, "learning_rate": 1.2464209121548848e-05, "loss": 5.1856, "step": 83950 }, { "epoch": 3.8656534394882476, "grad_norm": 2.2582340240478516, "learning_rate": 1.2416048639705757e-05, "loss": 5.1832, "step": 84000 }, { "epoch": 3.8656534394882476, "eval_loss": 5.601240158081055, "eval_runtime": 34.4022, "eval_samples_per_second": 11.162, "eval_steps_per_second": 5.581, "eval_tts_loss": 9.00224707553787, "step": 84000 }, { "epoch": 3.867954485313575, "grad_norm": 2.666868209838867, "learning_rate": 1.236796818815259e-05, "loss": 5.1795, "step": 84050 }, { "epoch": 3.870255531138903, "grad_norm": 1.95122492313385, "learning_rate": 1.231996786927061e-05, "loss": 5.1342, "step": 84100 }, { "epoch": 3.8725565769642305, "grad_norm": 2.664674758911133, "learning_rate": 1.2272047785270446e-05, "loss": 5.2027, "step": 84150 }, { "epoch": 3.8748576227895577, "grad_norm": 2.152007818222046, "learning_rate": 1.2224208038191886e-05, "loss": 5.1717, "step": 84200 }, { "epoch": 3.8771586686148853, "grad_norm": 2.2395548820495605, "learning_rate": 1.217644872990364e-05, "loss": 5.1786, "step": 84250 }, { "epoch": 3.879459714440213, "grad_norm": 2.2887299060821533, "learning_rate": 1.2128769962103142e-05, "loss": 5.1917, "step": 84300 }, { "epoch": 3.8817607602655406, "grad_norm": 2.3832855224609375, "learning_rate": 1.2081171836316325e-05, "loss": 5.2202, "step": 84350 }, { "epoch": 3.8840618060908683, "grad_norm": 2.2544591426849365, "learning_rate": 1.2033654453897391e-05, "loss": 5.1446, "step": 84400 }, { "epoch": 3.886362851916196, "grad_norm": 2.403583526611328, "learning_rate": 1.198621791602863e-05, "loss": 5.2115, "step": 84450 }, { "epoch": 3.8886638977415235, "grad_norm": 2.506659507751465, "learning_rate": 1.1939808641647338e-05, "loss": 5.1118, "step": 84500 }, { "epoch": 3.890964943566851, "grad_norm": 2.4026944637298584, "learning_rate": 1.1892532473821888e-05, "loss": 5.1479, "step": 84550 }, { "epoch": 3.893265989392179, "grad_norm": 2.1828866004943848, "learning_rate": 1.1845337451048095e-05, "loss": 5.1982, "step": 84600 }, { "epoch": 3.8955670352175065, "grad_norm": 2.180025339126587, "learning_rate": 1.17982236738218e-05, "loss": 5.1594, "step": 84650 }, { "epoch": 3.897868081042834, "grad_norm": 2.3280653953552246, "learning_rate": 1.175119124246587e-05, "loss": 5.2365, "step": 84700 }, { "epoch": 3.9001691268681613, "grad_norm": 2.364051580429077, "learning_rate": 1.1704240257129934e-05, "loss": 5.1986, "step": 84750 }, { "epoch": 3.902470172693489, "grad_norm": 2.603224754333496, "learning_rate": 1.1657370817790186e-05, "loss": 5.1579, "step": 84800 }, { "epoch": 3.9047712185188166, "grad_norm": 2.6435306072235107, "learning_rate": 1.1610583024249227e-05, "loss": 5.1749, "step": 84850 }, { "epoch": 3.9070722643441442, "grad_norm": 2.8594369888305664, "learning_rate": 1.1563876976135751e-05, "loss": 5.1947, "step": 84900 }, { "epoch": 3.909373310169472, "grad_norm": 2.1866133213043213, "learning_rate": 1.1517252772904408e-05, "loss": 5.186, "step": 84950 }, { "epoch": 3.9116743559947995, "grad_norm": 2.3272008895874023, "learning_rate": 1.1470710513835559e-05, "loss": 5.1384, "step": 85000 }, { "epoch": 3.9116743559947995, "eval_loss": 5.6002326011657715, "eval_runtime": 33.5323, "eval_samples_per_second": 11.452, "eval_steps_per_second": 5.726, "eval_tts_loss": 9.014526556404022, "step": 85000 }, { "epoch": 3.913975401820127, "grad_norm": 2.3301563262939453, "learning_rate": 1.1424250298035083e-05, "loss": 5.2276, "step": 85050 }, { "epoch": 3.916276447645455, "grad_norm": 2.2211480140686035, "learning_rate": 1.1377872224434178e-05, "loss": 5.2068, "step": 85100 }, { "epoch": 3.9185774934707824, "grad_norm": 2.2685728073120117, "learning_rate": 1.1331576391789062e-05, "loss": 5.1322, "step": 85150 }, { "epoch": 3.92087853929611, "grad_norm": 2.3483150005340576, "learning_rate": 1.1285362898680923e-05, "loss": 5.1722, "step": 85200 }, { "epoch": 3.9231795851214377, "grad_norm": 2.068774938583374, "learning_rate": 1.1239231843515552e-05, "loss": 5.1993, "step": 85250 }, { "epoch": 3.9254806309467654, "grad_norm": 2.5643553733825684, "learning_rate": 1.1193183324523232e-05, "loss": 5.169, "step": 85300 }, { "epoch": 3.927781676772093, "grad_norm": 2.1620678901672363, "learning_rate": 1.1147217439758473e-05, "loss": 5.1113, "step": 85350 }, { "epoch": 3.9300827225974206, "grad_norm": 2.5162718296051025, "learning_rate": 1.1101334287099835e-05, "loss": 5.1935, "step": 85400 }, { "epoch": 3.9323837684227483, "grad_norm": 2.4817066192626953, "learning_rate": 1.105553396424972e-05, "loss": 5.1891, "step": 85450 }, { "epoch": 3.934684814248076, "grad_norm": 2.493964433670044, "learning_rate": 1.1009816568734143e-05, "loss": 5.1429, "step": 85500 }, { "epoch": 3.9369858600734036, "grad_norm": 2.5530591011047363, "learning_rate": 1.096418219790254e-05, "loss": 5.1996, "step": 85550 }, { "epoch": 3.939286905898731, "grad_norm": 2.2496776580810547, "learning_rate": 1.0918630948927554e-05, "loss": 5.1454, "step": 85600 }, { "epoch": 3.941587951724059, "grad_norm": 2.241255521774292, "learning_rate": 1.0873162918804814e-05, "loss": 5.1635, "step": 85650 }, { "epoch": 3.943888997549386, "grad_norm": 2.414360284805298, "learning_rate": 1.082777820435279e-05, "loss": 5.132, "step": 85700 }, { "epoch": 3.9461900433747137, "grad_norm": 2.702965259552002, "learning_rate": 1.0782476902212507e-05, "loss": 5.1775, "step": 85750 }, { "epoch": 3.9484910892000413, "grad_norm": 2.7579190731048584, "learning_rate": 1.073725910884738e-05, "loss": 5.1708, "step": 85800 }, { "epoch": 3.950792135025369, "grad_norm": 2.170175313949585, "learning_rate": 1.0692124920542996e-05, "loss": 5.1411, "step": 85850 }, { "epoch": 3.9530931808506966, "grad_norm": 2.2946743965148926, "learning_rate": 1.0647074433406934e-05, "loss": 5.1557, "step": 85900 }, { "epoch": 3.9553942266760242, "grad_norm": 2.442544937133789, "learning_rate": 1.0602107743368522e-05, "loss": 5.1201, "step": 85950 }, { "epoch": 3.957695272501352, "grad_norm": 2.3080036640167236, "learning_rate": 1.0557224946178667e-05, "loss": 5.1919, "step": 86000 }, { "epoch": 3.957695272501352, "eval_loss": 5.599061489105225, "eval_runtime": 34.3103, "eval_samples_per_second": 11.192, "eval_steps_per_second": 5.596, "eval_tts_loss": 9.014872747805995, "step": 86000 }, { "epoch": 3.9599963183266795, "grad_norm": 2.5096185207366943, "learning_rate": 1.051242613740963e-05, "loss": 5.1564, "step": 86050 }, { "epoch": 3.962297364152007, "grad_norm": 2.575779914855957, "learning_rate": 1.0467711412454823e-05, "loss": 5.1363, "step": 86100 }, { "epoch": 3.964598409977335, "grad_norm": 2.8018746376037598, "learning_rate": 1.042308086652864e-05, "loss": 5.1536, "step": 86150 }, { "epoch": 3.9668994558026625, "grad_norm": 2.1475179195404053, "learning_rate": 1.0378534594666206e-05, "loss": 5.2278, "step": 86200 }, { "epoch": 3.9692005016279897, "grad_norm": 2.4993419647216797, "learning_rate": 1.0334072691723168e-05, "loss": 5.1654, "step": 86250 }, { "epoch": 3.9715015474533173, "grad_norm": 2.0179409980773926, "learning_rate": 1.028969525237557e-05, "loss": 5.2237, "step": 86300 }, { "epoch": 3.973802593278645, "grad_norm": 2.1702675819396973, "learning_rate": 1.0245402371119567e-05, "loss": 5.109, "step": 86350 }, { "epoch": 3.9761036391039726, "grad_norm": 2.7319464683532715, "learning_rate": 1.0201194142271281e-05, "loss": 5.1879, "step": 86400 }, { "epoch": 3.9784046849293, "grad_norm": 2.3209543228149414, "learning_rate": 1.015707065996656e-05, "loss": 5.0868, "step": 86450 }, { "epoch": 3.980705730754628, "grad_norm": 2.6626105308532715, "learning_rate": 1.0113032018160783e-05, "loss": 5.1558, "step": 86500 }, { "epoch": 3.9830067765799555, "grad_norm": 2.4389021396636963, "learning_rate": 1.0069078310628732e-05, "loss": 5.2351, "step": 86550 }, { "epoch": 3.985307822405283, "grad_norm": 2.3185486793518066, "learning_rate": 1.002608617067996e-05, "loss": 5.144, "step": 86600 }, { "epoch": 3.987608868230611, "grad_norm": 2.1815450191497803, "learning_rate": 9.982300908956022e-06, "loss": 5.1203, "step": 86650 }, { "epoch": 3.9899099140559384, "grad_norm": 2.7808938026428223, "learning_rate": 9.938600859881219e-06, "loss": 5.1624, "step": 86700 }, { "epoch": 3.992210959881266, "grad_norm": 2.6030192375183105, "learning_rate": 9.894986116509297e-06, "loss": 5.2076, "step": 86750 }, { "epoch": 3.9945120057065937, "grad_norm": 2.2136263847351074, "learning_rate": 9.851456771712342e-06, "loss": 5.2135, "step": 86800 }, { "epoch": 3.9968130515319213, "grad_norm": 2.76670241355896, "learning_rate": 9.808012918180643e-06, "loss": 5.2533, "step": 86850 }, { "epoch": 3.999114097357249, "grad_norm": 2.3571465015411377, "learning_rate": 9.764654648422394e-06, "loss": 5.2091, "step": 86900 }, { "epoch": 4.001380627495196, "grad_norm": 2.9005565643310547, "learning_rate": 9.721382054763572e-06, "loss": 5.0138, "step": 86950 }, { "epoch": 4.003681673320524, "grad_norm": 2.7534894943237305, "learning_rate": 9.678195229347714e-06, "loss": 4.9149, "step": 87000 }, { "epoch": 4.003681673320524, "eval_loss": 5.650022029876709, "eval_runtime": 34.0508, "eval_samples_per_second": 11.277, "eval_steps_per_second": 5.639, "eval_tts_loss": 9.280170685445977, "step": 87000 }, { "epoch": 4.005982719145852, "grad_norm": 2.4095113277435303, "learning_rate": 9.635094264135724e-06, "loss": 4.9212, "step": 87050 }, { "epoch": 4.008283764971179, "grad_norm": 2.556870222091675, "learning_rate": 9.592079250905684e-06, "loss": 4.9187, "step": 87100 }, { "epoch": 4.010584810796507, "grad_norm": 2.400278091430664, "learning_rate": 9.549150281252633e-06, "loss": 4.9294, "step": 87150 }, { "epoch": 4.012885856621835, "grad_norm": 2.6151013374328613, "learning_rate": 9.506307446588414e-06, "loss": 4.933, "step": 87200 }, { "epoch": 4.015186902447162, "grad_norm": 3.015284299850464, "learning_rate": 9.463550838141427e-06, "loss": 4.9152, "step": 87250 }, { "epoch": 4.01748794827249, "grad_norm": 2.4685239791870117, "learning_rate": 9.420880546956528e-06, "loss": 4.9318, "step": 87300 }, { "epoch": 4.0197889940978175, "grad_norm": 2.854645252227783, "learning_rate": 9.378296663894692e-06, "loss": 4.9648, "step": 87350 }, { "epoch": 4.022090039923145, "grad_norm": 2.287388563156128, "learning_rate": 9.335799279632934e-06, "loss": 4.954, "step": 87400 }, { "epoch": 4.024391085748473, "grad_norm": 2.6801013946533203, "learning_rate": 9.293388484664089e-06, "loss": 4.9327, "step": 87450 }, { "epoch": 4.0266921315738005, "grad_norm": 2.134859561920166, "learning_rate": 9.251064369296597e-06, "loss": 4.9183, "step": 87500 }, { "epoch": 4.028993177399128, "grad_norm": 2.5280423164367676, "learning_rate": 9.208827023654336e-06, "loss": 4.9208, "step": 87550 }, { "epoch": 4.031294223224456, "grad_norm": 2.4778597354888916, "learning_rate": 9.166676537676416e-06, "loss": 4.9699, "step": 87600 }, { "epoch": 4.033595269049783, "grad_norm": 2.7387208938598633, "learning_rate": 9.12461300111696e-06, "loss": 4.9429, "step": 87650 }, { "epoch": 4.035896314875111, "grad_norm": 2.26796555519104, "learning_rate": 9.08263650354501e-06, "loss": 4.932, "step": 87700 }, { "epoch": 4.038197360700439, "grad_norm": 2.425043821334839, "learning_rate": 9.040747134344208e-06, "loss": 4.8945, "step": 87750 }, { "epoch": 4.040498406525766, "grad_norm": 2.616593599319458, "learning_rate": 8.998944982712692e-06, "loss": 4.9216, "step": 87800 }, { "epoch": 4.042799452351094, "grad_norm": 2.8648416996002197, "learning_rate": 8.957230137662881e-06, "loss": 4.9731, "step": 87850 }, { "epoch": 4.045100498176422, "grad_norm": 2.239473581314087, "learning_rate": 8.91560268802128e-06, "loss": 4.9001, "step": 87900 }, { "epoch": 4.047401544001749, "grad_norm": 2.3506898880004883, "learning_rate": 8.874892663824048e-06, "loss": 4.9684, "step": 87950 }, { "epoch": 4.049702589827076, "grad_norm": 2.2261645793914795, "learning_rate": 8.833438518418107e-06, "loss": 5.0119, "step": 88000 }, { "epoch": 4.049702589827076, "eval_loss": 5.670416355133057, "eval_runtime": 33.0376, "eval_samples_per_second": 11.623, "eval_steps_per_second": 5.812, "eval_tts_loss": 9.288442969832703, "step": 88000 }, { "epoch": 4.052003635652404, "grad_norm": 2.7639026641845703, "learning_rate": 8.792072032019038e-06, "loss": 4.9038, "step": 88050 }, { "epoch": 4.054304681477731, "grad_norm": 2.569985866546631, "learning_rate": 8.750793292711563e-06, "loss": 4.977, "step": 88100 }, { "epoch": 4.056605727303059, "grad_norm": 3.0355279445648193, "learning_rate": 8.70960238839355e-06, "loss": 4.9272, "step": 88150 }, { "epoch": 4.0589067731283865, "grad_norm": 2.587646245956421, "learning_rate": 8.668499406775849e-06, "loss": 4.9789, "step": 88200 }, { "epoch": 4.061207818953714, "grad_norm": 2.128152847290039, "learning_rate": 8.627484435382072e-06, "loss": 4.9165, "step": 88250 }, { "epoch": 4.063508864779042, "grad_norm": 2.4285528659820557, "learning_rate": 8.586557561548442e-06, "loss": 4.9074, "step": 88300 }, { "epoch": 4.0658099106043695, "grad_norm": 2.435657501220703, "learning_rate": 8.545718872423581e-06, "loss": 4.9301, "step": 88350 }, { "epoch": 4.068110956429697, "grad_norm": 2.365025043487549, "learning_rate": 8.504968454968326e-06, "loss": 4.94, "step": 88400 }, { "epoch": 4.070412002255025, "grad_norm": 2.667067766189575, "learning_rate": 8.464306395955568e-06, "loss": 4.9485, "step": 88450 }, { "epoch": 4.072713048080352, "grad_norm": 2.2761590480804443, "learning_rate": 8.423732781970029e-06, "loss": 4.9872, "step": 88500 }, { "epoch": 4.07501409390568, "grad_norm": 2.8008153438568115, "learning_rate": 8.383247699408115e-06, "loss": 4.9176, "step": 88550 }, { "epoch": 4.077315139731008, "grad_norm": 2.3409249782562256, "learning_rate": 8.34285123447771e-06, "loss": 4.9863, "step": 88600 }, { "epoch": 4.079616185556335, "grad_norm": 2.465102195739746, "learning_rate": 8.302543473197993e-06, "loss": 4.9414, "step": 88650 }, { "epoch": 4.081917231381663, "grad_norm": 2.3558099269866943, "learning_rate": 8.262324501399265e-06, "loss": 4.8924, "step": 88700 }, { "epoch": 4.084218277206991, "grad_norm": 2.590440034866333, "learning_rate": 8.222194404722767e-06, "loss": 4.9291, "step": 88750 }, { "epoch": 4.086519323032318, "grad_norm": 2.5797536373138428, "learning_rate": 8.182153268620462e-06, "loss": 4.9662, "step": 88800 }, { "epoch": 4.088820368857646, "grad_norm": 2.68279767036438, "learning_rate": 8.142201178354936e-06, "loss": 4.9116, "step": 88850 }, { "epoch": 4.0911214146829735, "grad_norm": 2.7302539348602295, "learning_rate": 8.102338218999118e-06, "loss": 5.0191, "step": 88900 }, { "epoch": 4.093422460508301, "grad_norm": 2.9300999641418457, "learning_rate": 8.062564475436157e-06, "loss": 4.9311, "step": 88950 }, { "epoch": 4.095723506333629, "grad_norm": 2.688512086868286, "learning_rate": 8.022880032359231e-06, "loss": 4.863, "step": 89000 }, { "epoch": 4.095723506333629, "eval_loss": 5.680548191070557, "eval_runtime": 34.1834, "eval_samples_per_second": 11.234, "eval_steps_per_second": 5.617, "eval_tts_loss": 9.326126809397413, "step": 89000 }, { "epoch": 4.0980245521589564, "grad_norm": 2.644500732421875, "learning_rate": 7.983284974271365e-06, "loss": 4.943, "step": 89050 }, { "epoch": 4.100325597984284, "grad_norm": 2.59487247467041, "learning_rate": 7.943779385485244e-06, "loss": 4.9148, "step": 89100 }, { "epoch": 4.102626643809612, "grad_norm": 2.785487174987793, "learning_rate": 7.904363350123045e-06, "loss": 4.8976, "step": 89150 }, { "epoch": 4.104927689634939, "grad_norm": 2.6682450771331787, "learning_rate": 7.86503695211625e-06, "loss": 4.9197, "step": 89200 }, { "epoch": 4.107228735460267, "grad_norm": 3.0894181728363037, "learning_rate": 7.825800275205459e-06, "loss": 4.9131, "step": 89250 }, { "epoch": 4.109529781285595, "grad_norm": 2.421407461166382, "learning_rate": 7.786653402940247e-06, "loss": 4.9777, "step": 89300 }, { "epoch": 4.111830827110922, "grad_norm": 2.4230687618255615, "learning_rate": 7.747596418678949e-06, "loss": 4.9121, "step": 89350 }, { "epoch": 4.11413187293625, "grad_norm": 2.8178746700286865, "learning_rate": 7.708629405588486e-06, "loss": 4.9246, "step": 89400 }, { "epoch": 4.116432918761577, "grad_norm": 2.6915271282196045, "learning_rate": 7.669752446644196e-06, "loss": 4.9551, "step": 89450 }, { "epoch": 4.118733964586904, "grad_norm": 2.7895405292510986, "learning_rate": 7.630965624629693e-06, "loss": 4.9911, "step": 89500 }, { "epoch": 4.121035010412232, "grad_norm": 2.5442700386047363, "learning_rate": 7.592269022136594e-06, "loss": 4.9471, "step": 89550 }, { "epoch": 4.12333605623756, "grad_norm": 2.405792713165283, "learning_rate": 7.5536627215644585e-06, "loss": 4.9543, "step": 89600 }, { "epoch": 4.125637102062887, "grad_norm": 2.431641101837158, "learning_rate": 7.515146805120521e-06, "loss": 4.9375, "step": 89650 }, { "epoch": 4.127938147888215, "grad_norm": 2.513671398162842, "learning_rate": 7.476721354819604e-06, "loss": 4.9233, "step": 89700 }, { "epoch": 4.1302391937135425, "grad_norm": 2.221925735473633, "learning_rate": 7.4383864524838474e-06, "loss": 4.9451, "step": 89750 }, { "epoch": 4.13254023953887, "grad_norm": 2.2546591758728027, "learning_rate": 7.4001421797425975e-06, "loss": 4.9119, "step": 89800 }, { "epoch": 4.134841285364198, "grad_norm": 2.906501054763794, "learning_rate": 7.361988618032212e-06, "loss": 4.8577, "step": 89850 }, { "epoch": 4.1371423311895255, "grad_norm": 2.6346490383148193, "learning_rate": 7.323925848595903e-06, "loss": 4.9964, "step": 89900 }, { "epoch": 4.139443377014853, "grad_norm": 2.4832003116607666, "learning_rate": 7.285953952483537e-06, "loss": 4.9329, "step": 89950 }, { "epoch": 4.141744422840181, "grad_norm": 2.5018067359924316, "learning_rate": 7.248073010551481e-06, "loss": 4.9203, "step": 90000 }, { "epoch": 4.141744422840181, "eval_loss": 5.682537078857422, "eval_runtime": 34.224, "eval_samples_per_second": 11.22, "eval_steps_per_second": 5.61, "eval_tts_loss": 9.324252976809897, "step": 90000 }, { "epoch": 4.144045468665508, "grad_norm": 2.496283531188965, "learning_rate": 7.210283103462434e-06, "loss": 4.9322, "step": 90050 }, { "epoch": 4.146346514490836, "grad_norm": 2.9973666667938232, "learning_rate": 7.172584311685232e-06, "loss": 4.9629, "step": 90100 }, { "epoch": 4.148647560316164, "grad_norm": 2.648648500442505, "learning_rate": 7.134976715494723e-06, "loss": 4.9944, "step": 90150 }, { "epoch": 4.150948606141491, "grad_norm": 3.1687586307525635, "learning_rate": 7.097460394971534e-06, "loss": 4.8967, "step": 90200 }, { "epoch": 4.153249651966819, "grad_norm": 3.1536190509796143, "learning_rate": 7.060035430001943e-06, "loss": 4.9063, "step": 90250 }, { "epoch": 4.155550697792147, "grad_norm": 2.530733823776245, "learning_rate": 7.022701900277701e-06, "loss": 4.8927, "step": 90300 }, { "epoch": 4.157851743617474, "grad_norm": 2.789060592651367, "learning_rate": 6.985459885295853e-06, "loss": 4.9174, "step": 90350 }, { "epoch": 4.160152789442802, "grad_norm": 2.395866632461548, "learning_rate": 6.9483094643585826e-06, "loss": 4.9536, "step": 90400 }, { "epoch": 4.1624538352681295, "grad_norm": 2.376652717590332, "learning_rate": 6.911250716573031e-06, "loss": 4.9544, "step": 90450 }, { "epoch": 4.164754881093457, "grad_norm": 2.4753246307373047, "learning_rate": 6.874283720851132e-06, "loss": 4.9249, "step": 90500 }, { "epoch": 4.167055926918785, "grad_norm": 2.8535377979278564, "learning_rate": 6.837408555909447e-06, "loss": 5.0122, "step": 90550 }, { "epoch": 4.169356972744112, "grad_norm": 2.5676331520080566, "learning_rate": 6.800625300268987e-06, "loss": 4.913, "step": 90600 }, { "epoch": 4.17165801856944, "grad_norm": 2.30092716217041, "learning_rate": 6.763934032255076e-06, "loss": 4.942, "step": 90650 }, { "epoch": 4.173959064394768, "grad_norm": 3.224811553955078, "learning_rate": 6.727334829997134e-06, "loss": 4.9063, "step": 90700 }, { "epoch": 4.176260110220095, "grad_norm": 2.8014469146728516, "learning_rate": 6.690827771428559e-06, "loss": 4.9177, "step": 90750 }, { "epoch": 4.178561156045423, "grad_norm": 2.4171247482299805, "learning_rate": 6.65441293428653e-06, "loss": 4.9258, "step": 90800 }, { "epoch": 4.180862201870751, "grad_norm": 2.6210904121398926, "learning_rate": 6.618090396111859e-06, "loss": 4.885, "step": 90850 }, { "epoch": 4.183163247696078, "grad_norm": 2.62601637840271, "learning_rate": 6.58186023424881e-06, "loss": 4.9679, "step": 90900 }, { "epoch": 4.185464293521406, "grad_norm": 2.9907238483428955, "learning_rate": 6.545722525844944e-06, "loss": 4.9243, "step": 90950 }, { "epoch": 4.187765339346733, "grad_norm": 2.3237407207489014, "learning_rate": 6.509677347850973e-06, "loss": 4.96, "step": 91000 }, { "epoch": 4.187765339346733, "eval_loss": 5.679426670074463, "eval_runtime": 35.225, "eval_samples_per_second": 10.901, "eval_steps_per_second": 5.451, "eval_tts_loss": 9.328892827813839, "step": 91000 }, { "epoch": 4.19006638517206, "grad_norm": 2.776470422744751, "learning_rate": 6.473724777020557e-06, "loss": 4.9766, "step": 91050 }, { "epoch": 4.192367430997388, "grad_norm": 2.488623857498169, "learning_rate": 6.437864889910172e-06, "loss": 4.9468, "step": 91100 }, { "epoch": 4.194668476822716, "grad_norm": 2.666916608810425, "learning_rate": 6.402097762878917e-06, "loss": 4.896, "step": 91150 }, { "epoch": 4.196969522648043, "grad_norm": 2.505510091781616, "learning_rate": 6.366423472088396e-06, "loss": 4.9794, "step": 91200 }, { "epoch": 4.199270568473371, "grad_norm": 2.4801712036132812, "learning_rate": 6.33084209350251e-06, "loss": 4.9606, "step": 91250 }, { "epoch": 4.2015716142986985, "grad_norm": 2.5550425052642822, "learning_rate": 6.2953537028873245e-06, "loss": 4.8826, "step": 91300 }, { "epoch": 4.203872660124026, "grad_norm": 3.3445262908935547, "learning_rate": 6.259958375810898e-06, "loss": 4.9318, "step": 91350 }, { "epoch": 4.206173705949354, "grad_norm": 2.3678958415985107, "learning_rate": 6.2246561876431185e-06, "loss": 4.9158, "step": 91400 }, { "epoch": 4.208474751774681, "grad_norm": 2.5698840618133545, "learning_rate": 6.189447213555538e-06, "loss": 4.9356, "step": 91450 }, { "epoch": 4.210775797600009, "grad_norm": 2.515235185623169, "learning_rate": 6.154331528521251e-06, "loss": 4.9391, "step": 91500 }, { "epoch": 4.213076843425337, "grad_norm": 3.0568439960479736, "learning_rate": 6.119309207314683e-06, "loss": 4.8527, "step": 91550 }, { "epoch": 4.215377889250664, "grad_norm": 2.7625505924224854, "learning_rate": 6.084380324511446e-06, "loss": 4.8952, "step": 91600 }, { "epoch": 4.217678935075992, "grad_norm": 2.4264955520629883, "learning_rate": 6.049544954488201e-06, "loss": 4.9947, "step": 91650 }, { "epoch": 4.21997998090132, "grad_norm": 2.910615921020508, "learning_rate": 6.014803171422489e-06, "loss": 4.9183, "step": 91700 }, { "epoch": 4.222281026726647, "grad_norm": 2.3442256450653076, "learning_rate": 5.9801550492925404e-06, "loss": 4.8731, "step": 91750 }, { "epoch": 4.224582072551975, "grad_norm": 2.891833782196045, "learning_rate": 5.945600661877165e-06, "loss": 4.9631, "step": 91800 }, { "epoch": 4.226883118377303, "grad_norm": 2.956780433654785, "learning_rate": 5.911140082755595e-06, "loss": 4.9261, "step": 91850 }, { "epoch": 4.22918416420263, "grad_norm": 2.5378317832946777, "learning_rate": 5.876773385307294e-06, "loss": 5.0156, "step": 91900 }, { "epoch": 4.231485210027958, "grad_norm": 2.6025545597076416, "learning_rate": 5.842500642711807e-06, "loss": 4.9972, "step": 91950 }, { "epoch": 4.2337862558532855, "grad_norm": 3.2855794429779053, "learning_rate": 5.808321927948618e-06, "loss": 4.9, "step": 92000 }, { "epoch": 4.2337862558532855, "eval_loss": 5.68091344833374, "eval_runtime": 35.0218, "eval_samples_per_second": 10.965, "eval_steps_per_second": 5.482, "eval_tts_loss": 9.340122075743121, "step": 92000 }, { "epoch": 4.236087301678613, "grad_norm": 2.4028210639953613, "learning_rate": 5.774237313797004e-06, "loss": 4.9617, "step": 92050 }, { "epoch": 4.238388347503941, "grad_norm": 2.223472833633423, "learning_rate": 5.7402468728358524e-06, "loss": 4.9744, "step": 92100 }, { "epoch": 4.240689393329268, "grad_norm": 2.3175554275512695, "learning_rate": 5.706350677443522e-06, "loss": 4.9287, "step": 92150 }, { "epoch": 4.242990439154596, "grad_norm": 2.640993356704712, "learning_rate": 5.6725487997977e-06, "loss": 4.8685, "step": 92200 }, { "epoch": 4.245291484979924, "grad_norm": 2.1567435264587402, "learning_rate": 5.638841311875215e-06, "loss": 4.9202, "step": 92250 }, { "epoch": 4.247592530805251, "grad_norm": 2.5117077827453613, "learning_rate": 5.605899619794419e-06, "loss": 4.9533, "step": 92300 }, { "epoch": 4.249893576630579, "grad_norm": 2.356736421585083, "learning_rate": 5.5723792350834196e-06, "loss": 4.9547, "step": 92350 }, { "epoch": 4.252194622455907, "grad_norm": 2.7864232063293457, "learning_rate": 5.538953453394236e-06, "loss": 4.9201, "step": 92400 }, { "epoch": 4.254495668281233, "grad_norm": 2.60893177986145, "learning_rate": 5.50562234590285e-06, "loss": 4.9713, "step": 92450 }, { "epoch": 4.256796714106561, "grad_norm": 3.0111494064331055, "learning_rate": 5.472385983583667e-06, "loss": 4.8794, "step": 92500 }, { "epoch": 4.259097759931889, "grad_norm": 2.4298832416534424, "learning_rate": 5.439244437209329e-06, "loss": 4.9719, "step": 92550 }, { "epoch": 4.261398805757216, "grad_norm": 2.4979469776153564, "learning_rate": 5.406197777350574e-06, "loss": 4.9165, "step": 92600 }, { "epoch": 4.263699851582544, "grad_norm": 2.7282731533050537, "learning_rate": 5.3732460743761255e-06, "loss": 4.959, "step": 92650 }, { "epoch": 4.266000897407872, "grad_norm": 2.674407720565796, "learning_rate": 5.340389398452472e-06, "loss": 4.9879, "step": 92700 }, { "epoch": 4.268301943233199, "grad_norm": 2.50618577003479, "learning_rate": 5.307627819543759e-06, "loss": 4.9555, "step": 92750 }, { "epoch": 4.270602989058527, "grad_norm": 2.716242790222168, "learning_rate": 5.27496140741166e-06, "loss": 4.9114, "step": 92800 }, { "epoch": 4.2729040348838545, "grad_norm": 2.615030527114868, "learning_rate": 5.2423902316151575e-06, "loss": 4.9386, "step": 92850 }, { "epoch": 4.275205080709182, "grad_norm": 2.4920854568481445, "learning_rate": 5.209914361510487e-06, "loss": 4.9112, "step": 92900 }, { "epoch": 4.27750612653451, "grad_norm": 2.827641725540161, "learning_rate": 5.177533866250905e-06, "loss": 4.9114, "step": 92950 }, { "epoch": 4.279807172359837, "grad_norm": 2.755664587020874, "learning_rate": 5.145248814786619e-06, "loss": 4.9433, "step": 93000 }, { "epoch": 4.279807172359837, "eval_loss": 5.679144382476807, "eval_runtime": 34.4128, "eval_samples_per_second": 11.159, "eval_steps_per_second": 5.579, "eval_tts_loss": 9.318372178651252, "step": 93000 }, { "epoch": 4.282108218185165, "grad_norm": 2.069977045059204, "learning_rate": 5.113059275864573e-06, "loss": 4.8856, "step": 93050 }, { "epoch": 4.284409264010493, "grad_norm": 2.4945602416992188, "learning_rate": 5.08096531802833e-06, "loss": 4.937, "step": 93100 }, { "epoch": 4.28671030983582, "grad_norm": 2.9363367557525635, "learning_rate": 5.048967009617933e-06, "loss": 4.952, "step": 93150 }, { "epoch": 4.289011355661148, "grad_norm": 2.544668436050415, "learning_rate": 5.0170644187697455e-06, "loss": 4.8924, "step": 93200 }, { "epoch": 4.291312401486476, "grad_norm": 2.6560354232788086, "learning_rate": 4.9852576134163224e-06, "loss": 4.8599, "step": 93250 }, { "epoch": 4.293613447311803, "grad_norm": 2.506303548812866, "learning_rate": 4.95354666128624e-06, "loss": 4.9508, "step": 93300 }, { "epoch": 4.295914493137131, "grad_norm": 2.6585988998413086, "learning_rate": 4.921931629903986e-06, "loss": 4.8903, "step": 93350 }, { "epoch": 4.298215538962459, "grad_norm": 2.9596993923187256, "learning_rate": 4.890412586589777e-06, "loss": 4.9403, "step": 93400 }, { "epoch": 4.300516584787786, "grad_norm": 2.412736177444458, "learning_rate": 4.858989598459446e-06, "loss": 4.9516, "step": 93450 }, { "epoch": 4.302817630613114, "grad_norm": 2.6511383056640625, "learning_rate": 4.827662732424299e-06, "loss": 4.9228, "step": 93500 }, { "epoch": 4.3051186764384415, "grad_norm": 3.0368809700012207, "learning_rate": 4.796432055190947e-06, "loss": 4.8955, "step": 93550 }, { "epoch": 4.307419722263769, "grad_norm": 2.5606002807617188, "learning_rate": 4.7652976332611785e-06, "loss": 4.9485, "step": 93600 }, { "epoch": 4.309720768089097, "grad_norm": 2.8457987308502197, "learning_rate": 4.7342595329318225e-06, "loss": 4.9452, "step": 93650 }, { "epoch": 4.312021813914424, "grad_norm": 2.4588279724121094, "learning_rate": 4.703317820294612e-06, "loss": 4.9085, "step": 93700 }, { "epoch": 4.314322859739752, "grad_norm": 2.9369304180145264, "learning_rate": 4.6724725612360186e-06, "loss": 4.936, "step": 93750 }, { "epoch": 4.31662390556508, "grad_norm": 2.102524995803833, "learning_rate": 4.641723821437132e-06, "loss": 4.9798, "step": 93800 }, { "epoch": 4.318924951390407, "grad_norm": 2.3589839935302734, "learning_rate": 4.611071666373534e-06, "loss": 4.934, "step": 93850 }, { "epoch": 4.321225997215734, "grad_norm": 2.485720634460449, "learning_rate": 4.5805161613151124e-06, "loss": 4.8953, "step": 93900 }, { "epoch": 4.323527043041063, "grad_norm": 2.6778573989868164, "learning_rate": 4.5500573713259776e-06, "loss": 4.866, "step": 93950 }, { "epoch": 4.325828088866389, "grad_norm": 2.9395782947540283, "learning_rate": 4.519695361264281e-06, "loss": 4.9388, "step": 94000 }, { "epoch": 4.325828088866389, "eval_loss": 5.68023157119751, "eval_runtime": 34.8801, "eval_samples_per_second": 11.009, "eval_steps_per_second": 5.505, "eval_tts_loss": 9.319495803010797, "step": 94000 }, { "epoch": 4.328129134691717, "grad_norm": 2.3470025062561035, "learning_rate": 4.489430195782096e-06, "loss": 4.9749, "step": 94050 }, { "epoch": 4.330430180517045, "grad_norm": 2.6966683864593506, "learning_rate": 4.459261939325282e-06, "loss": 4.9726, "step": 94100 }, { "epoch": 4.332731226342372, "grad_norm": 2.843234062194824, "learning_rate": 4.4291906561333415e-06, "loss": 4.9067, "step": 94150 }, { "epoch": 4.3350322721677, "grad_norm": 2.3041133880615234, "learning_rate": 4.399216410239288e-06, "loss": 4.869, "step": 94200 }, { "epoch": 4.337333317993028, "grad_norm": 2.9902963638305664, "learning_rate": 4.369339265469496e-06, "loss": 4.8756, "step": 94250 }, { "epoch": 4.339634363818355, "grad_norm": 2.85142183303833, "learning_rate": 4.339559285443573e-06, "loss": 4.9174, "step": 94300 }, { "epoch": 4.341935409643683, "grad_norm": 3.339543581008911, "learning_rate": 4.309876533574258e-06, "loss": 4.9478, "step": 94350 }, { "epoch": 4.3442364554690105, "grad_norm": 2.467912435531616, "learning_rate": 4.280291073067216e-06, "loss": 4.8815, "step": 94400 }, { "epoch": 4.346537501294338, "grad_norm": 3.1716277599334717, "learning_rate": 4.250802966920964e-06, "loss": 4.893, "step": 94450 }, { "epoch": 4.348838547119666, "grad_norm": 2.9320003986358643, "learning_rate": 4.221412277926706e-06, "loss": 4.963, "step": 94500 }, { "epoch": 4.351139592944993, "grad_norm": 2.2422378063201904, "learning_rate": 4.19211906866821e-06, "loss": 4.9208, "step": 94550 }, { "epoch": 4.353440638770321, "grad_norm": 2.2541680335998535, "learning_rate": 4.162923401521684e-06, "loss": 4.9744, "step": 94600 }, { "epoch": 4.355741684595649, "grad_norm": 2.8150317668914795, "learning_rate": 4.1338253386556166e-06, "loss": 4.9426, "step": 94650 }, { "epoch": 4.358042730420976, "grad_norm": 2.552889108657837, "learning_rate": 4.104824942030666e-06, "loss": 4.9327, "step": 94700 }, { "epoch": 4.360343776246304, "grad_norm": 2.3204643726348877, "learning_rate": 4.075922273399524e-06, "loss": 4.9854, "step": 94750 }, { "epoch": 4.362644822071632, "grad_norm": 2.234245777130127, "learning_rate": 4.047117394306777e-06, "loss": 4.9436, "step": 94800 }, { "epoch": 4.364945867896959, "grad_norm": 2.6050384044647217, "learning_rate": 4.018410366088804e-06, "loss": 4.9197, "step": 94850 }, { "epoch": 4.367246913722287, "grad_norm": 2.8665106296539307, "learning_rate": 3.989801249873593e-06, "loss": 4.9342, "step": 94900 }, { "epoch": 4.3695479595476145, "grad_norm": 2.768986701965332, "learning_rate": 3.961859368918858e-06, "loss": 4.8924, "step": 94950 }, { "epoch": 4.371849005372942, "grad_norm": 2.7984561920166016, "learning_rate": 3.9334442979928e-06, "loss": 4.9995, "step": 95000 }, { "epoch": 4.371849005372942, "eval_loss": 5.678844451904297, "eval_runtime": 33.1028, "eval_samples_per_second": 11.6, "eval_steps_per_second": 5.8, "eval_tts_loss": 9.321806622066815, "step": 95000 }, { "epoch": 4.37415005119827, "grad_norm": 2.769041061401367, "learning_rate": 3.9051273199940455e-06, "loss": 4.9815, "step": 95050 }, { "epoch": 4.3764510970235975, "grad_norm": 2.3030972480773926, "learning_rate": 3.876908495220033e-06, "loss": 4.9004, "step": 95100 }, { "epoch": 4.378752142848925, "grad_norm": 2.612943172454834, "learning_rate": 3.848787883759192e-06, "loss": 4.9345, "step": 95150 }, { "epoch": 4.381053188674253, "grad_norm": 2.815553665161133, "learning_rate": 3.820765545490823e-06, "loss": 4.969, "step": 95200 }, { "epoch": 4.38335423449958, "grad_norm": 2.8260498046875, "learning_rate": 3.7928415400849683e-06, "loss": 4.9135, "step": 95250 }, { "epoch": 4.385655280324908, "grad_norm": 2.7037949562072754, "learning_rate": 3.7650159270022745e-06, "loss": 4.9189, "step": 95300 }, { "epoch": 4.387956326150236, "grad_norm": 2.645465850830078, "learning_rate": 3.737288765493885e-06, "loss": 4.9661, "step": 95350 }, { "epoch": 4.390257371975563, "grad_norm": 2.275775194168091, "learning_rate": 3.7096601146012877e-06, "loss": 4.9557, "step": 95400 }, { "epoch": 4.39255841780089, "grad_norm": 2.8104965686798096, "learning_rate": 3.6821300331562224e-06, "loss": 4.9196, "step": 95450 }, { "epoch": 4.394859463626218, "grad_norm": 2.825328826904297, "learning_rate": 3.65469857978053e-06, "loss": 4.9656, "step": 95500 }, { "epoch": 4.397160509451545, "grad_norm": 2.970402240753174, "learning_rate": 3.6273658128860357e-06, "loss": 5.0275, "step": 95550 }, { "epoch": 4.399461555276873, "grad_norm": 2.230492115020752, "learning_rate": 3.6001317906744216e-06, "loss": 5.0396, "step": 95600 }, { "epoch": 4.401762601102201, "grad_norm": 2.658487558364868, "learning_rate": 3.572996571137105e-06, "loss": 4.9603, "step": 95650 }, { "epoch": 4.404063646927528, "grad_norm": 2.7010092735290527, "learning_rate": 3.545960212055116e-06, "loss": 4.9494, "step": 95700 }, { "epoch": 4.406364692752856, "grad_norm": 2.616330623626709, "learning_rate": 3.5190227709989863e-06, "loss": 4.9521, "step": 95750 }, { "epoch": 4.408665738578184, "grad_norm": 2.6494617462158203, "learning_rate": 3.4921843053285886e-06, "loss": 4.9551, "step": 95800 }, { "epoch": 4.410966784403511, "grad_norm": 2.8855628967285156, "learning_rate": 3.465444872193069e-06, "loss": 4.9162, "step": 95850 }, { "epoch": 4.413267830228839, "grad_norm": 2.9472038745880127, "learning_rate": 3.4388045285306593e-06, "loss": 4.9305, "step": 95900 }, { "epoch": 4.4155688760541665, "grad_norm": 2.7934985160827637, "learning_rate": 3.4122633310686435e-06, "loss": 4.9496, "step": 95950 }, { "epoch": 4.417869921879494, "grad_norm": 2.439134120941162, "learning_rate": 3.3858213363231415e-06, "loss": 4.9566, "step": 96000 }, { "epoch": 4.417869921879494, "eval_loss": 5.680164337158203, "eval_runtime": 33.1921, "eval_samples_per_second": 11.569, "eval_steps_per_second": 5.785, "eval_tts_loss": 9.328472778896515, "step": 96000 }, { "epoch": 4.420170967704822, "grad_norm": 2.765312433242798, "learning_rate": 3.3594786005990586e-06, "loss": 4.9377, "step": 96050 }, { "epoch": 4.422472013530149, "grad_norm": 2.464045524597168, "learning_rate": 3.3332351799899354e-06, "loss": 4.9079, "step": 96100 }, { "epoch": 4.424773059355477, "grad_norm": 2.796971321105957, "learning_rate": 3.307091130377815e-06, "loss": 4.8846, "step": 96150 }, { "epoch": 4.427074105180805, "grad_norm": 2.300090789794922, "learning_rate": 3.281046507433161e-06, "loss": 4.9408, "step": 96200 }, { "epoch": 4.429375151006132, "grad_norm": 2.481484889984131, "learning_rate": 3.2551013666147266e-06, "loss": 4.9107, "step": 96250 }, { "epoch": 4.43167619683146, "grad_norm": 2.670475959777832, "learning_rate": 3.229255763169403e-06, "loss": 4.9368, "step": 96300 }, { "epoch": 4.433977242656788, "grad_norm": 2.8900489807128906, "learning_rate": 3.2035097521321665e-06, "loss": 4.9616, "step": 96350 }, { "epoch": 4.436278288482115, "grad_norm": 2.6688849925994873, "learning_rate": 3.1778633883259026e-06, "loss": 4.9095, "step": 96400 }, { "epoch": 4.438579334307443, "grad_norm": 2.6334035396575928, "learning_rate": 3.1523167263613042e-06, "loss": 4.9026, "step": 96450 }, { "epoch": 4.4408803801327705, "grad_norm": 2.841860055923462, "learning_rate": 3.1268698206367687e-06, "loss": 4.9327, "step": 96500 }, { "epoch": 4.443181425958098, "grad_norm": 2.5172505378723145, "learning_rate": 3.101522725338285e-06, "loss": 4.8795, "step": 96550 }, { "epoch": 4.445482471783426, "grad_norm": 2.743206024169922, "learning_rate": 3.0762754944392958e-06, "loss": 4.9006, "step": 96600 }, { "epoch": 4.4477835176087535, "grad_norm": 2.925295114517212, "learning_rate": 3.051128181700591e-06, "loss": 4.963, "step": 96650 }, { "epoch": 4.450084563434081, "grad_norm": 2.951690196990967, "learning_rate": 3.02608084067022e-06, "loss": 4.939, "step": 96700 }, { "epoch": 4.452385609259409, "grad_norm": 3.2038767337799072, "learning_rate": 3.0011335246833193e-06, "loss": 4.9009, "step": 96750 }, { "epoch": 4.454686655084736, "grad_norm": 2.2985875606536865, "learning_rate": 2.976286286862079e-06, "loss": 4.9868, "step": 96800 }, { "epoch": 4.456987700910064, "grad_norm": 2.6529111862182617, "learning_rate": 2.951539180115548e-06, "loss": 4.8608, "step": 96850 }, { "epoch": 4.459288746735391, "grad_norm": 2.827261209487915, "learning_rate": 2.9268922571395797e-06, "loss": 4.8992, "step": 96900 }, { "epoch": 4.461589792560719, "grad_norm": 2.702986717224121, "learning_rate": 2.902345570416687e-06, "loss": 4.9245, "step": 96950 }, { "epoch": 4.463890838386046, "grad_norm": 2.3246467113494873, "learning_rate": 2.8778991722159476e-06, "loss": 4.9339, "step": 97000 }, { "epoch": 4.463890838386046, "eval_loss": 5.68049955368042, "eval_runtime": 33.2125, "eval_samples_per_second": 11.562, "eval_steps_per_second": 5.781, "eval_tts_loss": 9.329744280777785, "step": 97000 }, { "epoch": 4.466191884211374, "grad_norm": 2.637733221054077, "learning_rate": 2.8535531145928884e-06, "loss": 4.8988, "step": 97050 }, { "epoch": 4.468492930036701, "grad_norm": 2.5782065391540527, "learning_rate": 2.829791378513441e-06, "loss": 4.9646, "step": 97100 }, { "epoch": 4.470793975862029, "grad_norm": 2.7204387187957764, "learning_rate": 2.805644147972042e-06, "loss": 4.9783, "step": 97150 }, { "epoch": 4.473095021687357, "grad_norm": 2.453364372253418, "learning_rate": 2.7815974118662824e-06, "loss": 4.9549, "step": 97200 }, { "epoch": 4.475396067512684, "grad_norm": 3.313816547393799, "learning_rate": 2.75765122140067e-06, "loss": 4.9501, "step": 97250 }, { "epoch": 4.477697113338012, "grad_norm": 3.019824981689453, "learning_rate": 2.7338056275656056e-06, "loss": 4.8977, "step": 97300 }, { "epoch": 4.4799981591633395, "grad_norm": 2.641376256942749, "learning_rate": 2.7100606811372664e-06, "loss": 4.9102, "step": 97350 }, { "epoch": 4.482299204988667, "grad_norm": 2.7181406021118164, "learning_rate": 2.6864164326775354e-06, "loss": 4.9606, "step": 97400 }, { "epoch": 4.484600250813995, "grad_norm": 2.706712245941162, "learning_rate": 2.6628729325338442e-06, "loss": 4.9555, "step": 97450 }, { "epoch": 4.4869012966393225, "grad_norm": 2.717064380645752, "learning_rate": 2.639430230839135e-06, "loss": 4.9223, "step": 97500 }, { "epoch": 4.48920234246465, "grad_norm": 2.9090964794158936, "learning_rate": 2.6160883775116718e-06, "loss": 4.9483, "step": 97550 }, { "epoch": 4.491503388289978, "grad_norm": 2.4921481609344482, "learning_rate": 2.5928474222550012e-06, "loss": 4.929, "step": 97600 }, { "epoch": 4.493804434115305, "grad_norm": 2.7136213779449463, "learning_rate": 2.569707414557798e-06, "loss": 4.8908, "step": 97650 }, { "epoch": 4.496105479940633, "grad_norm": 2.4080166816711426, "learning_rate": 2.546668403693797e-06, "loss": 4.8863, "step": 97700 }, { "epoch": 4.498406525765961, "grad_norm": 2.7067811489105225, "learning_rate": 2.5237304387216777e-06, "loss": 4.9743, "step": 97750 }, { "epoch": 4.500707571591288, "grad_norm": 2.41428542137146, "learning_rate": 2.5008935684849357e-06, "loss": 4.8811, "step": 97800 }, { "epoch": 4.503008617416616, "grad_norm": 2.815544843673706, "learning_rate": 2.4786115646308226e-06, "loss": 4.998, "step": 97850 }, { "epoch": 4.505309663241944, "grad_norm": 1.8985440731048584, "learning_rate": 2.4559750052256082e-06, "loss": 4.9947, "step": 97900 }, { "epoch": 4.507610709067271, "grad_norm": 2.4579977989196777, "learning_rate": 2.4334396848324337e-06, "loss": 4.9729, "step": 97950 }, { "epoch": 4.509911754892599, "grad_norm": 2.3383259773254395, "learning_rate": 2.4110056514374313e-06, "loss": 4.9476, "step": 98000 }, { "epoch": 4.509911754892599, "eval_loss": 5.68021821975708, "eval_runtime": 34.9078, "eval_samples_per_second": 11.0, "eval_steps_per_second": 5.5, "eval_tts_loss": 9.33405117698406, "step": 98000 }, { "epoch": 4.5122128007179265, "grad_norm": 2.444965362548828, "learning_rate": 2.3886729528110495e-06, "loss": 4.8585, "step": 98050 }, { "epoch": 4.514513846543254, "grad_norm": 2.5604538917541504, "learning_rate": 2.366441636507949e-06, "loss": 5.0007, "step": 98100 }, { "epoch": 4.516814892368582, "grad_norm": 2.7612502574920654, "learning_rate": 2.3443117498669178e-06, "loss": 4.9423, "step": 98150 }, { "epoch": 4.519115938193909, "grad_norm": 2.63486647605896, "learning_rate": 2.3222833400107615e-06, "loss": 4.949, "step": 98200 }, { "epoch": 4.521416984019237, "grad_norm": 2.565378427505493, "learning_rate": 2.3003564538462252e-06, "loss": 4.8717, "step": 98250 }, { "epoch": 4.523718029844565, "grad_norm": 2.9743268489837646, "learning_rate": 2.2785311380638218e-06, "loss": 4.9079, "step": 98300 }, { "epoch": 4.5260190756698915, "grad_norm": 2.619223117828369, "learning_rate": 2.256807439137826e-06, "loss": 4.9379, "step": 98350 }, { "epoch": 4.52832012149522, "grad_norm": 2.2771501541137695, "learning_rate": 2.2351854033261122e-06, "loss": 4.9124, "step": 98400 }, { "epoch": 4.530621167320547, "grad_norm": 2.3378474712371826, "learning_rate": 2.2136650766700908e-06, "loss": 4.9547, "step": 98450 }, { "epoch": 4.532922213145875, "grad_norm": 2.4339568614959717, "learning_rate": 2.1922465049945773e-06, "loss": 4.8822, "step": 98500 }, { "epoch": 4.535223258971202, "grad_norm": 2.3399932384490967, "learning_rate": 2.170929733907723e-06, "loss": 4.904, "step": 98550 }, { "epoch": 4.53752430479653, "grad_norm": 2.650240659713745, "learning_rate": 2.1497148088009068e-06, "loss": 4.9557, "step": 98600 }, { "epoch": 4.539825350621857, "grad_norm": 2.8209807872772217, "learning_rate": 2.1286017748486488e-06, "loss": 4.8756, "step": 98650 }, { "epoch": 4.542126396447185, "grad_norm": 2.7290356159210205, "learning_rate": 2.1075906770084863e-06, "loss": 4.8983, "step": 98700 }, { "epoch": 4.544427442272513, "grad_norm": 2.846173048019409, "learning_rate": 2.0866815600209034e-06, "loss": 4.9577, "step": 98750 }, { "epoch": 4.54672848809784, "grad_norm": 2.816352128982544, "learning_rate": 2.0658744684092403e-06, "loss": 4.9076, "step": 98800 }, { "epoch": 4.549029533923168, "grad_norm": 2.602374315261841, "learning_rate": 2.0451694464795725e-06, "loss": 4.9418, "step": 98850 }, { "epoch": 4.5513305797484955, "grad_norm": 2.383106231689453, "learning_rate": 2.0245665383206325e-06, "loss": 4.8878, "step": 98900 }, { "epoch": 4.553631625573823, "grad_norm": 2.5587480068206787, "learning_rate": 2.004065787803716e-06, "loss": 4.9493, "step": 98950 }, { "epoch": 4.555932671399151, "grad_norm": 2.5952608585357666, "learning_rate": 1.9836672385825983e-06, "loss": 4.9341, "step": 99000 }, { "epoch": 4.555932671399151, "eval_loss": 5.680870056152344, "eval_runtime": 32.9306, "eval_samples_per_second": 11.661, "eval_steps_per_second": 5.83, "eval_tts_loss": 9.328455072720036, "step": 99000 }, { "epoch": 4.5582337172244785, "grad_norm": 2.5696303844451904, "learning_rate": 1.9633709340934116e-06, "loss": 4.9358, "step": 99050 }, { "epoch": 4.560534763049806, "grad_norm": 3.0702054500579834, "learning_rate": 1.9431769175545846e-06, "loss": 4.9556, "step": 99100 }, { "epoch": 4.562835808875134, "grad_norm": 2.621656656265259, "learning_rate": 1.9230852319667316e-06, "loss": 4.9203, "step": 99150 }, { "epoch": 4.565136854700461, "grad_norm": 3.1234617233276367, "learning_rate": 1.9030959201125631e-06, "loss": 4.9338, "step": 99200 }, { "epoch": 4.567437900525789, "grad_norm": 2.6556556224823, "learning_rate": 1.8832090245568034e-06, "loss": 4.8807, "step": 99250 }, { "epoch": 4.569738946351117, "grad_norm": 2.8958590030670166, "learning_rate": 1.8634245876460953e-06, "loss": 4.9994, "step": 99300 }, { "epoch": 4.572039992176444, "grad_norm": 2.4126973152160645, "learning_rate": 1.8437426515089006e-06, "loss": 4.9562, "step": 99350 }, { "epoch": 4.574341038001772, "grad_norm": 2.059985876083374, "learning_rate": 1.8241632580554169e-06, "loss": 4.9111, "step": 99400 }, { "epoch": 4.5766420838271, "grad_norm": 2.619472026824951, "learning_rate": 1.804686448977494e-06, "loss": 4.8848, "step": 99450 }, { "epoch": 4.578943129652427, "grad_norm": 2.500753879547119, "learning_rate": 1.7853122657485567e-06, "loss": 4.9496, "step": 99500 }, { "epoch": 4.581244175477755, "grad_norm": 2.4524269104003906, "learning_rate": 1.7660407496234876e-06, "loss": 4.9079, "step": 99550 }, { "epoch": 4.5835452213030825, "grad_norm": 2.433594226837158, "learning_rate": 1.7468719416385393e-06, "loss": 4.937, "step": 99600 }, { "epoch": 4.58584626712841, "grad_norm": 2.4936323165893555, "learning_rate": 1.7278058826112886e-06, "loss": 4.9127, "step": 99650 }, { "epoch": 4.588147312953738, "grad_norm": 2.664189577102661, "learning_rate": 1.7088426131404933e-06, "loss": 4.9445, "step": 99700 }, { "epoch": 4.590448358779065, "grad_norm": 2.1606667041778564, "learning_rate": 1.6899821736060529e-06, "loss": 4.9695, "step": 99750 }, { "epoch": 4.592749404604393, "grad_norm": 2.475398302078247, "learning_rate": 1.6712246041688972e-06, "loss": 4.9838, "step": 99800 }, { "epoch": 4.595050450429721, "grad_norm": 2.5283830165863037, "learning_rate": 1.6525699447709098e-06, "loss": 4.9915, "step": 99850 }, { "epoch": 4.5973514962550475, "grad_norm": 2.233774185180664, "learning_rate": 1.6340182351348266e-06, "loss": 4.9693, "step": 99900 }, { "epoch": 4.599652542080376, "grad_norm": 2.912449598312378, "learning_rate": 1.6155695147641925e-06, "loss": 4.9467, "step": 99950 }, { "epoch": 4.601953587905703, "grad_norm": 2.5109288692474365, "learning_rate": 1.5972238229432279e-06, "loss": 4.9133, "step": 100000 }, { "epoch": 4.601953587905703, "eval_loss": 5.6785407066345215, "eval_runtime": 34.8739, "eval_samples_per_second": 11.011, "eval_steps_per_second": 5.506, "eval_tts_loss": 9.332583423361912, "step": 100000 }, { "epoch": 4.60425463373103, "grad_norm": 2.881924867630005, "learning_rate": 1.5789811987367675e-06, "loss": 4.8805, "step": 100050 }, { "epoch": 4.606555679556358, "grad_norm": 2.44765305519104, "learning_rate": 1.5608416809901882e-06, "loss": 4.8991, "step": 100100 }, { "epoch": 4.608856725381686, "grad_norm": 2.5008325576782227, "learning_rate": 1.542805308329298e-06, "loss": 4.9804, "step": 100150 }, { "epoch": 4.611157771207013, "grad_norm": 2.615161180496216, "learning_rate": 1.524872119160292e-06, "loss": 4.8991, "step": 100200 }, { "epoch": 4.613458817032341, "grad_norm": 2.3106582164764404, "learning_rate": 1.5070421516696353e-06, "loss": 4.8969, "step": 100250 }, { "epoch": 4.615759862857669, "grad_norm": 2.350234031677246, "learning_rate": 1.489315443823991e-06, "loss": 4.9601, "step": 100300 }, { "epoch": 4.618060908682996, "grad_norm": 2.6425936222076416, "learning_rate": 1.4716920333701534e-06, "loss": 4.9984, "step": 100350 }, { "epoch": 4.620361954508324, "grad_norm": 2.9497785568237305, "learning_rate": 1.4541719578349488e-06, "loss": 4.8953, "step": 100400 }, { "epoch": 4.6226630003336515, "grad_norm": 2.536822557449341, "learning_rate": 1.4367552545251905e-06, "loss": 4.9373, "step": 100450 }, { "epoch": 4.624964046158979, "grad_norm": 2.1799700260162354, "learning_rate": 1.41944196052754e-06, "loss": 4.9222, "step": 100500 }, { "epoch": 4.627265091984307, "grad_norm": 2.7928173542022705, "learning_rate": 1.4022321127084737e-06, "loss": 4.9494, "step": 100550 }, { "epoch": 4.629566137809634, "grad_norm": 2.839073896408081, "learning_rate": 1.3851257477142e-06, "loss": 4.8917, "step": 100600 }, { "epoch": 4.631867183634962, "grad_norm": 2.6892263889312744, "learning_rate": 1.3681229019705643e-06, "loss": 4.9696, "step": 100650 }, { "epoch": 4.63416822946029, "grad_norm": 2.6419594287872314, "learning_rate": 1.351223611682989e-06, "loss": 4.9139, "step": 100700 }, { "epoch": 4.636469275285617, "grad_norm": 2.7339677810668945, "learning_rate": 1.3344279128363778e-06, "loss": 5.0286, "step": 100750 }, { "epoch": 4.638770321110945, "grad_norm": 2.5983216762542725, "learning_rate": 1.3177358411950668e-06, "loss": 4.9363, "step": 100800 }, { "epoch": 4.641071366936273, "grad_norm": 2.3199923038482666, "learning_rate": 1.3011474323027074e-06, "loss": 4.9999, "step": 100850 }, { "epoch": 4.6433724127616, "grad_norm": 2.283487558364868, "learning_rate": 1.2846627214822382e-06, "loss": 4.8808, "step": 100900 }, { "epoch": 4.645673458586928, "grad_norm": 2.314702272415161, "learning_rate": 1.2682817438357642e-06, "loss": 4.9613, "step": 100950 }, { "epoch": 4.647974504412256, "grad_norm": 2.938246965408325, "learning_rate": 1.2520045342445164e-06, "loss": 4.9292, "step": 101000 }, { "epoch": 4.647974504412256, "eval_loss": 5.677786350250244, "eval_runtime": 34.7565, "eval_samples_per_second": 11.048, "eval_steps_per_second": 5.524, "eval_tts_loss": 9.33160708535609, "step": 101000 }, { "epoch": 4.650275550237583, "grad_norm": 3.057169198989868, "learning_rate": 1.2358311273687696e-06, "loss": 4.9679, "step": 101050 }, { "epoch": 4.652576596062911, "grad_norm": 2.5875933170318604, "learning_rate": 1.2197615576477417e-06, "loss": 4.9252, "step": 101100 }, { "epoch": 4.6548776418882385, "grad_norm": 2.5108046531677246, "learning_rate": 1.2037958592995724e-06, "loss": 4.8977, "step": 101150 }, { "epoch": 4.657178687713566, "grad_norm": 2.690922498703003, "learning_rate": 1.1879340663212002e-06, "loss": 4.9365, "step": 101200 }, { "epoch": 4.659479733538894, "grad_norm": 2.1565539836883545, "learning_rate": 1.1721762124883019e-06, "loss": 4.8915, "step": 101250 }, { "epoch": 4.661780779364221, "grad_norm": 2.7533767223358154, "learning_rate": 1.1565223313552643e-06, "loss": 4.912, "step": 101300 }, { "epoch": 4.664081825189548, "grad_norm": 3.154649257659912, "learning_rate": 1.1409724562550406e-06, "loss": 4.9524, "step": 101350 }, { "epoch": 4.666382871014877, "grad_norm": 2.9312243461608887, "learning_rate": 1.1255266202991389e-06, "loss": 4.8855, "step": 101400 }, { "epoch": 4.6686839168402035, "grad_norm": 2.4187536239624023, "learning_rate": 1.1101848563775163e-06, "loss": 4.9802, "step": 101450 }, { "epoch": 4.670984962665532, "grad_norm": 2.9323229789733887, "learning_rate": 1.0949471971585246e-06, "loss": 4.9253, "step": 101500 }, { "epoch": 4.673286008490859, "grad_norm": 2.3062210083007812, "learning_rate": 1.0798136750888476e-06, "loss": 4.844, "step": 101550 }, { "epoch": 4.675587054316186, "grad_norm": 3.0856621265411377, "learning_rate": 1.0647843223934028e-06, "loss": 4.9659, "step": 101600 }, { "epoch": 4.677888100141514, "grad_norm": 2.496002435684204, "learning_rate": 1.0498591710753126e-06, "loss": 4.9148, "step": 101650 }, { "epoch": 4.680189145966842, "grad_norm": 2.5412819385528564, "learning_rate": 1.0350382529158043e-06, "loss": 4.9533, "step": 101700 }, { "epoch": 4.682490191792169, "grad_norm": 2.6318113803863525, "learning_rate": 1.0203215994741556e-06, "loss": 4.9336, "step": 101750 }, { "epoch": 4.684791237617497, "grad_norm": 3.0665981769561768, "learning_rate": 1.0057092420876269e-06, "loss": 4.9408, "step": 101800 }, { "epoch": 4.687092283442825, "grad_norm": 2.4351532459259033, "learning_rate": 9.912012118713898e-07, "loss": 4.9675, "step": 101850 }, { "epoch": 4.689393329268152, "grad_norm": 2.9481053352355957, "learning_rate": 9.77084590253763e-07, "loss": 4.8896, "step": 101900 }, { "epoch": 4.69169437509348, "grad_norm": 2.7210726737976074, "learning_rate": 9.62783218761143e-07, "loss": 4.9012, "step": 101950 }, { "epoch": 4.6939954209188075, "grad_norm": 2.8957622051239014, "learning_rate": 9.485862658443812e-07, "loss": 4.9575, "step": 102000 }, { "epoch": 4.6939954209188075, "eval_loss": 5.678743362426758, "eval_runtime": 33.3391, "eval_samples_per_second": 11.518, "eval_steps_per_second": 5.759, "eval_tts_loss": 9.337093838506295, "step": 102000 }, { "epoch": 4.696296466744135, "grad_norm": 2.884580135345459, "learning_rate": 9.344937617340898e-07, "loss": 4.9248, "step": 102050 }, { "epoch": 4.698597512569463, "grad_norm": 2.360043525695801, "learning_rate": 9.205057364384872e-07, "loss": 4.9822, "step": 102100 }, { "epoch": 4.70089855839479, "grad_norm": 2.644174337387085, "learning_rate": 9.066222197433138e-07, "loss": 4.8702, "step": 102150 }, { "epoch": 4.703199604220118, "grad_norm": 2.578793525695801, "learning_rate": 8.92843241211766e-07, "loss": 4.9325, "step": 102200 }, { "epoch": 4.705500650045446, "grad_norm": 2.5854337215423584, "learning_rate": 8.791688301844403e-07, "loss": 4.9065, "step": 102250 }, { "epoch": 4.707801695870773, "grad_norm": 2.5102357864379883, "learning_rate": 8.655990157792726e-07, "loss": 4.9669, "step": 102300 }, { "epoch": 4.710102741696101, "grad_norm": 3.339261770248413, "learning_rate": 8.521338268914824e-07, "loss": 4.9129, "step": 102350 }, { "epoch": 4.712403787521429, "grad_norm": 2.8844716548919678, "learning_rate": 8.387732921934788e-07, "loss": 4.9146, "step": 102400 }, { "epoch": 4.714704833346756, "grad_norm": 2.4359424114227295, "learning_rate": 8.255174401348431e-07, "loss": 4.9007, "step": 102450 }, { "epoch": 4.717005879172084, "grad_norm": 2.860389232635498, "learning_rate": 8.12366298942252e-07, "loss": 4.9032, "step": 102500 }, { "epoch": 4.719306924997412, "grad_norm": 2.6299808025360107, "learning_rate": 7.993198966193938e-07, "loss": 4.9112, "step": 102550 }, { "epoch": 4.721607970822739, "grad_norm": 2.6053991317749023, "learning_rate": 7.863782609469406e-07, "loss": 4.8932, "step": 102600 }, { "epoch": 4.723909016648067, "grad_norm": 3.325043201446533, "learning_rate": 7.735414194824819e-07, "loss": 4.9195, "step": 102650 }, { "epoch": 4.7262100624733945, "grad_norm": 2.7334275245666504, "learning_rate": 7.608093995604415e-07, "loss": 4.9408, "step": 102700 }, { "epoch": 4.728511108298722, "grad_norm": 2.373932123184204, "learning_rate": 7.481822282920547e-07, "loss": 4.9523, "step": 102750 }, { "epoch": 4.73081215412405, "grad_norm": 2.570046901702881, "learning_rate": 7.356599325653024e-07, "loss": 4.9191, "step": 102800 }, { "epoch": 4.733113199949377, "grad_norm": 2.2466907501220703, "learning_rate": 7.232425390448272e-07, "loss": 4.9136, "step": 102850 }, { "epoch": 4.735414245774704, "grad_norm": 2.5075948238372803, "learning_rate": 7.109300741719116e-07, "loss": 4.9233, "step": 102900 }, { "epoch": 4.737715291600033, "grad_norm": 3.1715610027313232, "learning_rate": 6.987225641643891e-07, "loss": 4.939, "step": 102950 }, { "epoch": 4.740016337425359, "grad_norm": 2.431623697280884, "learning_rate": 6.866200350166218e-07, "loss": 4.9124, "step": 103000 }, { "epoch": 4.740016337425359, "eval_loss": 5.6778388023376465, "eval_runtime": 32.9075, "eval_samples_per_second": 11.669, "eval_steps_per_second": 5.835, "eval_tts_loss": 9.333679660500833, "step": 103000 }, { "epoch": 4.742317383250687, "grad_norm": 2.0969090461730957, "learning_rate": 6.746225124994176e-07, "loss": 4.9657, "step": 103050 }, { "epoch": 4.744618429076015, "grad_norm": 2.663628101348877, "learning_rate": 6.627300221599908e-07, "loss": 4.9273, "step": 103100 }, { "epoch": 4.746919474901342, "grad_norm": 3.7632904052734375, "learning_rate": 6.509425893219012e-07, "loss": 4.9684, "step": 103150 }, { "epoch": 4.74922052072667, "grad_norm": 2.8351118564605713, "learning_rate": 6.392602390849988e-07, "loss": 4.9627, "step": 103200 }, { "epoch": 4.751521566551998, "grad_norm": 2.918797254562378, "learning_rate": 6.276829963253794e-07, "loss": 4.9821, "step": 103250 }, { "epoch": 4.753822612377325, "grad_norm": 3.037306308746338, "learning_rate": 6.162108856953286e-07, "loss": 4.9707, "step": 103300 }, { "epoch": 4.756123658202653, "grad_norm": 2.749039888381958, "learning_rate": 6.048439316232502e-07, "loss": 4.9084, "step": 103350 }, { "epoch": 4.758424704027981, "grad_norm": 2.600416898727417, "learning_rate": 5.935821583136492e-07, "loss": 4.9592, "step": 103400 }, { "epoch": 4.760725749853308, "grad_norm": 2.7607455253601074, "learning_rate": 5.824255897470487e-07, "loss": 4.9454, "step": 103450 }, { "epoch": 4.763026795678636, "grad_norm": 2.995832920074463, "learning_rate": 5.713742496799623e-07, "loss": 4.9448, "step": 103500 }, { "epoch": 4.7653278415039635, "grad_norm": 2.9551029205322266, "learning_rate": 5.60428161644816e-07, "loss": 4.9486, "step": 103550 }, { "epoch": 4.767628887329291, "grad_norm": 2.6423027515411377, "learning_rate": 5.49587348949937e-07, "loss": 4.9013, "step": 103600 }, { "epoch": 4.769929933154619, "grad_norm": 2.4621798992156982, "learning_rate": 5.388518346794602e-07, "loss": 4.9932, "step": 103650 }, { "epoch": 4.772230978979946, "grad_norm": 2.4298455715179443, "learning_rate": 5.282216416933161e-07, "loss": 4.9677, "step": 103700 }, { "epoch": 4.774532024805274, "grad_norm": 2.3390984535217285, "learning_rate": 5.176967926271647e-07, "loss": 4.928, "step": 103750 }, { "epoch": 4.776833070630602, "grad_norm": 2.9800028800964355, "learning_rate": 5.072773098923401e-07, "loss": 4.8998, "step": 103800 }, { "epoch": 4.779134116455929, "grad_norm": 2.249694347381592, "learning_rate": 4.969632156758219e-07, "loss": 4.9175, "step": 103850 }, { "epoch": 4.781435162281257, "grad_norm": 2.700409173965454, "learning_rate": 4.867545319401701e-07, "loss": 4.9035, "step": 103900 }, { "epoch": 4.783736208106585, "grad_norm": 2.519153118133545, "learning_rate": 4.7665128042349573e-07, "loss": 4.9496, "step": 103950 }, { "epoch": 4.786037253931912, "grad_norm": 3.556790828704834, "learning_rate": 4.6665348263939536e-07, "loss": 4.9067, "step": 104000 }, { "epoch": 4.786037253931912, "eval_loss": 5.678002834320068, "eval_runtime": 34.3512, "eval_samples_per_second": 11.179, "eval_steps_per_second": 5.589, "eval_tts_loss": 9.335159319797523, "step": 104000 }, { "epoch": 4.78833829975724, "grad_norm": 2.6343655586242676, "learning_rate": 4.567611598769228e-07, "loss": 4.9074, "step": 104050 }, { "epoch": 4.7906393455825675, "grad_norm": 2.6444284915924072, "learning_rate": 4.469743332005338e-07, "loss": 4.9405, "step": 104100 }, { "epoch": 4.792940391407895, "grad_norm": 2.464726686477661, "learning_rate": 4.3729302345004166e-07, "loss": 4.9708, "step": 104150 }, { "epoch": 4.795241437233223, "grad_norm": 3.8831639289855957, "learning_rate": 4.2771725124057806e-07, "loss": 4.8792, "step": 104200 }, { "epoch": 4.7975424830585505, "grad_norm": 2.8383708000183105, "learning_rate": 4.184354066495755e-07, "loss": 4.9004, "step": 104250 }, { "epoch": 4.799843528883878, "grad_norm": 2.8688228130340576, "learning_rate": 4.090686587104875e-07, "loss": 4.9201, "step": 104300 }, { "epoch": 4.802144574709205, "grad_norm": 2.409350633621216, "learning_rate": 3.998075084126529e-07, "loss": 4.8971, "step": 104350 }, { "epoch": 4.804445620534533, "grad_norm": 2.6027088165283203, "learning_rate": 3.906519754765248e-07, "loss": 5.0035, "step": 104400 }, { "epoch": 4.80674666635986, "grad_norm": 2.6355624198913574, "learning_rate": 3.816020793976638e-07, "loss": 4.919, "step": 104450 }, { "epoch": 4.809047712185188, "grad_norm": 2.6161468029022217, "learning_rate": 3.726578394466773e-07, "loss": 4.9573, "step": 104500 }, { "epoch": 4.811348758010515, "grad_norm": 2.807112455368042, "learning_rate": 3.638192746692026e-07, "loss": 4.968, "step": 104550 }, { "epoch": 4.813649803835843, "grad_norm": 2.690364122390747, "learning_rate": 3.550864038858459e-07, "loss": 4.9339, "step": 104600 }, { "epoch": 4.815950849661171, "grad_norm": 2.5228426456451416, "learning_rate": 3.4645924569216e-07, "loss": 4.8549, "step": 104650 }, { "epoch": 4.818251895486498, "grad_norm": 2.660872459411621, "learning_rate": 3.379378184585835e-07, "loss": 4.9103, "step": 104700 }, { "epoch": 4.820552941311826, "grad_norm": 2.904383659362793, "learning_rate": 3.295221403304294e-07, "loss": 4.9719, "step": 104750 }, { "epoch": 4.822853987137154, "grad_norm": 2.389962673187256, "learning_rate": 3.212122292278241e-07, "loss": 4.9299, "step": 104800 }, { "epoch": 4.825155032962481, "grad_norm": 2.7333154678344727, "learning_rate": 3.130081028456688e-07, "loss": 4.9329, "step": 104850 }, { "epoch": 4.827456078787809, "grad_norm": 2.3956751823425293, "learning_rate": 3.0490977865362815e-07, "loss": 4.9459, "step": 104900 }, { "epoch": 4.8297571246131366, "grad_norm": 2.398859977722168, "learning_rate": 2.969172738960524e-07, "loss": 4.9526, "step": 104950 }, { "epoch": 4.832058170438464, "grad_norm": 2.804928779602051, "learning_rate": 2.8903060559198335e-07, "loss": 4.9288, "step": 105000 }, { "epoch": 4.832058170438464, "eval_loss": 5.678267002105713, "eval_runtime": 34.178, "eval_samples_per_second": 11.235, "eval_steps_per_second": 5.618, "eval_tts_loss": 9.336976003650609, "step": 105000 }, { "epoch": 4.834359216263792, "grad_norm": 2.7038073539733887, "learning_rate": 2.812497905350819e-07, "loss": 4.8927, "step": 105050 }, { "epoch": 4.8366602620891195, "grad_norm": 2.4471280574798584, "learning_rate": 2.735748452936171e-07, "loss": 4.9623, "step": 105100 }, { "epoch": 4.838961307914447, "grad_norm": 2.022722005844116, "learning_rate": 2.660057862104104e-07, "loss": 4.9395, "step": 105150 }, { "epoch": 4.841262353739775, "grad_norm": 2.6244399547576904, "learning_rate": 2.585426294028304e-07, "loss": 4.9201, "step": 105200 }, { "epoch": 4.843563399565102, "grad_norm": 2.5239951610565186, "learning_rate": 2.511853907627204e-07, "loss": 4.9401, "step": 105250 }, { "epoch": 4.84586444539043, "grad_norm": 1.9543540477752686, "learning_rate": 2.4393408595639875e-07, "loss": 4.9751, "step": 105300 }, { "epoch": 4.848165491215758, "grad_norm": 2.3956873416900635, "learning_rate": 2.3678873042460302e-07, "loss": 4.8881, "step": 105350 }, { "epoch": 4.850466537041085, "grad_norm": 2.3740742206573486, "learning_rate": 2.297493393824679e-07, "loss": 4.8846, "step": 105400 }, { "epoch": 4.852767582866413, "grad_norm": 2.7427990436553955, "learning_rate": 2.228159278194919e-07, "loss": 4.9626, "step": 105450 }, { "epoch": 4.855068628691741, "grad_norm": 2.6279170513153076, "learning_rate": 2.1598851049949854e-07, "loss": 4.947, "step": 105500 }, { "epoch": 4.857369674517068, "grad_norm": 2.6171083450317383, "learning_rate": 2.0926710196061404e-07, "loss": 4.9544, "step": 105550 }, { "epoch": 4.859670720342396, "grad_norm": 2.7386815547943115, "learning_rate": 2.026517165152342e-07, "loss": 4.9288, "step": 105600 }, { "epoch": 4.8619717661677235, "grad_norm": 2.7789573669433594, "learning_rate": 1.9614236824998523e-07, "loss": 4.9173, "step": 105650 }, { "epoch": 4.864272811993051, "grad_norm": 2.6970458030700684, "learning_rate": 1.8973907102570189e-07, "loss": 4.9651, "step": 105700 }, { "epoch": 4.866573857818379, "grad_norm": 2.866112232208252, "learning_rate": 1.8344183847740503e-07, "loss": 4.922, "step": 105750 }, { "epoch": 4.868874903643706, "grad_norm": 2.490802764892578, "learning_rate": 1.7725068401425739e-07, "loss": 4.8887, "step": 105800 }, { "epoch": 4.871175949469034, "grad_norm": 2.6515631675720215, "learning_rate": 1.7116562081954113e-07, "loss": 4.8661, "step": 105850 }, { "epoch": 4.873476995294361, "grad_norm": 2.1467347145080566, "learning_rate": 1.6518666185063037e-07, "loss": 4.957, "step": 105900 }, { "epoch": 4.875778041119689, "grad_norm": 3.2990024089813232, "learning_rate": 1.5931381983896877e-07, "loss": 4.918, "step": 105950 }, { "epoch": 4.878079086945016, "grad_norm": 2.501412868499756, "learning_rate": 1.535471072900252e-07, "loss": 4.8902, "step": 106000 }, { "epoch": 4.878079086945016, "eval_loss": 5.677829742431641, "eval_runtime": 32.6971, "eval_samples_per_second": 11.744, "eval_steps_per_second": 5.872, "eval_tts_loss": 9.334307465951118, "step": 106000 }, { "epoch": 4.880380132770344, "grad_norm": 3.173262119293213, "learning_rate": 1.478865364832993e-07, "loss": 4.9177, "step": 106050 }, { "epoch": 4.882681178595671, "grad_norm": 2.9706802368164062, "learning_rate": 1.4233211947224934e-07, "loss": 4.99, "step": 106100 }, { "epoch": 4.884982224420999, "grad_norm": 3.015517473220825, "learning_rate": 1.3688386808431985e-07, "loss": 4.859, "step": 106150 }, { "epoch": 4.887283270246327, "grad_norm": 3.1198346614837646, "learning_rate": 1.3154179392086962e-07, "loss": 4.8891, "step": 106200 }, { "epoch": 4.889584316071654, "grad_norm": 3.3529605865478516, "learning_rate": 1.2630590835717715e-07, "loss": 4.958, "step": 106250 }, { "epoch": 4.891885361896982, "grad_norm": 3.0760412216186523, "learning_rate": 1.211762225423907e-07, "loss": 4.9196, "step": 106300 }, { "epoch": 4.89418640772231, "grad_norm": 2.513789653778076, "learning_rate": 1.1615274739954495e-07, "loss": 4.8735, "step": 106350 }, { "epoch": 4.896487453547637, "grad_norm": 2.1810100078582764, "learning_rate": 1.1123549362548336e-07, "loss": 4.848, "step": 106400 }, { "epoch": 4.898788499372965, "grad_norm": 2.3476059436798096, "learning_rate": 1.0642447169089132e-07, "loss": 4.9512, "step": 106450 }, { "epoch": 4.9010895451982925, "grad_norm": 2.690737009048462, "learning_rate": 1.0171969184021857e-07, "loss": 4.9523, "step": 106500 }, { "epoch": 4.90339059102362, "grad_norm": 2.9952900409698486, "learning_rate": 9.712116409171801e-08, "loss": 4.87, "step": 106550 }, { "epoch": 4.905691636848948, "grad_norm": 2.496338367462158, "learning_rate": 9.262889823736243e-08, "loss": 4.8978, "step": 106600 }, { "epoch": 4.9079926826742755, "grad_norm": 2.607712984085083, "learning_rate": 8.824290384287227e-08, "loss": 4.9671, "step": 106650 }, { "epoch": 4.910293728499603, "grad_norm": 2.359705686569214, "learning_rate": 8.396319024766564e-08, "loss": 4.8706, "step": 106700 }, { "epoch": 4.912594774324931, "grad_norm": 2.2572617530822754, "learning_rate": 7.978976656485837e-08, "loss": 4.9829, "step": 106750 }, { "epoch": 4.914895820150258, "grad_norm": 2.9874086380004883, "learning_rate": 7.572264168123067e-08, "loss": 4.9902, "step": 106800 }, { "epoch": 4.917196865975586, "grad_norm": 2.823206901550293, "learning_rate": 7.17618242572049e-08, "loss": 4.944, "step": 106850 }, { "epoch": 4.919497911800914, "grad_norm": 2.656986951828003, "learning_rate": 6.790732272685119e-08, "loss": 4.9266, "step": 106900 }, { "epoch": 4.921798957626241, "grad_norm": 2.7516186237335205, "learning_rate": 6.415914529784295e-08, "loss": 4.9177, "step": 106950 }, { "epoch": 4.924100003451569, "grad_norm": 3.0680367946624756, "learning_rate": 6.05172999514514e-08, "loss": 4.9755, "step": 107000 }, { "epoch": 4.924100003451569, "eval_loss": 5.677335262298584, "eval_runtime": 34.2155, "eval_samples_per_second": 11.223, "eval_steps_per_second": 5.611, "eval_tts_loss": 9.332648532018462, "step": 107000 }, { "epoch": 4.926401049276897, "grad_norm": 2.8450987339019775, "learning_rate": 5.6981794442528824e-08, "loss": 4.9353, "step": 107050 }, { "epoch": 4.928702095102224, "grad_norm": 2.833491086959839, "learning_rate": 5.355263629948648e-08, "loss": 4.9363, "step": 107100 }, { "epoch": 4.931003140927552, "grad_norm": 2.4455604553222656, "learning_rate": 5.0229832824283396e-08, "loss": 4.9182, "step": 107150 }, { "epoch": 4.9333041867528795, "grad_norm": 2.173336982727051, "learning_rate": 4.701339109240977e-08, "loss": 4.951, "step": 107200 }, { "epoch": 4.935605232578207, "grad_norm": 2.993123769760132, "learning_rate": 4.3903317952881385e-08, "loss": 4.8801, "step": 107250 }, { "epoch": 4.937906278403535, "grad_norm": 2.1944525241851807, "learning_rate": 4.0899620028200804e-08, "loss": 4.921, "step": 107300 }, { "epoch": 4.9402073242288616, "grad_norm": 2.7427523136138916, "learning_rate": 3.8002303714362864e-08, "loss": 4.9377, "step": 107350 }, { "epoch": 4.94250837005419, "grad_norm": 2.353147506713867, "learning_rate": 3.5211375180843606e-08, "loss": 4.9212, "step": 107400 }, { "epoch": 4.944809415879517, "grad_norm": 2.595211982727051, "learning_rate": 3.252684037056697e-08, "loss": 4.9031, "step": 107450 }, { "epoch": 4.9471104617048445, "grad_norm": 2.404395341873169, "learning_rate": 2.994870499992142e-08, "loss": 4.9661, "step": 107500 }, { "epoch": 4.949411507530172, "grad_norm": 2.551229476928711, "learning_rate": 2.7476974558710013e-08, "loss": 4.918, "step": 107550 }, { "epoch": 4.9517125533555, "grad_norm": 2.5151207447052, "learning_rate": 2.5111654310178144e-08, "loss": 4.9241, "step": 107600 }, { "epoch": 4.954013599180827, "grad_norm": 2.604194402694702, "learning_rate": 2.2852749290974696e-08, "loss": 4.9312, "step": 107650 }, { "epoch": 4.956314645006155, "grad_norm": 3.0182950496673584, "learning_rate": 2.0700264311157568e-08, "loss": 4.8306, "step": 107700 }, { "epoch": 4.958615690831483, "grad_norm": 2.690030336380005, "learning_rate": 1.865420395416595e-08, "loss": 4.8744, "step": 107750 }, { "epoch": 4.96091673665681, "grad_norm": 2.436689853668213, "learning_rate": 1.67145725768314e-08, "loss": 4.8807, "step": 107800 }, { "epoch": 4.963217782482138, "grad_norm": 2.4815070629119873, "learning_rate": 1.4881374309355656e-08, "loss": 4.8777, "step": 107850 }, { "epoch": 4.965518828307466, "grad_norm": 2.296535015106201, "learning_rate": 1.3154613055299525e-08, "loss": 4.8823, "step": 107900 }, { "epoch": 4.967819874132793, "grad_norm": 2.163233518600464, "learning_rate": 1.1534292491588438e-08, "loss": 4.994, "step": 107950 }, { "epoch": 4.970120919958121, "grad_norm": 2.8297841548919678, "learning_rate": 1.0020416068490245e-08, "loss": 4.9759, "step": 108000 }, { "epoch": 4.970120919958121, "eval_loss": 5.677609920501709, "eval_runtime": 33.7841, "eval_samples_per_second": 11.366, "eval_steps_per_second": 5.683, "eval_tts_loss": 9.33482673060466, "step": 108000 }, { "epoch": 4.9724219657834485, "grad_norm": 2.1278953552246094, "learning_rate": 8.612987009609663e-09, "loss": 4.9734, "step": 108050 }, { "epoch": 4.974723011608776, "grad_norm": 2.8732552528381348, "learning_rate": 7.312008311888274e-09, "loss": 4.986, "step": 108100 }, { "epoch": 4.977024057434104, "grad_norm": 2.6053032875061035, "learning_rate": 6.117482745610082e-09, "loss": 4.9327, "step": 108150 }, { "epoch": 4.9793251032594315, "grad_norm": 2.6100637912750244, "learning_rate": 5.029412854351545e-09, "loss": 4.9559, "step": 108200 }, { "epoch": 4.981626149084759, "grad_norm": 2.834904193878174, "learning_rate": 4.066389890794442e-09, "loss": 4.8548, "step": 108250 }, { "epoch": 4.983927194910087, "grad_norm": 2.3319642543792725, "learning_rate": 3.189108852957867e-09, "loss": 4.9586, "step": 108300 }, { "epoch": 4.986228240735414, "grad_norm": 2.7623283863067627, "learning_rate": 2.4182897257529757e-09, "loss": 4.9031, "step": 108350 }, { "epoch": 4.988529286560742, "grad_norm": 2.888908624649048, "learning_rate": 1.7539341505390383e-09, "loss": 4.926, "step": 108400 }, { "epoch": 4.99083033238607, "grad_norm": 2.4926486015319824, "learning_rate": 1.196043541973335e-09, "loss": 4.8925, "step": 108450 }, { "epoch": 4.993131378211397, "grad_norm": 2.533823251724243, "learning_rate": 7.446190880222581e-10, "loss": 4.9152, "step": 108500 }, { "epoch": 4.995432424036725, "grad_norm": 2.3294498920440674, "learning_rate": 3.996617499280042e-10, "loss": 4.9371, "step": 108550 }, { "epoch": 4.997733469862053, "grad_norm": 3.2538695335388184, "learning_rate": 1.6117226224188208e-10, "loss": 4.9338, "step": 108600 } ], "logging_steps": 50, "max_steps": 108645, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.757344604245459e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }