{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.4261345293796084, "eval_steps": 3000, "global_step": 99000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002152194592826305, "grad_norm": 2.6073930263519287, "learning_rate": 9.800000000000001e-06, "loss": 14.3937, "step": 50 }, { "epoch": 0.000430438918565261, "grad_norm": 48.563026428222656, "learning_rate": 1.9800000000000004e-05, "loss": 11.6426, "step": 100 }, { "epoch": 0.0006456583778478914, "grad_norm": 119.36799621582031, "learning_rate": 2.98e-05, "loss": 9.0542, "step": 150 }, { "epoch": 0.000860877837130522, "grad_norm": 9.547628402709961, "learning_rate": 3.9800000000000005e-05, "loss": 8.1742, "step": 200 }, { "epoch": 0.0010760972964131525, "grad_norm": 17.469379425048828, "learning_rate": 4.9800000000000004e-05, "loss": 7.8025, "step": 250 }, { "epoch": 0.001291316755695783, "grad_norm": 13.026924133300781, "learning_rate": 5.9800000000000003e-05, "loss": 7.8664, "step": 300 }, { "epoch": 0.0015065362149784135, "grad_norm": 4.849294185638428, "learning_rate": 6.98e-05, "loss": 8.077, "step": 350 }, { "epoch": 0.001721755674261044, "grad_norm": 4.3049187660217285, "learning_rate": 7.98e-05, "loss": 7.3555, "step": 400 }, { "epoch": 0.0019369751335436745, "grad_norm": 1.7300336360931396, "learning_rate": 8.98e-05, "loss": 7.0314, "step": 450 }, { "epoch": 0.002152194592826305, "grad_norm": 7.320311546325684, "learning_rate": 9.98e-05, "loss": 7.0003, "step": 500 }, { "epoch": 0.0023674140521089356, "grad_norm": 1.680544137954712, "learning_rate": 0.00010980000000000001, "loss": 6.6457, "step": 550 }, { "epoch": 0.002582633511391566, "grad_norm": 4.155694007873535, "learning_rate": 0.0001198, "loss": 7.3553, "step": 600 }, { "epoch": 0.0027978529706741964, "grad_norm": 3.384514570236206, "learning_rate": 0.0001298, "loss": 7.2104, "step": 650 }, { "epoch": 0.003013072429956827, "grad_norm": 4.987912178039551, "learning_rate": 0.0001398, "loss": 6.6966, "step": 700 }, { "epoch": 0.0032282918892394576, "grad_norm": 2.180516242980957, "learning_rate": 0.0001498, "loss": 7.2091, "step": 750 }, { "epoch": 0.003443511348522088, "grad_norm": 3.591073751449585, "learning_rate": 0.0001598, "loss": 7.2767, "step": 800 }, { "epoch": 0.0036587308078047183, "grad_norm": 1.9516600370407104, "learning_rate": 0.0001698, "loss": 6.8345, "step": 850 }, { "epoch": 0.003873950267087349, "grad_norm": 1.1745686531066895, "learning_rate": 0.0001798, "loss": 6.9251, "step": 900 }, { "epoch": 0.004089169726369979, "grad_norm": 2.7080631256103516, "learning_rate": 0.0001898, "loss": 6.3135, "step": 950 }, { "epoch": 0.00430438918565261, "grad_norm": 1.5969077348709106, "learning_rate": 0.0001998, "loss": 7.1232, "step": 1000 }, { "epoch": 0.00451960864493524, "grad_norm": 2.4516732692718506, "learning_rate": 0.0001999999995514373, "loss": 6.9686, "step": 1050 }, { "epoch": 0.004734828104217871, "grad_norm": 4.165557861328125, "learning_rate": 0.000199999998168945, "loss": 7.4058, "step": 1100 }, { "epoch": 0.004950047563500501, "grad_norm": 3.53717041015625, "learning_rate": 0.0001999999958523363, "loss": 6.8303, "step": 1150 }, { "epoch": 0.005165267022783132, "grad_norm": 2.105712413787842, "learning_rate": 0.0001999999926016112, "loss": 6.7287, "step": 1200 }, { "epoch": 0.005380486482065763, "grad_norm": 1.546506404876709, "learning_rate": 0.00019999998841676976, "loss": 6.9124, "step": 1250 }, { "epoch": 0.005595705941348393, "grad_norm": 2.323193073272705, "learning_rate": 0.000199999983297812, "loss": 6.8458, "step": 1300 }, { "epoch": 0.005810925400631024, "grad_norm": 1.3599646091461182, "learning_rate": 0.00019999997724473798, "loss": 6.2248, "step": 1350 }, { "epoch": 0.006026144859913654, "grad_norm": 1.221240520477295, "learning_rate": 0.00019999997025754772, "loss": 6.511, "step": 1400 }, { "epoch": 0.006241364319196284, "grad_norm": 6.495981693267822, "learning_rate": 0.00019999996233624134, "loss": 7.8413, "step": 1450 }, { "epoch": 0.006456583778478915, "grad_norm": 2.1630427837371826, "learning_rate": 0.00019999995348081887, "loss": 7.0452, "step": 1500 }, { "epoch": 0.006671803237761545, "grad_norm": 2.6963202953338623, "learning_rate": 0.00019999994369128047, "loss": 6.9581, "step": 1550 }, { "epoch": 0.006887022697044176, "grad_norm": 2.5625011920928955, "learning_rate": 0.0001999999329676261, "loss": 6.2387, "step": 1600 }, { "epoch": 0.0071022421563268065, "grad_norm": 1.3538457155227661, "learning_rate": 0.00019999992130985595, "loss": 7.2057, "step": 1650 }, { "epoch": 0.007317461615609437, "grad_norm": 0.46421146392822266, "learning_rate": 0.00019999990871797017, "loss": 6.2604, "step": 1700 }, { "epoch": 0.007532681074892068, "grad_norm": 2.8024303913116455, "learning_rate": 0.00019999989519196875, "loss": 6.976, "step": 1750 }, { "epoch": 0.007747900534174698, "grad_norm": 3.3166704177856445, "learning_rate": 0.00019999988073185192, "loss": 6.5064, "step": 1800 }, { "epoch": 0.007963119993457328, "grad_norm": 1.6762725114822388, "learning_rate": 0.00019999986533761978, "loss": 6.9396, "step": 1850 }, { "epoch": 0.008178339452739958, "grad_norm": 2.568375587463379, "learning_rate": 0.00019999984900927244, "loss": 6.8964, "step": 1900 }, { "epoch": 0.00839355891202259, "grad_norm": 3.00893497467041, "learning_rate": 0.00019999983174681015, "loss": 6.7745, "step": 1950 }, { "epoch": 0.00860877837130522, "grad_norm": 1.1491914987564087, "learning_rate": 0.00019999981355023295, "loss": 7.3272, "step": 2000 }, { "epoch": 0.00882399783058785, "grad_norm": 2.075793743133545, "learning_rate": 0.00019999979441954108, "loss": 6.9433, "step": 2050 }, { "epoch": 0.00903921728987048, "grad_norm": 1.6542335748672485, "learning_rate": 0.00019999977435473473, "loss": 7.1048, "step": 2100 }, { "epoch": 0.00925443674915311, "grad_norm": 3.5281338691711426, "learning_rate": 0.00019999975335581406, "loss": 6.9366, "step": 2150 }, { "epoch": 0.009469656208435743, "grad_norm": 2.0397300720214844, "learning_rate": 0.00019999973142277927, "loss": 7.1942, "step": 2200 }, { "epoch": 0.009684875667718373, "grad_norm": 4.444468975067139, "learning_rate": 0.00019999970855563058, "loss": 6.9963, "step": 2250 }, { "epoch": 0.009900095127001003, "grad_norm": 2.2611148357391357, "learning_rate": 0.00019999968475436814, "loss": 6.7503, "step": 2300 }, { "epoch": 0.010115314586283633, "grad_norm": 4.652219295501709, "learning_rate": 0.00019999966001899225, "loss": 6.5062, "step": 2350 }, { "epoch": 0.010330534045566263, "grad_norm": 0.6238192915916443, "learning_rate": 0.00019999963434950312, "loss": 7.2539, "step": 2400 }, { "epoch": 0.010545753504848895, "grad_norm": 1.5304611921310425, "learning_rate": 0.000199999607745901, "loss": 6.9958, "step": 2450 }, { "epoch": 0.010760972964131525, "grad_norm": 2.5568177700042725, "learning_rate": 0.00019999958020818607, "loss": 6.9074, "step": 2500 }, { "epoch": 0.010976192423414155, "grad_norm": 3.291961908340454, "learning_rate": 0.0001999995517363587, "loss": 6.8066, "step": 2550 }, { "epoch": 0.011191411882696786, "grad_norm": 3.9224555492401123, "learning_rate": 0.00019999952233041906, "loss": 6.8734, "step": 2600 }, { "epoch": 0.011406631341979416, "grad_norm": 2.0873894691467285, "learning_rate": 0.00019999949199036744, "loss": 6.4464, "step": 2650 }, { "epoch": 0.011621850801262048, "grad_norm": 0.4970775842666626, "learning_rate": 0.00019999946071620417, "loss": 6.7282, "step": 2700 }, { "epoch": 0.011837070260544678, "grad_norm": 2.0582058429718018, "learning_rate": 0.0001999994285079295, "loss": 6.3106, "step": 2750 }, { "epoch": 0.012052289719827308, "grad_norm": 2.1863248348236084, "learning_rate": 0.00019999939536554376, "loss": 7.1542, "step": 2800 }, { "epoch": 0.012267509179109938, "grad_norm": 0.6202731132507324, "learning_rate": 0.00019999936128904723, "loss": 5.967, "step": 2850 }, { "epoch": 0.012482728638392568, "grad_norm": 1.653417944908142, "learning_rate": 0.00019999932627844027, "loss": 6.842, "step": 2900 }, { "epoch": 0.0126979480976752, "grad_norm": 1.9701952934265137, "learning_rate": 0.00019999929033372315, "loss": 7.3646, "step": 2950 }, { "epoch": 0.01291316755695783, "grad_norm": 1.5888023376464844, "learning_rate": 0.00019999925345489625, "loss": 6.8507, "step": 3000 }, { "epoch": 0.01291316755695783, "eval_loss": 7.3343915939331055, "eval_runtime": 49.6627, "eval_samples_per_second": 12.887, "eval_steps_per_second": 6.443, "eval_tts_loss": 6.164592638475798, "step": 3000 }, { "epoch": 0.01312838701624046, "grad_norm": 2.1609249114990234, "learning_rate": 0.0001999992156419599, "loss": 7.197, "step": 3050 }, { "epoch": 0.01334360647552309, "grad_norm": 0.4406762719154358, "learning_rate": 0.00019999917689491445, "loss": 6.0601, "step": 3100 }, { "epoch": 0.01355882593480572, "grad_norm": 1.7598097324371338, "learning_rate": 0.00019999913721376026, "loss": 7.105, "step": 3150 }, { "epoch": 0.013774045394088353, "grad_norm": 1.8830268383026123, "learning_rate": 0.00019999909659849775, "loss": 6.4049, "step": 3200 }, { "epoch": 0.013989264853370983, "grad_norm": 0.5762608051300049, "learning_rate": 0.00019999905504912722, "loss": 6.4268, "step": 3250 }, { "epoch": 0.014204484312653613, "grad_norm": 3.0572683811187744, "learning_rate": 0.00019999901256564906, "loss": 6.9949, "step": 3300 }, { "epoch": 0.014419703771936243, "grad_norm": 1.173756718635559, "learning_rate": 0.00019999896914806375, "loss": 6.4548, "step": 3350 }, { "epoch": 0.014634923231218873, "grad_norm": 1.338756799697876, "learning_rate": 0.00019999892479637162, "loss": 6.5965, "step": 3400 }, { "epoch": 0.014850142690501505, "grad_norm": 3.4090278148651123, "learning_rate": 0.00019999887951057313, "loss": 6.742, "step": 3450 }, { "epoch": 0.015065362149784135, "grad_norm": 4.307085990905762, "learning_rate": 0.00019999883329066868, "loss": 6.6925, "step": 3500 }, { "epoch": 0.015280581609066765, "grad_norm": 1.0881993770599365, "learning_rate": 0.0001999987861366587, "loss": 6.4193, "step": 3550 }, { "epoch": 0.015495801068349396, "grad_norm": 2.6481592655181885, "learning_rate": 0.00019999873804854365, "loss": 6.6122, "step": 3600 }, { "epoch": 0.015711020527632027, "grad_norm": 2.3462109565734863, "learning_rate": 0.00019999868902632396, "loss": 7.0011, "step": 3650 }, { "epoch": 0.015926239986914656, "grad_norm": 2.486067056655884, "learning_rate": 0.00019999863907000008, "loss": 6.5551, "step": 3700 }, { "epoch": 0.016141459446197288, "grad_norm": 0.6153894066810608, "learning_rate": 0.0001999985881795725, "loss": 6.5401, "step": 3750 }, { "epoch": 0.016356678905479916, "grad_norm": 2.104681968688965, "learning_rate": 0.00019999853635504166, "loss": 6.9376, "step": 3800 }, { "epoch": 0.016571898364762548, "grad_norm": 2.123143196105957, "learning_rate": 0.0001999984835964081, "loss": 6.3114, "step": 3850 }, { "epoch": 0.01678711782404518, "grad_norm": 2.1063411235809326, "learning_rate": 0.00019999842990367226, "loss": 6.4982, "step": 3900 }, { "epoch": 0.01700233728332781, "grad_norm": 2.31050705909729, "learning_rate": 0.00019999837527683468, "loss": 6.7584, "step": 3950 }, { "epoch": 0.01721755674261044, "grad_norm": 2.675370693206787, "learning_rate": 0.00019999831971589585, "loss": 6.7453, "step": 4000 }, { "epoch": 0.01743277620189307, "grad_norm": 2.255415678024292, "learning_rate": 0.0001999982632208563, "loss": 7.1191, "step": 4050 }, { "epoch": 0.0176479956611757, "grad_norm": 2.7438063621520996, "learning_rate": 0.00019999820579171655, "loss": 7.4197, "step": 4100 }, { "epoch": 0.017863215120458333, "grad_norm": 1.4139623641967773, "learning_rate": 0.00019999814742847712, "loss": 6.4279, "step": 4150 }, { "epoch": 0.01807843457974096, "grad_norm": 0.9164754748344421, "learning_rate": 0.0001999980881311386, "loss": 6.4897, "step": 4200 }, { "epoch": 0.018293654039023593, "grad_norm": 1.496140956878662, "learning_rate": 0.0001999980278997015, "loss": 6.5977, "step": 4250 }, { "epoch": 0.01850887349830622, "grad_norm": 1.86050546169281, "learning_rate": 0.00019999796673416637, "loss": 6.6604, "step": 4300 }, { "epoch": 0.018724092957588853, "grad_norm": 1.4085026979446411, "learning_rate": 0.00019999790463453386, "loss": 6.7403, "step": 4350 }, { "epoch": 0.018939312416871485, "grad_norm": 2.2738709449768066, "learning_rate": 0.00019999784160080447, "loss": 6.2971, "step": 4400 }, { "epoch": 0.019154531876154113, "grad_norm": 1.0178765058517456, "learning_rate": 0.0001999977776329788, "loss": 6.535, "step": 4450 }, { "epoch": 0.019369751335436745, "grad_norm": 2.553838014602661, "learning_rate": 0.00019999771273105748, "loss": 6.4153, "step": 4500 }, { "epoch": 0.019584970794719374, "grad_norm": 2.5008866786956787, "learning_rate": 0.00019999764689504116, "loss": 7.1747, "step": 4550 }, { "epoch": 0.019800190254002006, "grad_norm": 1.8582850694656372, "learning_rate": 0.00019999758012493034, "loss": 6.9555, "step": 4600 }, { "epoch": 0.020015409713284638, "grad_norm": 2.053887128829956, "learning_rate": 0.00019999751242072571, "loss": 6.1801, "step": 4650 }, { "epoch": 0.020230629172567266, "grad_norm": 0.5575258731842041, "learning_rate": 0.0001999974437824279, "loss": 6.7331, "step": 4700 }, { "epoch": 0.020445848631849898, "grad_norm": 3.4684932231903076, "learning_rate": 0.00019999737421003754, "loss": 6.9731, "step": 4750 }, { "epoch": 0.020661068091132526, "grad_norm": 2.3210811614990234, "learning_rate": 0.0001999973037035553, "loss": 6.6728, "step": 4800 }, { "epoch": 0.020876287550415158, "grad_norm": 1.4208295345306396, "learning_rate": 0.0001999972322629818, "loss": 6.4296, "step": 4850 }, { "epoch": 0.02109150700969779, "grad_norm": 1.7994465827941895, "learning_rate": 0.00019999715988831774, "loss": 6.6867, "step": 4900 }, { "epoch": 0.02130672646898042, "grad_norm": 140.22865295410156, "learning_rate": 0.0001999970865795638, "loss": 6.839, "step": 4950 }, { "epoch": 0.02152194592826305, "grad_norm": 4.320616245269775, "learning_rate": 0.00019999701233672064, "loss": 6.8145, "step": 5000 }, { "epoch": 0.02173716538754568, "grad_norm": 2.0404884815216064, "learning_rate": 0.00019999693715978895, "loss": 6.2631, "step": 5050 }, { "epoch": 0.02195238484682831, "grad_norm": 1.7448879480361938, "learning_rate": 0.00019999686104876948, "loss": 6.6981, "step": 5100 }, { "epoch": 0.022167604306110943, "grad_norm": 1.8464744091033936, "learning_rate": 0.0001999967840036629, "loss": 6.1681, "step": 5150 }, { "epoch": 0.02238282376539357, "grad_norm": 2.9619882106781006, "learning_rate": 0.00019999670602446993, "loss": 6.4849, "step": 5200 }, { "epoch": 0.022598043224676203, "grad_norm": 1.745737075805664, "learning_rate": 0.00019999662711119132, "loss": 6.8544, "step": 5250 }, { "epoch": 0.02281326268395883, "grad_norm": 2.039433240890503, "learning_rate": 0.00019999654726382776, "loss": 6.378, "step": 5300 }, { "epoch": 0.023028482143241463, "grad_norm": 1.9177124500274658, "learning_rate": 0.00019999646648238007, "loss": 7.2146, "step": 5350 }, { "epoch": 0.023243701602524095, "grad_norm": 1.5384106636047363, "learning_rate": 0.00019999638476684894, "loss": 6.4529, "step": 5400 }, { "epoch": 0.023458921061806724, "grad_norm": 2.6210649013519287, "learning_rate": 0.00019999630211723517, "loss": 6.6326, "step": 5450 }, { "epoch": 0.023674140521089355, "grad_norm": 2.439842462539673, "learning_rate": 0.00019999621853353952, "loss": 6.5269, "step": 5500 }, { "epoch": 0.023889359980371984, "grad_norm": 2.1007227897644043, "learning_rate": 0.00019999613401576273, "loss": 6.7317, "step": 5550 }, { "epoch": 0.024104579439654616, "grad_norm": 2.07796573638916, "learning_rate": 0.0001999960485639057, "loss": 6.6005, "step": 5600 }, { "epoch": 0.024319798898937248, "grad_norm": 2.3872644901275635, "learning_rate": 0.0001999959621779691, "loss": 6.2064, "step": 5650 }, { "epoch": 0.024535018358219876, "grad_norm": 1.382962703704834, "learning_rate": 0.0001999958748579538, "loss": 6.6654, "step": 5700 }, { "epoch": 0.024750237817502508, "grad_norm": 0.9705161452293396, "learning_rate": 0.00019999578660386065, "loss": 6.3412, "step": 5750 }, { "epoch": 0.024965457276785136, "grad_norm": 1.5257275104522705, "learning_rate": 0.0001999956974156904, "loss": 6.4978, "step": 5800 }, { "epoch": 0.02518067673606777, "grad_norm": 4.197699069976807, "learning_rate": 0.00019999560729344392, "loss": 6.0099, "step": 5850 }, { "epoch": 0.0253958961953504, "grad_norm": 1.9446665048599243, "learning_rate": 0.00019999551623712207, "loss": 5.9875, "step": 5900 }, { "epoch": 0.02561111565463303, "grad_norm": 1.982170820236206, "learning_rate": 0.00019999542424672568, "loss": 6.4992, "step": 5950 }, { "epoch": 0.02582633511391566, "grad_norm": 1.7469029426574707, "learning_rate": 0.00019999533132225559, "loss": 6.3633, "step": 6000 }, { "epoch": 0.02582633511391566, "eval_loss": 7.0566253662109375, "eval_runtime": 49.5806, "eval_samples_per_second": 12.908, "eval_steps_per_second": 6.454, "eval_tts_loss": 5.875931752473971, "step": 6000 }, { "epoch": 0.02604155457319829, "grad_norm": 2.591013193130493, "learning_rate": 0.0001999952374637127, "loss": 6.8583, "step": 6050 }, { "epoch": 0.02625677403248092, "grad_norm": 2.089803695678711, "learning_rate": 0.00019999514267109788, "loss": 6.438, "step": 6100 }, { "epoch": 0.026471993491763553, "grad_norm": 0.5964639186859131, "learning_rate": 0.00019999504694441197, "loss": 6.7851, "step": 6150 }, { "epoch": 0.02668721295104618, "grad_norm": 1.7756956815719604, "learning_rate": 0.00019999495028365596, "loss": 6.2447, "step": 6200 }, { "epoch": 0.026902432410328813, "grad_norm": 0.22942449152469635, "learning_rate": 0.00019999485268883067, "loss": 6.2284, "step": 6250 }, { "epoch": 0.02711765186961144, "grad_norm": 2.517723560333252, "learning_rate": 0.00019999475415993705, "loss": 6.7713, "step": 6300 }, { "epoch": 0.027332871328894073, "grad_norm": 0.6978651881217957, "learning_rate": 0.000199994654696976, "loss": 6.0546, "step": 6350 }, { "epoch": 0.027548090788176705, "grad_norm": 1.194336175918579, "learning_rate": 0.00019999455429994847, "loss": 6.4271, "step": 6400 }, { "epoch": 0.027763310247459334, "grad_norm": 1.9801650047302246, "learning_rate": 0.00019999445296885537, "loss": 6.4588, "step": 6450 }, { "epoch": 0.027978529706741966, "grad_norm": 1.1488522291183472, "learning_rate": 0.0001999943507036977, "loss": 6.2738, "step": 6500 }, { "epoch": 0.028193749166024594, "grad_norm": 3.9114561080932617, "learning_rate": 0.0001999942475044763, "loss": 6.0109, "step": 6550 }, { "epoch": 0.028408968625307226, "grad_norm": 2.5312509536743164, "learning_rate": 0.00019999414337119227, "loss": 6.7639, "step": 6600 }, { "epoch": 0.028624188084589858, "grad_norm": 0.8824054002761841, "learning_rate": 0.0001999940383038465, "loss": 5.7257, "step": 6650 }, { "epoch": 0.028839407543872486, "grad_norm": 0.8979548811912537, "learning_rate": 0.00019999393230244004, "loss": 6.1756, "step": 6700 }, { "epoch": 0.029054627003155118, "grad_norm": 1.7945047616958618, "learning_rate": 0.0001999938253669738, "loss": 5.9278, "step": 6750 }, { "epoch": 0.029269846462437746, "grad_norm": 1.945326566696167, "learning_rate": 0.00019999371749744883, "loss": 5.817, "step": 6800 }, { "epoch": 0.02948506592172038, "grad_norm": 2.253286123275757, "learning_rate": 0.00019999360869386612, "loss": 7.0382, "step": 6850 }, { "epoch": 0.02970028538100301, "grad_norm": 1.7811031341552734, "learning_rate": 0.00019999349895622669, "loss": 6.156, "step": 6900 }, { "epoch": 0.02991550484028564, "grad_norm": 2.0919275283813477, "learning_rate": 0.00019999338828453157, "loss": 6.4961, "step": 6950 }, { "epoch": 0.03013072429956827, "grad_norm": 1.9256055355072021, "learning_rate": 0.00019999327667878175, "loss": 7.096, "step": 7000 }, { "epoch": 0.0303459437588509, "grad_norm": 1.8411363363265991, "learning_rate": 0.00019999316413897832, "loss": 6.8436, "step": 7050 }, { "epoch": 0.03056116321813353, "grad_norm": 1.8480933904647827, "learning_rate": 0.00019999305066512236, "loss": 6.8193, "step": 7100 }, { "epoch": 0.030776382677416163, "grad_norm": 2.2632932662963867, "learning_rate": 0.00019999293625721487, "loss": 6.402, "step": 7150 }, { "epoch": 0.03099160213669879, "grad_norm": 3.648071765899658, "learning_rate": 0.00019999282091525693, "loss": 6.1128, "step": 7200 }, { "epoch": 0.031206821595981423, "grad_norm": 2.308161973953247, "learning_rate": 0.00019999270463924966, "loss": 5.7086, "step": 7250 }, { "epoch": 0.031422041055264055, "grad_norm": 2.890892505645752, "learning_rate": 0.0001999925874291941, "loss": 6.5998, "step": 7300 }, { "epoch": 0.03163726051454668, "grad_norm": 2.657132148742676, "learning_rate": 0.00019999246928509136, "loss": 6.4928, "step": 7350 }, { "epoch": 0.03185247997382931, "grad_norm": 2.48996639251709, "learning_rate": 0.0001999923502069425, "loss": 6.7611, "step": 7400 }, { "epoch": 0.03206769943311195, "grad_norm": 1.876556396484375, "learning_rate": 0.00019999223019474873, "loss": 6.5462, "step": 7450 }, { "epoch": 0.032282918892394576, "grad_norm": 2.5409133434295654, "learning_rate": 0.0001999921092485111, "loss": 6.8456, "step": 7500 }, { "epoch": 0.032498138351677204, "grad_norm": 2.754026412963867, "learning_rate": 0.00019999198736823074, "loss": 6.1733, "step": 7550 }, { "epoch": 0.03271335781095983, "grad_norm": 1.9323228597640991, "learning_rate": 0.00019999186455390883, "loss": 6.2054, "step": 7600 }, { "epoch": 0.03292857727024247, "grad_norm": 3.4594311714172363, "learning_rate": 0.00019999174080554647, "loss": 6.767, "step": 7650 }, { "epoch": 0.033143796729525096, "grad_norm": 2.7913849353790283, "learning_rate": 0.00019999161612314486, "loss": 6.8333, "step": 7700 }, { "epoch": 0.033359016188807725, "grad_norm": 2.448808193206787, "learning_rate": 0.0001999914905067051, "loss": 6.2841, "step": 7750 }, { "epoch": 0.03357423564809036, "grad_norm": 1.970625877380371, "learning_rate": 0.00019999136395622847, "loss": 6.9459, "step": 7800 }, { "epoch": 0.03378945510737299, "grad_norm": 3.076575756072998, "learning_rate": 0.00019999123647171605, "loss": 5.8749, "step": 7850 }, { "epoch": 0.03400467456665562, "grad_norm": 1.665256381034851, "learning_rate": 0.00019999110805316906, "loss": 6.8159, "step": 7900 }, { "epoch": 0.03421989402593825, "grad_norm": 2.336392402648926, "learning_rate": 0.00019999097870058874, "loss": 6.5222, "step": 7950 }, { "epoch": 0.03443511348522088, "grad_norm": 1.6112903356552124, "learning_rate": 0.00019999084841397621, "loss": 5.965, "step": 8000 }, { "epoch": 0.03465033294450351, "grad_norm": 0.4891376495361328, "learning_rate": 0.00019999071719333279, "loss": 6.2601, "step": 8050 }, { "epoch": 0.03486555240378614, "grad_norm": 2.2167160511016846, "learning_rate": 0.00019999058503865965, "loss": 6.7744, "step": 8100 }, { "epoch": 0.03508077186306877, "grad_norm": 2.8674168586730957, "learning_rate": 0.00019999045194995802, "loss": 6.5251, "step": 8150 }, { "epoch": 0.0352959913223514, "grad_norm": 2.226119041442871, "learning_rate": 0.0001999903179272292, "loss": 6.8678, "step": 8200 }, { "epoch": 0.03551121078163403, "grad_norm": 2.152615547180176, "learning_rate": 0.00019999018297047437, "loss": 6.6898, "step": 8250 }, { "epoch": 0.035726430240916665, "grad_norm": 0.5630964636802673, "learning_rate": 0.00019999004707969479, "loss": 6.2041, "step": 8300 }, { "epoch": 0.035941649700199294, "grad_norm": 1.2390971183776855, "learning_rate": 0.0001999899102548918, "loss": 6.1318, "step": 8350 }, { "epoch": 0.03615686915948192, "grad_norm": 3.8274381160736084, "learning_rate": 0.00019998977249606663, "loss": 6.133, "step": 8400 }, { "epoch": 0.03637208861876456, "grad_norm": 0.9620758295059204, "learning_rate": 0.00019998963380322054, "loss": 6.4686, "step": 8450 }, { "epoch": 0.036587308078047186, "grad_norm": 1.3624557256698608, "learning_rate": 0.00019998949417635492, "loss": 5.8248, "step": 8500 }, { "epoch": 0.036802527537329814, "grad_norm": 2.6003293991088867, "learning_rate": 0.00019998935361547097, "loss": 5.7632, "step": 8550 }, { "epoch": 0.03701774699661244, "grad_norm": 1.0451719760894775, "learning_rate": 0.00019998921212057008, "loss": 6.3645, "step": 8600 }, { "epoch": 0.03723296645589508, "grad_norm": 2.616468906402588, "learning_rate": 0.0001999890696916535, "loss": 5.7281, "step": 8650 }, { "epoch": 0.037448185915177706, "grad_norm": 2.5309996604919434, "learning_rate": 0.0001999889263287226, "loss": 5.8425, "step": 8700 }, { "epoch": 0.037663405374460335, "grad_norm": 2.8271148204803467, "learning_rate": 0.00019998878203177873, "loss": 5.7994, "step": 8750 }, { "epoch": 0.03787862483374297, "grad_norm": 2.7787251472473145, "learning_rate": 0.00019998863680082324, "loss": 6.3239, "step": 8800 }, { "epoch": 0.0380938442930256, "grad_norm": 3.018136978149414, "learning_rate": 0.00019998849063585747, "loss": 5.6854, "step": 8850 }, { "epoch": 0.03830906375230823, "grad_norm": 0.6548991799354553, "learning_rate": 0.00019998834353688276, "loss": 5.7831, "step": 8900 }, { "epoch": 0.03852428321159086, "grad_norm": 1.1760902404785156, "learning_rate": 0.00019998819550390054, "loss": 6.4366, "step": 8950 }, { "epoch": 0.03873950267087349, "grad_norm": 2.670318603515625, "learning_rate": 0.00019998804653691217, "loss": 6.3096, "step": 9000 }, { "epoch": 0.03873950267087349, "eval_loss": 6.8365373611450195, "eval_runtime": 49.9266, "eval_samples_per_second": 12.819, "eval_steps_per_second": 6.409, "eval_tts_loss": 5.803004110855314, "step": 9000 }, { "epoch": 0.03895472213015612, "grad_norm": 2.490490674972534, "learning_rate": 0.000199987896635919, "loss": 6.578, "step": 9050 }, { "epoch": 0.03916994158943875, "grad_norm": 1.8307572603225708, "learning_rate": 0.0001999877458009225, "loss": 6.1396, "step": 9100 }, { "epoch": 0.03938516104872138, "grad_norm": 1.7277525663375854, "learning_rate": 0.00019998759403192401, "loss": 6.1055, "step": 9150 }, { "epoch": 0.03960038050800401, "grad_norm": 0.5808207988739014, "learning_rate": 0.000199987441328925, "loss": 6.072, "step": 9200 }, { "epoch": 0.03981559996728664, "grad_norm": 1.9747388362884521, "learning_rate": 0.00019998728769192692, "loss": 6.1097, "step": 9250 }, { "epoch": 0.040030819426569275, "grad_norm": 1.2256169319152832, "learning_rate": 0.0001999871331209311, "loss": 5.8204, "step": 9300 }, { "epoch": 0.040246038885851904, "grad_norm": 1.7142612934112549, "learning_rate": 0.0001999869776159391, "loss": 6.3311, "step": 9350 }, { "epoch": 0.04046125834513453, "grad_norm": 0.3137798011302948, "learning_rate": 0.00019998682117695232, "loss": 5.9763, "step": 9400 }, { "epoch": 0.04067647780441717, "grad_norm": 1.2712684869766235, "learning_rate": 0.0001999866638039722, "loss": 6.8616, "step": 9450 }, { "epoch": 0.040891697263699796, "grad_norm": 1.4483245611190796, "learning_rate": 0.00019998650549700022, "loss": 6.6387, "step": 9500 }, { "epoch": 0.041106916722982424, "grad_norm": 1.6184136867523193, "learning_rate": 0.0001999863462560379, "loss": 6.3182, "step": 9550 }, { "epoch": 0.04132213618226505, "grad_norm": 2.311795473098755, "learning_rate": 0.0001999861860810867, "loss": 5.7295, "step": 9600 }, { "epoch": 0.04153735564154769, "grad_norm": 0.9464931488037109, "learning_rate": 0.00019998602497214808, "loss": 5.6891, "step": 9650 }, { "epoch": 0.041752575100830316, "grad_norm": 1.5330201387405396, "learning_rate": 0.0001999858629292236, "loss": 6.0954, "step": 9700 }, { "epoch": 0.041967794560112945, "grad_norm": 1.5301202535629272, "learning_rate": 0.00019998569995231474, "loss": 5.9131, "step": 9750 }, { "epoch": 0.04218301401939558, "grad_norm": 1.1338653564453125, "learning_rate": 0.00019998553604142307, "loss": 5.9046, "step": 9800 }, { "epoch": 0.04239823347867821, "grad_norm": 4.303573131561279, "learning_rate": 0.00019998537119655008, "loss": 5.9846, "step": 9850 }, { "epoch": 0.04261345293796084, "grad_norm": 1.9508590698242188, "learning_rate": 0.0001999852054176973, "loss": 5.9664, "step": 9900 }, { "epoch": 0.04282867239724347, "grad_norm": 2.622410774230957, "learning_rate": 0.0001999850387048663, "loss": 6.3333, "step": 9950 }, { "epoch": 0.0430438918565261, "grad_norm": 2.6119353771209717, "learning_rate": 0.0001999848710580586, "loss": 5.7802, "step": 10000 }, { "epoch": 0.04325911131580873, "grad_norm": 1.684345006942749, "learning_rate": 0.00019998470247727584, "loss": 5.8853, "step": 10050 }, { "epoch": 0.04347433077509136, "grad_norm": 1.4450182914733887, "learning_rate": 0.00019998453296251958, "loss": 6.5049, "step": 10100 }, { "epoch": 0.04368955023437399, "grad_norm": 2.113241672515869, "learning_rate": 0.00019998436251379134, "loss": 5.9399, "step": 10150 }, { "epoch": 0.04390476969365662, "grad_norm": 1.862652063369751, "learning_rate": 0.00019998419113109275, "loss": 6.2848, "step": 10200 }, { "epoch": 0.04411998915293925, "grad_norm": 3.1360459327697754, "learning_rate": 0.00019998401881442542, "loss": 6.0439, "step": 10250 }, { "epoch": 0.044335208612221885, "grad_norm": 1.903953194618225, "learning_rate": 0.00019998384556379093, "loss": 5.9969, "step": 10300 }, { "epoch": 0.044550428071504514, "grad_norm": 1.5163097381591797, "learning_rate": 0.00019998367137919094, "loss": 6.2781, "step": 10350 }, { "epoch": 0.04476564753078714, "grad_norm": 1.2600642442703247, "learning_rate": 0.00019998349626062703, "loss": 5.9311, "step": 10400 }, { "epoch": 0.04498086699006978, "grad_norm": 1.4845930337905884, "learning_rate": 0.00019998332020810088, "loss": 6.3513, "step": 10450 }, { "epoch": 0.045196086449352406, "grad_norm": 3.19741153717041, "learning_rate": 0.0001999831432216141, "loss": 6.2932, "step": 10500 }, { "epoch": 0.045411305908635034, "grad_norm": 1.843291997909546, "learning_rate": 0.0001999829653011684, "loss": 6.3465, "step": 10550 }, { "epoch": 0.04562652536791766, "grad_norm": 2.138098955154419, "learning_rate": 0.00019998278644676537, "loss": 5.8819, "step": 10600 }, { "epoch": 0.0458417448272003, "grad_norm": 0.46005672216415405, "learning_rate": 0.00019998260665840675, "loss": 5.9104, "step": 10650 }, { "epoch": 0.046056964286482927, "grad_norm": 1.2751468420028687, "learning_rate": 0.00019998242593609412, "loss": 6.3372, "step": 10700 }, { "epoch": 0.046272183745765555, "grad_norm": 1.385887861251831, "learning_rate": 0.00019998224427982924, "loss": 6.5581, "step": 10750 }, { "epoch": 0.04648740320504819, "grad_norm": 0.8021640777587891, "learning_rate": 0.00019998206168961384, "loss": 6.0399, "step": 10800 }, { "epoch": 0.04670262266433082, "grad_norm": 2.2508010864257812, "learning_rate": 0.00019998187816544952, "loss": 6.4395, "step": 10850 }, { "epoch": 0.04691784212361345, "grad_norm": 1.5441851615905762, "learning_rate": 0.00019998169370733814, "loss": 6.138, "step": 10900 }, { "epoch": 0.04713306158289608, "grad_norm": 3.2657670974731445, "learning_rate": 0.00019998150831528126, "loss": 6.0005, "step": 10950 }, { "epoch": 0.04734828104217871, "grad_norm": 3.098504066467285, "learning_rate": 0.00019998132198928073, "loss": 6.027, "step": 11000 }, { "epoch": 0.04756350050146134, "grad_norm": 2.4909443855285645, "learning_rate": 0.00019998113472933824, "loss": 6.0344, "step": 11050 }, { "epoch": 0.04777871996074397, "grad_norm": 1.9509649276733398, "learning_rate": 0.00019998094653545553, "loss": 5.8786, "step": 11100 }, { "epoch": 0.0479939394200266, "grad_norm": 2.4075751304626465, "learning_rate": 0.0001999807574076344, "loss": 5.8857, "step": 11150 }, { "epoch": 0.04820915887930923, "grad_norm": 1.79831862449646, "learning_rate": 0.0001999805673458766, "loss": 6.384, "step": 11200 }, { "epoch": 0.04842437833859186, "grad_norm": 2.080692768096924, "learning_rate": 0.00019998037635018392, "loss": 5.7048, "step": 11250 }, { "epoch": 0.048639597797874495, "grad_norm": 0.9270420670509338, "learning_rate": 0.00019998018442055808, "loss": 5.8681, "step": 11300 }, { "epoch": 0.048854817257157124, "grad_norm": 2.5996997356414795, "learning_rate": 0.00019997999155700095, "loss": 6.3557, "step": 11350 }, { "epoch": 0.04907003671643975, "grad_norm": 2.470367908477783, "learning_rate": 0.00019997979775951427, "loss": 6.3286, "step": 11400 }, { "epoch": 0.04928525617572239, "grad_norm": 2.5481393337249756, "learning_rate": 0.0001999796030280999, "loss": 5.9273, "step": 11450 }, { "epoch": 0.049500475635005016, "grad_norm": 1.8851689100265503, "learning_rate": 0.00019997940736275966, "loss": 6.2376, "step": 11500 }, { "epoch": 0.049715695094287644, "grad_norm": 2.121269464492798, "learning_rate": 0.00019997921076349534, "loss": 5.9948, "step": 11550 }, { "epoch": 0.04993091455357027, "grad_norm": 1.7271016836166382, "learning_rate": 0.00019997901323030875, "loss": 5.77, "step": 11600 }, { "epoch": 0.05014613401285291, "grad_norm": 2.514357328414917, "learning_rate": 0.00019997881476320183, "loss": 6.5634, "step": 11650 }, { "epoch": 0.05036135347213554, "grad_norm": 2.115257740020752, "learning_rate": 0.00019997861536217638, "loss": 6.4403, "step": 11700 }, { "epoch": 0.050576572931418165, "grad_norm": 0.45553264021873474, "learning_rate": 0.00019997841502723422, "loss": 6.2061, "step": 11750 }, { "epoch": 0.0507917923907008, "grad_norm": 1.6110501289367676, "learning_rate": 0.0001999782137583773, "loss": 5.9793, "step": 11800 }, { "epoch": 0.05100701184998343, "grad_norm": 2.176588773727417, "learning_rate": 0.00019997801155560748, "loss": 6.1148, "step": 11850 }, { "epoch": 0.05122223130926606, "grad_norm": 2.086477279663086, "learning_rate": 0.00019997780841892658, "loss": 6.2692, "step": 11900 }, { "epoch": 0.051437450768548686, "grad_norm": 1.3375746011734009, "learning_rate": 0.00019997760434833662, "loss": 6.5075, "step": 11950 }, { "epoch": 0.05165267022783132, "grad_norm": 0.49182364344596863, "learning_rate": 0.0001999773993438394, "loss": 5.4787, "step": 12000 }, { "epoch": 0.05165267022783132, "eval_loss": 6.623365879058838, "eval_runtime": 49.574, "eval_samples_per_second": 12.91, "eval_steps_per_second": 6.455, "eval_tts_loss": 5.784761963718602, "step": 12000 }, { "epoch": 0.05186788968711395, "grad_norm": 2.018699884414673, "learning_rate": 0.00019997719340543687, "loss": 6.4145, "step": 12050 }, { "epoch": 0.05208310914639658, "grad_norm": 2.7786993980407715, "learning_rate": 0.000199976986533131, "loss": 6.5735, "step": 12100 }, { "epoch": 0.05229832860567921, "grad_norm": 2.132110595703125, "learning_rate": 0.00019997677872692364, "loss": 6.1921, "step": 12150 }, { "epoch": 0.05251354806496184, "grad_norm": 2.1844096183776855, "learning_rate": 0.00019997656998681676, "loss": 6.1033, "step": 12200 }, { "epoch": 0.05272876752424447, "grad_norm": 0.5526061654090881, "learning_rate": 0.00019997636031281234, "loss": 5.609, "step": 12250 }, { "epoch": 0.052943986983527105, "grad_norm": 1.4592640399932861, "learning_rate": 0.00019997614970491232, "loss": 5.2584, "step": 12300 }, { "epoch": 0.053159206442809734, "grad_norm": 1.3534646034240723, "learning_rate": 0.00019997593816311867, "loss": 5.7567, "step": 12350 }, { "epoch": 0.05337442590209236, "grad_norm": 1.6809126138687134, "learning_rate": 0.00019997572568743334, "loss": 6.1116, "step": 12400 }, { "epoch": 0.05358964536137499, "grad_norm": 1.6263248920440674, "learning_rate": 0.00019997551227785837, "loss": 6.3186, "step": 12450 }, { "epoch": 0.053804864820657626, "grad_norm": 1.8647878170013428, "learning_rate": 0.0001999752979343957, "loss": 5.5683, "step": 12500 }, { "epoch": 0.054020084279940254, "grad_norm": 1.4841105937957764, "learning_rate": 0.00019997508265704735, "loss": 6.0048, "step": 12550 }, { "epoch": 0.05423530373922288, "grad_norm": 1.013608694076538, "learning_rate": 0.00019997486644581533, "loss": 5.8209, "step": 12600 }, { "epoch": 0.05445052319850552, "grad_norm": 1.648074984550476, "learning_rate": 0.0001999746493007017, "loss": 5.9624, "step": 12650 }, { "epoch": 0.05466574265778815, "grad_norm": 0.8728966116905212, "learning_rate": 0.0001999744312217084, "loss": 6.4457, "step": 12700 }, { "epoch": 0.054880962117070775, "grad_norm": 1.9282983541488647, "learning_rate": 0.00019997421220883759, "loss": 6.1858, "step": 12750 }, { "epoch": 0.05509618157635341, "grad_norm": 3.152851104736328, "learning_rate": 0.00019997399226209117, "loss": 6.1478, "step": 12800 }, { "epoch": 0.05531140103563604, "grad_norm": 1.2537286281585693, "learning_rate": 0.0001999737713814713, "loss": 6.2248, "step": 12850 }, { "epoch": 0.05552662049491867, "grad_norm": 0.884758472442627, "learning_rate": 0.00019997354956698002, "loss": 6.7941, "step": 12900 }, { "epoch": 0.055741839954201296, "grad_norm": 1.5350348949432373, "learning_rate": 0.0001999733268186194, "loss": 5.9895, "step": 12950 }, { "epoch": 0.05595705941348393, "grad_norm": 2.1208536624908447, "learning_rate": 0.0001999731031363915, "loss": 6.3084, "step": 13000 }, { "epoch": 0.05617227887276656, "grad_norm": 2.4153099060058594, "learning_rate": 0.00019997287852029842, "loss": 5.8474, "step": 13050 }, { "epoch": 0.05638749833204919, "grad_norm": 1.8414849042892456, "learning_rate": 0.0001999726529703423, "loss": 6.055, "step": 13100 }, { "epoch": 0.05660271779133182, "grad_norm": 2.1909987926483154, "learning_rate": 0.00019997242648652516, "loss": 6.5164, "step": 13150 }, { "epoch": 0.05681793725061445, "grad_norm": 1.5044007301330566, "learning_rate": 0.0001999721990688492, "loss": 6.0312, "step": 13200 }, { "epoch": 0.05703315670989708, "grad_norm": 1.7146600484848022, "learning_rate": 0.0001999719707173165, "loss": 6.523, "step": 13250 }, { "epoch": 0.057248376169179716, "grad_norm": 1.8632339239120483, "learning_rate": 0.00019997174143192919, "loss": 5.7861, "step": 13300 }, { "epoch": 0.057463595628462344, "grad_norm": 1.0893120765686035, "learning_rate": 0.00019997151121268944, "loss": 6.3586, "step": 13350 }, { "epoch": 0.05767881508774497, "grad_norm": 0.5059545636177063, "learning_rate": 0.00019997128005959939, "loss": 5.9897, "step": 13400 }, { "epoch": 0.0578940345470276, "grad_norm": 1.8292659521102905, "learning_rate": 0.00019997104797266118, "loss": 6.1702, "step": 13450 }, { "epoch": 0.058109254006310236, "grad_norm": 2.2367465496063232, "learning_rate": 0.000199970814951877, "loss": 6.1767, "step": 13500 }, { "epoch": 0.058324473465592865, "grad_norm": 2.4136533737182617, "learning_rate": 0.000199970580997249, "loss": 6.1249, "step": 13550 }, { "epoch": 0.05853969292487549, "grad_norm": 1.612929105758667, "learning_rate": 0.00019997034610877944, "loss": 5.8287, "step": 13600 }, { "epoch": 0.05875491238415813, "grad_norm": 0.9251105785369873, "learning_rate": 0.0001999701102864704, "loss": 6.0461, "step": 13650 }, { "epoch": 0.05897013184344076, "grad_norm": 2.172238826751709, "learning_rate": 0.0001999698735303242, "loss": 5.8905, "step": 13700 }, { "epoch": 0.059185351302723385, "grad_norm": 1.4060155153274536, "learning_rate": 0.00019996963584034295, "loss": 5.6214, "step": 13750 }, { "epoch": 0.05940057076200602, "grad_norm": 2.0800819396972656, "learning_rate": 0.00019996939721652888, "loss": 6.5706, "step": 13800 }, { "epoch": 0.05961579022128865, "grad_norm": 1.402781367301941, "learning_rate": 0.00019996915765888433, "loss": 6.2123, "step": 13850 }, { "epoch": 0.05983100968057128, "grad_norm": 1.6385306119918823, "learning_rate": 0.00019996891716741143, "loss": 5.4061, "step": 13900 }, { "epoch": 0.060046229139853906, "grad_norm": 0.31861889362335205, "learning_rate": 0.00019996867574211243, "loss": 5.3454, "step": 13950 }, { "epoch": 0.06026144859913654, "grad_norm": 2.9108541011810303, "learning_rate": 0.00019996843338298965, "loss": 5.4115, "step": 14000 }, { "epoch": 0.06047666805841917, "grad_norm": 0.47495198249816895, "learning_rate": 0.0001999681900900453, "loss": 5.5743, "step": 14050 }, { "epoch": 0.0606918875177018, "grad_norm": 1.9878536462783813, "learning_rate": 0.00019996794586328166, "loss": 6.1532, "step": 14100 }, { "epoch": 0.06090710697698443, "grad_norm": 1.391926646232605, "learning_rate": 0.00019996770070270103, "loss": 5.9684, "step": 14150 }, { "epoch": 0.06112232643626706, "grad_norm": 2.491175413131714, "learning_rate": 0.00019996745460830568, "loss": 5.9218, "step": 14200 }, { "epoch": 0.06133754589554969, "grad_norm": 1.5352635383605957, "learning_rate": 0.00019996720758009792, "loss": 5.4476, "step": 14250 }, { "epoch": 0.061552765354832326, "grad_norm": 1.4185391664505005, "learning_rate": 0.00019996695961808008, "loss": 6.2033, "step": 14300 }, { "epoch": 0.061767984814114954, "grad_norm": 1.5892852544784546, "learning_rate": 0.00019996671072225445, "loss": 6.2819, "step": 14350 }, { "epoch": 0.06198320427339758, "grad_norm": 0.38056132197380066, "learning_rate": 0.00019996646089262332, "loss": 5.9927, "step": 14400 }, { "epoch": 0.06219842373268021, "grad_norm": 1.9075366258621216, "learning_rate": 0.00019996621012918908, "loss": 5.8884, "step": 14450 }, { "epoch": 0.062413643191962846, "grad_norm": 1.7801440954208374, "learning_rate": 0.00019996595843195408, "loss": 6.243, "step": 14500 }, { "epoch": 0.06262886265124548, "grad_norm": 1.9789460897445679, "learning_rate": 0.0001999657058009206, "loss": 5.2851, "step": 14550 }, { "epoch": 0.06284408211052811, "grad_norm": 1.8612463474273682, "learning_rate": 0.00019996545223609108, "loss": 5.7764, "step": 14600 }, { "epoch": 0.06305930156981074, "grad_norm": 2.149993658065796, "learning_rate": 0.00019996519773746784, "loss": 5.7601, "step": 14650 }, { "epoch": 0.06327452102909337, "grad_norm": 1.8025063276290894, "learning_rate": 0.00019996494230505328, "loss": 5.9883, "step": 14700 }, { "epoch": 0.063489740488376, "grad_norm": 0.9869495034217834, "learning_rate": 0.00019996468593884976, "loss": 5.6959, "step": 14750 }, { "epoch": 0.06370495994765862, "grad_norm": 3.1229183673858643, "learning_rate": 0.00019996442863885972, "loss": 6.1251, "step": 14800 }, { "epoch": 0.06392017940694125, "grad_norm": 1.4746047258377075, "learning_rate": 0.0001999641704050855, "loss": 6.0352, "step": 14850 }, { "epoch": 0.0641353988662239, "grad_norm": 1.3455562591552734, "learning_rate": 0.00019996391123752955, "loss": 5.3674, "step": 14900 }, { "epoch": 0.06435061832550652, "grad_norm": 1.8665653467178345, "learning_rate": 0.00019996365113619433, "loss": 6.0864, "step": 14950 }, { "epoch": 0.06456583778478915, "grad_norm": 0.7121979594230652, "learning_rate": 0.00019996339010108218, "loss": 5.5788, "step": 15000 }, { "epoch": 0.06456583778478915, "eval_loss": 6.473628044128418, "eval_runtime": 49.5914, "eval_samples_per_second": 12.905, "eval_steps_per_second": 6.453, "eval_tts_loss": 5.745539475633379, "step": 15000 }, { "epoch": 0.06478105724407178, "grad_norm": 0.6924996972084045, "learning_rate": 0.0001999631281321956, "loss": 6.1741, "step": 15050 }, { "epoch": 0.06499627670335441, "grad_norm": 1.3486802577972412, "learning_rate": 0.00019996286522953703, "loss": 5.3251, "step": 15100 }, { "epoch": 0.06521149616263704, "grad_norm": 2.111424446105957, "learning_rate": 0.00019996260139310893, "loss": 6.1477, "step": 15150 }, { "epoch": 0.06542671562191966, "grad_norm": 2.849172830581665, "learning_rate": 0.00019996233662291374, "loss": 5.3829, "step": 15200 }, { "epoch": 0.06564193508120231, "grad_norm": 1.7484654188156128, "learning_rate": 0.00019996207091895393, "loss": 6.2822, "step": 15250 }, { "epoch": 0.06585715454048494, "grad_norm": 2.255788564682007, "learning_rate": 0.00019996180428123205, "loss": 5.6831, "step": 15300 }, { "epoch": 0.06607237399976756, "grad_norm": 1.1666499376296997, "learning_rate": 0.00019996153670975047, "loss": 6.0367, "step": 15350 }, { "epoch": 0.06628759345905019, "grad_norm": 3.034298896789551, "learning_rate": 0.00019996126820451181, "loss": 5.9626, "step": 15400 }, { "epoch": 0.06650281291833282, "grad_norm": 2.2734391689300537, "learning_rate": 0.00019996099876551853, "loss": 5.7789, "step": 15450 }, { "epoch": 0.06671803237761545, "grad_norm": 1.6445180177688599, "learning_rate": 0.0001999607283927731, "loss": 6.1434, "step": 15500 }, { "epoch": 0.06693325183689809, "grad_norm": 1.7269871234893799, "learning_rate": 0.00019996045708627812, "loss": 5.2998, "step": 15550 }, { "epoch": 0.06714847129618072, "grad_norm": 0.25782594084739685, "learning_rate": 0.0001999601848460361, "loss": 5.6093, "step": 15600 }, { "epoch": 0.06736369075546335, "grad_norm": 0.5305419564247131, "learning_rate": 0.00019995991167204958, "loss": 5.4733, "step": 15650 }, { "epoch": 0.06757891021474598, "grad_norm": 1.7923744916915894, "learning_rate": 0.00019995963756432107, "loss": 5.8425, "step": 15700 }, { "epoch": 0.0677941296740286, "grad_norm": 1.5148489475250244, "learning_rate": 0.00019995936252285322, "loss": 5.7757, "step": 15750 }, { "epoch": 0.06800934913331123, "grad_norm": 1.1658241748809814, "learning_rate": 0.0001999590865476485, "loss": 6.5055, "step": 15800 }, { "epoch": 0.06822456859259386, "grad_norm": 0.537838339805603, "learning_rate": 0.00019995880963870956, "loss": 5.9404, "step": 15850 }, { "epoch": 0.0684397880518765, "grad_norm": 1.525205373764038, "learning_rate": 0.00019995853179603898, "loss": 6.2541, "step": 15900 }, { "epoch": 0.06865500751115913, "grad_norm": 2.2363901138305664, "learning_rate": 0.0001999582530196393, "loss": 5.9106, "step": 15950 }, { "epoch": 0.06887022697044176, "grad_norm": 3.7833199501037598, "learning_rate": 0.0001999579733095132, "loss": 5.9148, "step": 16000 }, { "epoch": 0.06908544642972439, "grad_norm": 1.4124741554260254, "learning_rate": 0.00019995769266566323, "loss": 6.6484, "step": 16050 }, { "epoch": 0.06930066588900702, "grad_norm": 2.293710708618164, "learning_rate": 0.00019995741108809204, "loss": 5.5159, "step": 16100 }, { "epoch": 0.06951588534828965, "grad_norm": 3.107032299041748, "learning_rate": 0.00019995712857680226, "loss": 6.0746, "step": 16150 }, { "epoch": 0.06973110480757228, "grad_norm": 2.14458966255188, "learning_rate": 0.00019995684513179654, "loss": 5.7623, "step": 16200 }, { "epoch": 0.06994632426685492, "grad_norm": 2.7619552612304688, "learning_rate": 0.00019995656075307748, "loss": 5.7459, "step": 16250 }, { "epoch": 0.07016154372613755, "grad_norm": 2.292412519454956, "learning_rate": 0.00019995627544064782, "loss": 5.5526, "step": 16300 }, { "epoch": 0.07037676318542017, "grad_norm": 1.2620292901992798, "learning_rate": 0.00019995598919451014, "loss": 5.5151, "step": 16350 }, { "epoch": 0.0705919826447028, "grad_norm": 1.5071910619735718, "learning_rate": 0.00019995570201466713, "loss": 5.9367, "step": 16400 }, { "epoch": 0.07080720210398543, "grad_norm": 1.4440970420837402, "learning_rate": 0.00019995541390112153, "loss": 5.4564, "step": 16450 }, { "epoch": 0.07102242156326806, "grad_norm": 1.0968797206878662, "learning_rate": 0.00019995512485387597, "loss": 5.517, "step": 16500 }, { "epoch": 0.0712376410225507, "grad_norm": 2.6469056606292725, "learning_rate": 0.0001999548348729332, "loss": 6.4836, "step": 16550 }, { "epoch": 0.07145286048183333, "grad_norm": 1.7991894483566284, "learning_rate": 0.0001999545439582959, "loss": 5.534, "step": 16600 }, { "epoch": 0.07166807994111596, "grad_norm": 1.4423532485961914, "learning_rate": 0.00019995425210996676, "loss": 6.1768, "step": 16650 }, { "epoch": 0.07188329940039859, "grad_norm": 1.2330061197280884, "learning_rate": 0.00019995395932794857, "loss": 5.6375, "step": 16700 }, { "epoch": 0.07209851885968122, "grad_norm": 1.4921482801437378, "learning_rate": 0.000199953665612244, "loss": 6.3679, "step": 16750 }, { "epoch": 0.07231373831896384, "grad_norm": 2.5475311279296875, "learning_rate": 0.0001999533709628558, "loss": 6.3564, "step": 16800 }, { "epoch": 0.07252895777824647, "grad_norm": 2.933581829071045, "learning_rate": 0.00019995307537978683, "loss": 5.7804, "step": 16850 }, { "epoch": 0.07274417723752911, "grad_norm": 1.4602488279342651, "learning_rate": 0.00019995277886303972, "loss": 5.8952, "step": 16900 }, { "epoch": 0.07295939669681174, "grad_norm": 1.3859577178955078, "learning_rate": 0.00019995248141261728, "loss": 6.1608, "step": 16950 }, { "epoch": 0.07317461615609437, "grad_norm": 0.8692163228988647, "learning_rate": 0.0001999521830285223, "loss": 5.6304, "step": 17000 }, { "epoch": 0.073389835615377, "grad_norm": 2.576009750366211, "learning_rate": 0.00019995188371075754, "loss": 5.9063, "step": 17050 }, { "epoch": 0.07360505507465963, "grad_norm": 1.8625093698501587, "learning_rate": 0.00019995158345932588, "loss": 6.2452, "step": 17100 }, { "epoch": 0.07382027453394226, "grad_norm": 1.5731500387191772, "learning_rate": 0.00019995128227423003, "loss": 6.0671, "step": 17150 }, { "epoch": 0.07403549399322489, "grad_norm": 0.4131663143634796, "learning_rate": 0.00019995098015547283, "loss": 5.8824, "step": 17200 }, { "epoch": 0.07425071345250753, "grad_norm": 0.8354901075363159, "learning_rate": 0.00019995067710305712, "loss": 5.8642, "step": 17250 }, { "epoch": 0.07446593291179016, "grad_norm": 1.5117969512939453, "learning_rate": 0.0001999503731169857, "loss": 5.6949, "step": 17300 }, { "epoch": 0.07468115237107278, "grad_norm": 1.678210735321045, "learning_rate": 0.00019995006819726146, "loss": 5.8878, "step": 17350 }, { "epoch": 0.07489637183035541, "grad_norm": 1.3499431610107422, "learning_rate": 0.0001999497623438872, "loss": 5.6163, "step": 17400 }, { "epoch": 0.07511159128963804, "grad_norm": 0.5428993701934814, "learning_rate": 0.0001999494555568658, "loss": 5.6279, "step": 17450 }, { "epoch": 0.07532681074892067, "grad_norm": 1.8066754341125488, "learning_rate": 0.00019994914783620012, "loss": 6.257, "step": 17500 }, { "epoch": 0.07554203020820331, "grad_norm": 1.0412993431091309, "learning_rate": 0.00019994883918189303, "loss": 5.4711, "step": 17550 }, { "epoch": 0.07575724966748594, "grad_norm": 2.0055477619171143, "learning_rate": 0.00019994852959394742, "loss": 6.132, "step": 17600 }, { "epoch": 0.07597246912676857, "grad_norm": 1.5970449447631836, "learning_rate": 0.00019994821907236618, "loss": 6.1875, "step": 17650 }, { "epoch": 0.0761876885860512, "grad_norm": 1.3863506317138672, "learning_rate": 0.00019994790761715224, "loss": 5.4176, "step": 17700 }, { "epoch": 0.07640290804533383, "grad_norm": 1.4371259212493896, "learning_rate": 0.00019994759522830844, "loss": 5.9098, "step": 17750 }, { "epoch": 0.07661812750461645, "grad_norm": 1.0196858644485474, "learning_rate": 0.00019994728190583775, "loss": 5.9273, "step": 17800 }, { "epoch": 0.07683334696389908, "grad_norm": 1.851461410522461, "learning_rate": 0.0001999469676497431, "loss": 5.5738, "step": 17850 }, { "epoch": 0.07704856642318172, "grad_norm": 1.1883164644241333, "learning_rate": 0.00019994665246002738, "loss": 5.7435, "step": 17900 }, { "epoch": 0.07726378588246435, "grad_norm": 3.046671152114868, "learning_rate": 0.00019994633633669363, "loss": 6.0122, "step": 17950 }, { "epoch": 0.07747900534174698, "grad_norm": 1.2815881967544556, "learning_rate": 0.0001999460192797447, "loss": 5.4561, "step": 18000 }, { "epoch": 0.07747900534174698, "eval_loss": 6.404990196228027, "eval_runtime": 49.5251, "eval_samples_per_second": 12.923, "eval_steps_per_second": 6.461, "eval_tts_loss": 5.82405799120409, "step": 18000 }, { "epoch": 0.07769422480102961, "grad_norm": 1.6492704153060913, "learning_rate": 0.00019994570128918358, "loss": 5.8528, "step": 18050 }, { "epoch": 0.07790944426031224, "grad_norm": 3.2377874851226807, "learning_rate": 0.00019994538236501325, "loss": 6.0166, "step": 18100 }, { "epoch": 0.07812466371959487, "grad_norm": 1.5120192766189575, "learning_rate": 0.00019994506250723668, "loss": 6.3162, "step": 18150 }, { "epoch": 0.0783398831788775, "grad_norm": 1.6424208879470825, "learning_rate": 0.0001999447417158569, "loss": 5.6143, "step": 18200 }, { "epoch": 0.07855510263816014, "grad_norm": 2.1206583976745605, "learning_rate": 0.00019994441999087684, "loss": 5.4974, "step": 18250 }, { "epoch": 0.07877032209744277, "grad_norm": 1.7122966051101685, "learning_rate": 0.00019994409733229955, "loss": 5.8445, "step": 18300 }, { "epoch": 0.0789855415567254, "grad_norm": 2.3482401371002197, "learning_rate": 0.00019994377374012805, "loss": 5.8277, "step": 18350 }, { "epoch": 0.07920076101600802, "grad_norm": 3.078629732131958, "learning_rate": 0.00019994344921436534, "loss": 5.319, "step": 18400 }, { "epoch": 0.07941598047529065, "grad_norm": 0.6193755269050598, "learning_rate": 0.00019994312375501446, "loss": 6.0665, "step": 18450 }, { "epoch": 0.07963119993457328, "grad_norm": 1.496768593788147, "learning_rate": 0.00019994279736207843, "loss": 5.5588, "step": 18500 }, { "epoch": 0.07984641939385591, "grad_norm": 1.4955095052719116, "learning_rate": 0.00019994247003556032, "loss": 5.8826, "step": 18550 }, { "epoch": 0.08006163885313855, "grad_norm": 0.6703450083732605, "learning_rate": 0.00019994214177546319, "loss": 5.2912, "step": 18600 }, { "epoch": 0.08027685831242118, "grad_norm": 1.2676513195037842, "learning_rate": 0.0001999418125817901, "loss": 6.1078, "step": 18650 }, { "epoch": 0.08049207777170381, "grad_norm": 1.202744483947754, "learning_rate": 0.00019994148245454412, "loss": 5.4566, "step": 18700 }, { "epoch": 0.08070729723098644, "grad_norm": 1.044953465461731, "learning_rate": 0.00019994115139372834, "loss": 5.3546, "step": 18750 }, { "epoch": 0.08092251669026906, "grad_norm": 1.0811514854431152, "learning_rate": 0.00019994081939934586, "loss": 5.9288, "step": 18800 }, { "epoch": 0.08113773614955169, "grad_norm": 2.0052831172943115, "learning_rate": 0.00019994048647139977, "loss": 5.8342, "step": 18850 }, { "epoch": 0.08135295560883433, "grad_norm": 1.2067780494689941, "learning_rate": 0.0001999401526098932, "loss": 5.6739, "step": 18900 }, { "epoch": 0.08156817506811696, "grad_norm": 0.4020227789878845, "learning_rate": 0.0001999398178148292, "loss": 5.5966, "step": 18950 }, { "epoch": 0.08178339452739959, "grad_norm": 2.052269697189331, "learning_rate": 0.00019993948208621098, "loss": 5.9416, "step": 19000 }, { "epoch": 0.08199861398668222, "grad_norm": 1.8629567623138428, "learning_rate": 0.00019993914542404168, "loss": 5.8055, "step": 19050 }, { "epoch": 0.08221383344596485, "grad_norm": 0.7409083843231201, "learning_rate": 0.00019993880782832432, "loss": 5.9739, "step": 19100 }, { "epoch": 0.08242905290524748, "grad_norm": 1.4553385972976685, "learning_rate": 0.0001999384692990622, "loss": 5.7744, "step": 19150 }, { "epoch": 0.0826442723645301, "grad_norm": 1.211584448814392, "learning_rate": 0.00019993812983625843, "loss": 5.9604, "step": 19200 }, { "epoch": 0.08285949182381275, "grad_norm": 1.2810535430908203, "learning_rate": 0.00019993778943991618, "loss": 5.7684, "step": 19250 }, { "epoch": 0.08307471128309538, "grad_norm": 1.5037965774536133, "learning_rate": 0.00019993744811003863, "loss": 5.9016, "step": 19300 }, { "epoch": 0.083289930742378, "grad_norm": 2.1033639907836914, "learning_rate": 0.00019993710584662893, "loss": 5.7813, "step": 19350 }, { "epoch": 0.08350515020166063, "grad_norm": 2.1486220359802246, "learning_rate": 0.0001999367626496903, "loss": 6.1157, "step": 19400 }, { "epoch": 0.08372036966094326, "grad_norm": 1.3567312955856323, "learning_rate": 0.000199936418519226, "loss": 5.9387, "step": 19450 }, { "epoch": 0.08393558912022589, "grad_norm": 1.303710699081421, "learning_rate": 0.0001999360734552392, "loss": 5.4414, "step": 19500 }, { "epoch": 0.08415080857950852, "grad_norm": 1.0848397016525269, "learning_rate": 0.00019993572745773315, "loss": 6.2168, "step": 19550 }, { "epoch": 0.08436602803879116, "grad_norm": 2.0063588619232178, "learning_rate": 0.00019993538052671101, "loss": 6.0967, "step": 19600 }, { "epoch": 0.08458124749807379, "grad_norm": 1.0752456188201904, "learning_rate": 0.0001999350326621761, "loss": 5.7904, "step": 19650 }, { "epoch": 0.08479646695735642, "grad_norm": 1.660332441329956, "learning_rate": 0.0001999346838641316, "loss": 5.2532, "step": 19700 }, { "epoch": 0.08501168641663905, "grad_norm": 1.4136866331100464, "learning_rate": 0.00019993433413258082, "loss": 5.8131, "step": 19750 }, { "epoch": 0.08522690587592167, "grad_norm": 0.26315972208976746, "learning_rate": 0.00019993398346752702, "loss": 5.7124, "step": 19800 }, { "epoch": 0.0854421253352043, "grad_norm": 1.3278565406799316, "learning_rate": 0.00019993363186897347, "loss": 6.0263, "step": 19850 }, { "epoch": 0.08565734479448694, "grad_norm": 1.3003625869750977, "learning_rate": 0.00019993327933692345, "loss": 5.7515, "step": 19900 }, { "epoch": 0.08587256425376957, "grad_norm": 2.5729832649230957, "learning_rate": 0.00019993292587138024, "loss": 5.8048, "step": 19950 }, { "epoch": 0.0860877837130522, "grad_norm": 1.6819454431533813, "learning_rate": 0.0001999325714723472, "loss": 5.9692, "step": 20000 }, { "epoch": 0.08630300317233483, "grad_norm": 1.5779386758804321, "learning_rate": 0.00019993221613982755, "loss": 5.9163, "step": 20050 }, { "epoch": 0.08651822263161746, "grad_norm": 2.1650054454803467, "learning_rate": 0.00019993185987382468, "loss": 5.6334, "step": 20100 }, { "epoch": 0.08673344209090009, "grad_norm": 1.8448052406311035, "learning_rate": 0.00019993150267434188, "loss": 5.2351, "step": 20150 }, { "epoch": 0.08694866155018272, "grad_norm": 1.7112996578216553, "learning_rate": 0.00019993114454138253, "loss": 6.1807, "step": 20200 }, { "epoch": 0.08716388100946536, "grad_norm": 1.3212093114852905, "learning_rate": 0.0001999307854749499, "loss": 5.6485, "step": 20250 }, { "epoch": 0.08737910046874799, "grad_norm": 1.3698363304138184, "learning_rate": 0.00019993042547504747, "loss": 5.9835, "step": 20300 }, { "epoch": 0.08759431992803061, "grad_norm": 1.5495718717575073, "learning_rate": 0.00019993006454167847, "loss": 5.7647, "step": 20350 }, { "epoch": 0.08780953938731324, "grad_norm": 0.9871811270713806, "learning_rate": 0.00019992970267484632, "loss": 5.7333, "step": 20400 }, { "epoch": 0.08802475884659587, "grad_norm": 3.238474130630493, "learning_rate": 0.00019992933987455442, "loss": 5.8684, "step": 20450 }, { "epoch": 0.0882399783058785, "grad_norm": 1.609344720840454, "learning_rate": 0.00019992897614080614, "loss": 5.8609, "step": 20500 }, { "epoch": 0.08845519776516113, "grad_norm": 1.6140156984329224, "learning_rate": 0.0001999286114736049, "loss": 5.9072, "step": 20550 }, { "epoch": 0.08867041722444377, "grad_norm": 1.660849690437317, "learning_rate": 0.00019992824587295406, "loss": 5.7746, "step": 20600 }, { "epoch": 0.0888856366837264, "grad_norm": 1.2376216650009155, "learning_rate": 0.00019992787933885707, "loss": 5.789, "step": 20650 }, { "epoch": 0.08910085614300903, "grad_norm": 0.14461947977542877, "learning_rate": 0.00019992751187131739, "loss": 5.4646, "step": 20700 }, { "epoch": 0.08931607560229166, "grad_norm": 2.796215772628784, "learning_rate": 0.00019992714347033834, "loss": 5.8981, "step": 20750 }, { "epoch": 0.08953129506157428, "grad_norm": 2.2618072032928467, "learning_rate": 0.00019992677413592348, "loss": 5.7584, "step": 20800 }, { "epoch": 0.08974651452085691, "grad_norm": 4.238938808441162, "learning_rate": 0.0001999264038680762, "loss": 5.4638, "step": 20850 }, { "epoch": 0.08996173398013955, "grad_norm": 2.422128915786743, "learning_rate": 0.0001999260326668, "loss": 5.8727, "step": 20900 }, { "epoch": 0.09017695343942218, "grad_norm": 2.265758752822876, "learning_rate": 0.00019992566053209828, "loss": 5.4803, "step": 20950 }, { "epoch": 0.09039217289870481, "grad_norm": 1.1186612844467163, "learning_rate": 0.00019992528746397455, "loss": 5.962, "step": 21000 }, { "epoch": 0.09039217289870481, "eval_loss": 6.2909040451049805, "eval_runtime": 49.7885, "eval_samples_per_second": 12.854, "eval_steps_per_second": 6.427, "eval_tts_loss": 5.815406650647389, "step": 21000 }, { "epoch": 0.09060739235798744, "grad_norm": 1.6426743268966675, "learning_rate": 0.00019992491346243233, "loss": 5.4301, "step": 21050 }, { "epoch": 0.09082261181727007, "grad_norm": 0.8008338809013367, "learning_rate": 0.00019992453852747506, "loss": 6.08, "step": 21100 }, { "epoch": 0.0910378312765527, "grad_norm": 1.4610543251037598, "learning_rate": 0.00019992416265910627, "loss": 6.0996, "step": 21150 }, { "epoch": 0.09125305073583533, "grad_norm": 1.2644808292388916, "learning_rate": 0.00019992378585732948, "loss": 5.8896, "step": 21200 }, { "epoch": 0.09146827019511797, "grad_norm": 2.0503859519958496, "learning_rate": 0.00019992340812214815, "loss": 5.9489, "step": 21250 }, { "epoch": 0.0916834896544006, "grad_norm": 1.3490443229675293, "learning_rate": 0.00019992302945356588, "loss": 5.6218, "step": 21300 }, { "epoch": 0.09189870911368322, "grad_norm": 1.1548471450805664, "learning_rate": 0.00019992264985158623, "loss": 5.1465, "step": 21350 }, { "epoch": 0.09211392857296585, "grad_norm": 1.0206573009490967, "learning_rate": 0.00019992226931621264, "loss": 5.7689, "step": 21400 }, { "epoch": 0.09232914803224848, "grad_norm": 0.36905255913734436, "learning_rate": 0.00019992188784744873, "loss": 5.4248, "step": 21450 }, { "epoch": 0.09254436749153111, "grad_norm": 1.2325526475906372, "learning_rate": 0.00019992150544529807, "loss": 5.6658, "step": 21500 }, { "epoch": 0.09275958695081374, "grad_norm": 1.8993648290634155, "learning_rate": 0.00019992112210976418, "loss": 5.323, "step": 21550 }, { "epoch": 0.09297480641009638, "grad_norm": 0.8752487301826477, "learning_rate": 0.00019992073784085075, "loss": 5.7318, "step": 21600 }, { "epoch": 0.09319002586937901, "grad_norm": 1.5711878538131714, "learning_rate": 0.00019992035263856123, "loss": 5.8729, "step": 21650 }, { "epoch": 0.09340524532866164, "grad_norm": 0.3212529718875885, "learning_rate": 0.00019991996650289931, "loss": 5.9012, "step": 21700 }, { "epoch": 0.09362046478794427, "grad_norm": 0.9851751327514648, "learning_rate": 0.0001999195794338686, "loss": 5.6023, "step": 21750 }, { "epoch": 0.0938356842472269, "grad_norm": 1.3344018459320068, "learning_rate": 0.00019991919143147265, "loss": 6.3393, "step": 21800 }, { "epoch": 0.09405090370650952, "grad_norm": 1.7369003295898438, "learning_rate": 0.00019991880249571516, "loss": 5.6145, "step": 21850 }, { "epoch": 0.09426612316579216, "grad_norm": 1.8828001022338867, "learning_rate": 0.00019991841262659968, "loss": 5.6144, "step": 21900 }, { "epoch": 0.0944813426250748, "grad_norm": 2.2298882007598877, "learning_rate": 0.00019991802182412993, "loss": 5.428, "step": 21950 }, { "epoch": 0.09469656208435742, "grad_norm": 2.3802340030670166, "learning_rate": 0.00019991763008830952, "loss": 5.7847, "step": 22000 }, { "epoch": 0.09491178154364005, "grad_norm": 1.7031786441802979, "learning_rate": 0.00019991723741914213, "loss": 6.1874, "step": 22050 }, { "epoch": 0.09512700100292268, "grad_norm": 2.1063942909240723, "learning_rate": 0.0001999168438166314, "loss": 5.8033, "step": 22100 }, { "epoch": 0.09534222046220531, "grad_norm": 1.99550199508667, "learning_rate": 0.00019991644928078102, "loss": 5.7063, "step": 22150 }, { "epoch": 0.09555743992148794, "grad_norm": 1.9723435640335083, "learning_rate": 0.00019991605381159468, "loss": 6.0364, "step": 22200 }, { "epoch": 0.09577265938077058, "grad_norm": 1.6845579147338867, "learning_rate": 0.0001999156574090761, "loss": 5.8157, "step": 22250 }, { "epoch": 0.0959878788400532, "grad_norm": 1.8619921207427979, "learning_rate": 0.00019991526007322893, "loss": 5.8394, "step": 22300 }, { "epoch": 0.09620309829933583, "grad_norm": 1.0366944074630737, "learning_rate": 0.0001999148618040569, "loss": 5.9506, "step": 22350 }, { "epoch": 0.09641831775861846, "grad_norm": 1.169640064239502, "learning_rate": 0.00019991446260156376, "loss": 5.9154, "step": 22400 }, { "epoch": 0.09663353721790109, "grad_norm": 0.5184738636016846, "learning_rate": 0.0001999140624657532, "loss": 5.7072, "step": 22450 }, { "epoch": 0.09684875667718372, "grad_norm": 1.3242932558059692, "learning_rate": 0.00019991366139662897, "loss": 5.8444, "step": 22500 }, { "epoch": 0.09706397613646635, "grad_norm": 0.7979602813720703, "learning_rate": 0.00019991325939419484, "loss": 5.4438, "step": 22550 }, { "epoch": 0.09727919559574899, "grad_norm": 1.5615895986557007, "learning_rate": 0.0001999128564584545, "loss": 6.14, "step": 22600 }, { "epoch": 0.09749441505503162, "grad_norm": 1.672521710395813, "learning_rate": 0.0001999124525894118, "loss": 5.1855, "step": 22650 }, { "epoch": 0.09770963451431425, "grad_norm": 1.0949991941452026, "learning_rate": 0.00019991204778707048, "loss": 5.7224, "step": 22700 }, { "epoch": 0.09792485397359688, "grad_norm": 0.8543679118156433, "learning_rate": 0.00019991164205143428, "loss": 5.9455, "step": 22750 }, { "epoch": 0.0981400734328795, "grad_norm": 1.1134306192398071, "learning_rate": 0.00019991123538250704, "loss": 5.2943, "step": 22800 }, { "epoch": 0.09835529289216213, "grad_norm": 2.1342930793762207, "learning_rate": 0.00019991082778029253, "loss": 5.7644, "step": 22850 }, { "epoch": 0.09857051235144478, "grad_norm": 2.0855610370635986, "learning_rate": 0.00019991041924479462, "loss": 6.0002, "step": 22900 }, { "epoch": 0.0987857318107274, "grad_norm": 0.5152562260627747, "learning_rate": 0.00019991000977601702, "loss": 5.8704, "step": 22950 }, { "epoch": 0.09900095127001003, "grad_norm": 0.6348848938941956, "learning_rate": 0.00019990959937396361, "loss": 5.7402, "step": 23000 }, { "epoch": 0.09921617072929266, "grad_norm": 1.433713674545288, "learning_rate": 0.00019990918803863825, "loss": 6.0831, "step": 23050 }, { "epoch": 0.09943139018857529, "grad_norm": 1.6799544095993042, "learning_rate": 0.00019990877577004474, "loss": 5.5117, "step": 23100 }, { "epoch": 0.09964660964785792, "grad_norm": 2.434643268585205, "learning_rate": 0.00019990836256818698, "loss": 6.3327, "step": 23150 }, { "epoch": 0.09986182910714055, "grad_norm": 0.9821347594261169, "learning_rate": 0.00019990794843306874, "loss": 5.8924, "step": 23200 }, { "epoch": 0.10007704856642319, "grad_norm": 1.0059661865234375, "learning_rate": 0.000199907533364694, "loss": 5.471, "step": 23250 }, { "epoch": 0.10029226802570582, "grad_norm": 1.9866455793380737, "learning_rate": 0.00019990711736306654, "loss": 5.6078, "step": 23300 }, { "epoch": 0.10050748748498844, "grad_norm": 0.15037888288497925, "learning_rate": 0.0001999067004281903, "loss": 5.4157, "step": 23350 }, { "epoch": 0.10072270694427107, "grad_norm": 1.6282198429107666, "learning_rate": 0.00019990628256006917, "loss": 5.8707, "step": 23400 }, { "epoch": 0.1009379264035537, "grad_norm": 3.3550591468811035, "learning_rate": 0.00019990586375870706, "loss": 6.1169, "step": 23450 }, { "epoch": 0.10115314586283633, "grad_norm": 1.7812178134918213, "learning_rate": 0.00019990544402410788, "loss": 5.8557, "step": 23500 }, { "epoch": 0.10136836532211896, "grad_norm": 1.471010446548462, "learning_rate": 0.0001999050233562755, "loss": 5.9594, "step": 23550 }, { "epoch": 0.1015835847814016, "grad_norm": 1.7927289009094238, "learning_rate": 0.00019990460175521392, "loss": 5.8421, "step": 23600 }, { "epoch": 0.10179880424068423, "grad_norm": 1.5211669206619263, "learning_rate": 0.00019990417922092704, "loss": 5.5385, "step": 23650 }, { "epoch": 0.10201402369996686, "grad_norm": 2.1302831172943115, "learning_rate": 0.0001999037557534188, "loss": 5.1209, "step": 23700 }, { "epoch": 0.10222924315924949, "grad_norm": 1.2189171314239502, "learning_rate": 0.0001999033313526932, "loss": 5.8434, "step": 23750 }, { "epoch": 0.10244446261853211, "grad_norm": 2.371162176132202, "learning_rate": 0.00019990290601875415, "loss": 6.0877, "step": 23800 }, { "epoch": 0.10265968207781474, "grad_norm": 1.2115380764007568, "learning_rate": 0.00019990247975160564, "loss": 5.779, "step": 23850 }, { "epoch": 0.10287490153709737, "grad_norm": 2.0238332748413086, "learning_rate": 0.00019990205255125166, "loss": 5.3852, "step": 23900 }, { "epoch": 0.10309012099638001, "grad_norm": 1.6233253479003906, "learning_rate": 0.00019990162441769626, "loss": 5.7042, "step": 23950 }, { "epoch": 0.10330534045566264, "grad_norm": 1.0732369422912598, "learning_rate": 0.00019990119535094332, "loss": 5.8434, "step": 24000 }, { "epoch": 0.10330534045566264, "eval_loss": 6.24224853515625, "eval_runtime": 49.5309, "eval_samples_per_second": 12.921, "eval_steps_per_second": 6.461, "eval_tts_loss": 5.905483429914947, "step": 24000 }, { "epoch": 0.10352055991494527, "grad_norm": 0.6568212509155273, "learning_rate": 0.00019990076535099691, "loss": 5.3753, "step": 24050 }, { "epoch": 0.1037357793742279, "grad_norm": 1.2204896211624146, "learning_rate": 0.00019990033441786106, "loss": 5.8193, "step": 24100 }, { "epoch": 0.10395099883351053, "grad_norm": 1.3121346235275269, "learning_rate": 0.00019989990255153976, "loss": 5.6626, "step": 24150 }, { "epoch": 0.10416621829279316, "grad_norm": 2.8243863582611084, "learning_rate": 0.0001998994697520371, "loss": 6.0255, "step": 24200 }, { "epoch": 0.1043814377520758, "grad_norm": 2.167590856552124, "learning_rate": 0.00019989903601935706, "loss": 5.4189, "step": 24250 }, { "epoch": 0.10459665721135843, "grad_norm": 1.3141735792160034, "learning_rate": 0.0001998986013535037, "loss": 5.8852, "step": 24300 }, { "epoch": 0.10481187667064105, "grad_norm": 2.2350897789001465, "learning_rate": 0.00019989816575448113, "loss": 5.6946, "step": 24350 }, { "epoch": 0.10502709612992368, "grad_norm": 1.437633991241455, "learning_rate": 0.0001998977292222934, "loss": 6.2396, "step": 24400 }, { "epoch": 0.10524231558920631, "grad_norm": 1.286749243736267, "learning_rate": 0.00019989729175694455, "loss": 5.6797, "step": 24450 }, { "epoch": 0.10545753504848894, "grad_norm": 1.4086824655532837, "learning_rate": 0.00019989685335843868, "loss": 5.4428, "step": 24500 }, { "epoch": 0.10567275450777157, "grad_norm": 1.3007848262786865, "learning_rate": 0.00019989641402677993, "loss": 5.423, "step": 24550 }, { "epoch": 0.10588797396705421, "grad_norm": 1.54371178150177, "learning_rate": 0.00019989597376197232, "loss": 5.4932, "step": 24600 }, { "epoch": 0.10610319342633684, "grad_norm": 1.058266520500183, "learning_rate": 0.00019989553256402007, "loss": 5.6363, "step": 24650 }, { "epoch": 0.10631841288561947, "grad_norm": 2.476527214050293, "learning_rate": 0.00019989509043292722, "loss": 6.0965, "step": 24700 }, { "epoch": 0.1065336323449021, "grad_norm": 2.0618507862091064, "learning_rate": 0.00019989464736869792, "loss": 5.3415, "step": 24750 }, { "epoch": 0.10674885180418472, "grad_norm": 0.8434683084487915, "learning_rate": 0.00019989420337133634, "loss": 5.079, "step": 24800 }, { "epoch": 0.10696407126346735, "grad_norm": 1.7271143198013306, "learning_rate": 0.00019989375844084654, "loss": 5.6947, "step": 24850 }, { "epoch": 0.10717929072274998, "grad_norm": 1.7582595348358154, "learning_rate": 0.0001998933125772328, "loss": 5.903, "step": 24900 }, { "epoch": 0.10739451018203262, "grad_norm": 1.9345180988311768, "learning_rate": 0.0001998928657804992, "loss": 5.3986, "step": 24950 }, { "epoch": 0.10760972964131525, "grad_norm": 1.560867428779602, "learning_rate": 0.00019989241805064994, "loss": 5.9008, "step": 25000 }, { "epoch": 0.10782494910059788, "grad_norm": 1.8246324062347412, "learning_rate": 0.00019989196938768918, "loss": 5.3205, "step": 25050 }, { "epoch": 0.10804016855988051, "grad_norm": 1.2432165145874023, "learning_rate": 0.00019989151979162112, "loss": 6.0183, "step": 25100 }, { "epoch": 0.10825538801916314, "grad_norm": 2.5812430381774902, "learning_rate": 0.00019989106926245, "loss": 5.7058, "step": 25150 }, { "epoch": 0.10847060747844577, "grad_norm": 1.5686339139938354, "learning_rate": 0.00019989061780018, "loss": 5.3641, "step": 25200 }, { "epoch": 0.10868582693772841, "grad_norm": 0.9212952852249146, "learning_rate": 0.00019989016540481532, "loss": 5.3621, "step": 25250 }, { "epoch": 0.10890104639701104, "grad_norm": 1.4971938133239746, "learning_rate": 0.0001998897120763602, "loss": 5.9256, "step": 25300 }, { "epoch": 0.10911626585629366, "grad_norm": 1.2954185009002686, "learning_rate": 0.00019988925781481886, "loss": 5.5944, "step": 25350 }, { "epoch": 0.1093314853155763, "grad_norm": 2.062845230102539, "learning_rate": 0.00019988880262019557, "loss": 5.5364, "step": 25400 }, { "epoch": 0.10954670477485892, "grad_norm": 1.2483621835708618, "learning_rate": 0.00019988834649249456, "loss": 5.9148, "step": 25450 }, { "epoch": 0.10976192423414155, "grad_norm": 2.104804039001465, "learning_rate": 0.00019988788943172013, "loss": 5.4299, "step": 25500 }, { "epoch": 0.10997714369342418, "grad_norm": 0.33023127913475037, "learning_rate": 0.00019988743143787652, "loss": 5.2553, "step": 25550 }, { "epoch": 0.11019236315270682, "grad_norm": 0.6736541986465454, "learning_rate": 0.00019988697251096797, "loss": 5.1407, "step": 25600 }, { "epoch": 0.11040758261198945, "grad_norm": 1.6496710777282715, "learning_rate": 0.00019988651265099883, "loss": 5.794, "step": 25650 }, { "epoch": 0.11062280207127208, "grad_norm": 0.8787000775337219, "learning_rate": 0.00019988605185797335, "loss": 5.5898, "step": 25700 }, { "epoch": 0.1108380215305547, "grad_norm": 1.4060803651809692, "learning_rate": 0.0001998855901318959, "loss": 5.7085, "step": 25750 }, { "epoch": 0.11105324098983733, "grad_norm": 1.605080246925354, "learning_rate": 0.00019988512747277072, "loss": 5.7787, "step": 25800 }, { "epoch": 0.11126846044911996, "grad_norm": 1.804853081703186, "learning_rate": 0.00019988466388060213, "loss": 4.9681, "step": 25850 }, { "epoch": 0.11148367990840259, "grad_norm": 1.311269760131836, "learning_rate": 0.00019988419935539455, "loss": 5.781, "step": 25900 }, { "epoch": 0.11169889936768523, "grad_norm": 1.0356370210647583, "learning_rate": 0.00019988373389715223, "loss": 5.5331, "step": 25950 }, { "epoch": 0.11191411882696786, "grad_norm": 1.8825137615203857, "learning_rate": 0.00019988326750587955, "loss": 5.7136, "step": 26000 }, { "epoch": 0.11212933828625049, "grad_norm": 1.8037456274032593, "learning_rate": 0.00019988280018158085, "loss": 5.4935, "step": 26050 }, { "epoch": 0.11234455774553312, "grad_norm": 0.7363218665122986, "learning_rate": 0.00019988233192426053, "loss": 5.5364, "step": 26100 }, { "epoch": 0.11255977720481575, "grad_norm": 0.7933793663978577, "learning_rate": 0.00019988186273392296, "loss": 5.5593, "step": 26150 }, { "epoch": 0.11277499666409838, "grad_norm": 1.783206820487976, "learning_rate": 0.00019988139261057248, "loss": 5.5515, "step": 26200 }, { "epoch": 0.11299021612338102, "grad_norm": 1.3489131927490234, "learning_rate": 0.0001998809215542135, "loss": 5.3354, "step": 26250 }, { "epoch": 0.11320543558266365, "grad_norm": 1.143532156944275, "learning_rate": 0.00019988044956485042, "loss": 5.6071, "step": 26300 }, { "epoch": 0.11342065504194628, "grad_norm": 1.6742401123046875, "learning_rate": 0.0001998799766424877, "loss": 5.9032, "step": 26350 }, { "epoch": 0.1136358745012289, "grad_norm": 1.5394868850708008, "learning_rate": 0.00019987950278712967, "loss": 5.6004, "step": 26400 }, { "epoch": 0.11385109396051153, "grad_norm": 1.1985183954238892, "learning_rate": 0.00019987902799878084, "loss": 6.1478, "step": 26450 }, { "epoch": 0.11406631341979416, "grad_norm": 1.1922097206115723, "learning_rate": 0.0001998785522774456, "loss": 6.0627, "step": 26500 }, { "epoch": 0.11428153287907679, "grad_norm": 1.6295381784439087, "learning_rate": 0.00019987807562312838, "loss": 5.9679, "step": 26550 }, { "epoch": 0.11449675233835943, "grad_norm": 1.8900880813598633, "learning_rate": 0.0001998775980358337, "loss": 5.6452, "step": 26600 }, { "epoch": 0.11471197179764206, "grad_norm": 1.9875677824020386, "learning_rate": 0.00019987711951556593, "loss": 5.2439, "step": 26650 }, { "epoch": 0.11492719125692469, "grad_norm": 1.8133513927459717, "learning_rate": 0.0001998766400623296, "loss": 5.9694, "step": 26700 }, { "epoch": 0.11514241071620732, "grad_norm": 0.4232478141784668, "learning_rate": 0.00019987615967612913, "loss": 5.8166, "step": 26750 }, { "epoch": 0.11535763017548994, "grad_norm": 1.5024044513702393, "learning_rate": 0.0001998756783569691, "loss": 6.0176, "step": 26800 }, { "epoch": 0.11557284963477257, "grad_norm": 1.7622610330581665, "learning_rate": 0.00019987519610485395, "loss": 5.8349, "step": 26850 }, { "epoch": 0.1157880690940552, "grad_norm": 1.3520721197128296, "learning_rate": 0.00019987471291978818, "loss": 5.3621, "step": 26900 }, { "epoch": 0.11600328855333784, "grad_norm": 1.2591204643249512, "learning_rate": 0.0001998742288017763, "loss": 5.8021, "step": 26950 }, { "epoch": 0.11621850801262047, "grad_norm": 0.4446702003479004, "learning_rate": 0.00019987374375082285, "loss": 5.1568, "step": 27000 }, { "epoch": 0.11621850801262047, "eval_loss": 6.19075870513916, "eval_runtime": 49.5408, "eval_samples_per_second": 12.919, "eval_steps_per_second": 6.459, "eval_tts_loss": 5.90734129316522, "step": 27000 }, { "epoch": 0.1164337274719031, "grad_norm": 1.408015489578247, "learning_rate": 0.00019987325776693238, "loss": 5.7044, "step": 27050 }, { "epoch": 0.11664894693118573, "grad_norm": 1.8005096912384033, "learning_rate": 0.0001998727708501094, "loss": 5.6772, "step": 27100 }, { "epoch": 0.11686416639046836, "grad_norm": 1.4133096933364868, "learning_rate": 0.00019987228300035844, "loss": 6.1483, "step": 27150 }, { "epoch": 0.11707938584975099, "grad_norm": 1.2703144550323486, "learning_rate": 0.00019987179421768412, "loss": 5.1933, "step": 27200 }, { "epoch": 0.11729460530903363, "grad_norm": 0.9948371052742004, "learning_rate": 0.00019987130450209095, "loss": 5.7162, "step": 27250 }, { "epoch": 0.11750982476831626, "grad_norm": 1.9665820598602295, "learning_rate": 0.00019987081385358353, "loss": 5.8128, "step": 27300 }, { "epoch": 0.11772504422759889, "grad_norm": 0.9630419611930847, "learning_rate": 0.00019987032227216641, "loss": 5.5708, "step": 27350 }, { "epoch": 0.11794026368688151, "grad_norm": 1.8731855154037476, "learning_rate": 0.00019986982975784426, "loss": 5.5406, "step": 27400 }, { "epoch": 0.11815548314616414, "grad_norm": 0.5045666694641113, "learning_rate": 0.00019986933631062156, "loss": 5.508, "step": 27450 }, { "epoch": 0.11837070260544677, "grad_norm": 2.080347776412964, "learning_rate": 0.00019986884193050304, "loss": 5.7528, "step": 27500 }, { "epoch": 0.1185859220647294, "grad_norm": 1.278510332107544, "learning_rate": 0.00019986834661749324, "loss": 5.3166, "step": 27550 }, { "epoch": 0.11880114152401204, "grad_norm": 0.941729724407196, "learning_rate": 0.00019986785037159684, "loss": 5.4626, "step": 27600 }, { "epoch": 0.11901636098329467, "grad_norm": 0.9016689658164978, "learning_rate": 0.00019986735319281845, "loss": 5.6751, "step": 27650 }, { "epoch": 0.1192315804425773, "grad_norm": 1.0939890146255493, "learning_rate": 0.00019986685508116268, "loss": 5.7138, "step": 27700 }, { "epoch": 0.11944679990185993, "grad_norm": 2.225604295730591, "learning_rate": 0.00019986635603663424, "loss": 5.8822, "step": 27750 }, { "epoch": 0.11966201936114255, "grad_norm": 1.2729984521865845, "learning_rate": 0.00019986585605923774, "loss": 5.4267, "step": 27800 }, { "epoch": 0.11987723882042518, "grad_norm": 1.0371261835098267, "learning_rate": 0.0001998653551489779, "loss": 5.6694, "step": 27850 }, { "epoch": 0.12009245827970781, "grad_norm": 1.1765762567520142, "learning_rate": 0.00019986485330585939, "loss": 5.5096, "step": 27900 }, { "epoch": 0.12030767773899045, "grad_norm": 1.7503007650375366, "learning_rate": 0.00019986435052988686, "loss": 5.6327, "step": 27950 }, { "epoch": 0.12052289719827308, "grad_norm": 0.34061622619628906, "learning_rate": 0.00019986384682106503, "loss": 5.7661, "step": 28000 }, { "epoch": 0.12073811665755571, "grad_norm": 1.0800831317901611, "learning_rate": 0.0001998633421793986, "loss": 5.3194, "step": 28050 }, { "epoch": 0.12095333611683834, "grad_norm": 1.671716332435608, "learning_rate": 0.00019986283660489233, "loss": 5.6265, "step": 28100 }, { "epoch": 0.12116855557612097, "grad_norm": 1.3313302993774414, "learning_rate": 0.00019986233009755086, "loss": 5.427, "step": 28150 }, { "epoch": 0.1213837750354036, "grad_norm": 1.476624608039856, "learning_rate": 0.00019986182265737898, "loss": 5.5261, "step": 28200 }, { "epoch": 0.12159899449468624, "grad_norm": 1.3727617263793945, "learning_rate": 0.00019986131428438141, "loss": 5.7686, "step": 28250 }, { "epoch": 0.12181421395396887, "grad_norm": 1.8785264492034912, "learning_rate": 0.00019986080497856292, "loss": 5.7476, "step": 28300 }, { "epoch": 0.1220294334132515, "grad_norm": 1.380555272102356, "learning_rate": 0.00019986029473992824, "loss": 5.8758, "step": 28350 }, { "epoch": 0.12224465287253412, "grad_norm": 1.447871208190918, "learning_rate": 0.00019985978356848217, "loss": 5.4412, "step": 28400 }, { "epoch": 0.12245987233181675, "grad_norm": 1.6066268682479858, "learning_rate": 0.00019985927146422942, "loss": 5.5418, "step": 28450 }, { "epoch": 0.12267509179109938, "grad_norm": 1.2039726972579956, "learning_rate": 0.00019985875842717482, "loss": 5.208, "step": 28500 }, { "epoch": 0.12289031125038201, "grad_norm": 1.4972678422927856, "learning_rate": 0.00019985824445732317, "loss": 4.8046, "step": 28550 }, { "epoch": 0.12310553070966465, "grad_norm": 0.5245859622955322, "learning_rate": 0.00019985772955467926, "loss": 5.6823, "step": 28600 }, { "epoch": 0.12332075016894728, "grad_norm": 1.3844263553619385, "learning_rate": 0.0001998572137192479, "loss": 5.2662, "step": 28650 }, { "epoch": 0.12353596962822991, "grad_norm": 1.526535153388977, "learning_rate": 0.0001998566969510339, "loss": 5.3498, "step": 28700 }, { "epoch": 0.12375118908751254, "grad_norm": 1.622756004333496, "learning_rate": 0.0001998561792500421, "loss": 5.7072, "step": 28750 }, { "epoch": 0.12396640854679516, "grad_norm": 1.3114421367645264, "learning_rate": 0.00019985566061627734, "loss": 4.9002, "step": 28800 }, { "epoch": 0.1241816280060778, "grad_norm": 1.5549030303955078, "learning_rate": 0.00019985514104974445, "loss": 5.4442, "step": 28850 }, { "epoch": 0.12439684746536042, "grad_norm": 0.14270412921905518, "learning_rate": 0.00019985462055044828, "loss": 5.5777, "step": 28900 }, { "epoch": 0.12461206692464306, "grad_norm": 1.556518316268921, "learning_rate": 0.0001998540991183937, "loss": 5.7363, "step": 28950 }, { "epoch": 0.12482728638392569, "grad_norm": 2.213775634765625, "learning_rate": 0.0001998535767535856, "loss": 5.6727, "step": 29000 }, { "epoch": 0.1250425058432083, "grad_norm": 2.1670968532562256, "learning_rate": 0.00019985305345602883, "loss": 5.7356, "step": 29050 }, { "epoch": 0.12525772530249096, "grad_norm": 0.7682716250419617, "learning_rate": 0.00019985252922572827, "loss": 5.644, "step": 29100 }, { "epoch": 0.1254729447617736, "grad_norm": 1.434509515762329, "learning_rate": 0.00019985200406268884, "loss": 5.5145, "step": 29150 }, { "epoch": 0.12568816422105622, "grad_norm": 1.236522912979126, "learning_rate": 0.00019985147796691544, "loss": 5.5184, "step": 29200 }, { "epoch": 0.12590338368033885, "grad_norm": 1.1303000450134277, "learning_rate": 0.000199850950938413, "loss": 5.8237, "step": 29250 }, { "epoch": 0.12611860313962148, "grad_norm": 1.4842880964279175, "learning_rate": 0.00019985042297718643, "loss": 5.5043, "step": 29300 }, { "epoch": 0.1263338225989041, "grad_norm": 1.331786036491394, "learning_rate": 0.00019984989408324067, "loss": 5.6198, "step": 29350 }, { "epoch": 0.12654904205818673, "grad_norm": 0.7231821417808533, "learning_rate": 0.00019984936425658062, "loss": 5.6148, "step": 29400 }, { "epoch": 0.12676426151746936, "grad_norm": 0.29656919836997986, "learning_rate": 0.00019984883349721127, "loss": 5.569, "step": 29450 }, { "epoch": 0.126979480976752, "grad_norm": 1.393602728843689, "learning_rate": 0.0001998483018051376, "loss": 5.7558, "step": 29500 }, { "epoch": 0.12719470043603462, "grad_norm": 1.8038228750228882, "learning_rate": 0.00019984776918036452, "loss": 5.8036, "step": 29550 }, { "epoch": 0.12740991989531725, "grad_norm": 1.806577444076538, "learning_rate": 0.00019984723562289705, "loss": 5.3018, "step": 29600 }, { "epoch": 0.12762513935459988, "grad_norm": 0.9906163811683655, "learning_rate": 0.0001998467011327401, "loss": 5.6327, "step": 29650 }, { "epoch": 0.1278403588138825, "grad_norm": 1.9192959070205688, "learning_rate": 0.00019984616570989876, "loss": 5.4614, "step": 29700 }, { "epoch": 0.12805557827316513, "grad_norm": 1.3354132175445557, "learning_rate": 0.000199845629354378, "loss": 5.454, "step": 29750 }, { "epoch": 0.1282707977324478, "grad_norm": 0.9859639406204224, "learning_rate": 0.00019984509206618278, "loss": 5.9884, "step": 29800 }, { "epoch": 0.12848601719173042, "grad_norm": 0.9074296951293945, "learning_rate": 0.0001998445538453182, "loss": 5.31, "step": 29850 }, { "epoch": 0.12870123665101305, "grad_norm": 1.5731911659240723, "learning_rate": 0.00019984401469178923, "loss": 5.3923, "step": 29900 }, { "epoch": 0.12891645611029567, "grad_norm": 1.4788398742675781, "learning_rate": 0.0001998434746056009, "loss": 5.6249, "step": 29950 }, { "epoch": 0.1291316755695783, "grad_norm": 0.8341617584228516, "learning_rate": 0.0001998429335867583, "loss": 5.5265, "step": 30000 }, { "epoch": 0.1291316755695783, "eval_loss": 6.147937774658203, "eval_runtime": 49.6095, "eval_samples_per_second": 12.901, "eval_steps_per_second": 6.45, "eval_tts_loss": 6.082416508944742, "step": 30000 }, { "epoch": 0.12934689502886093, "grad_norm": 1.0650166273117065, "learning_rate": 0.00019984239163526644, "loss": 5.4653, "step": 30050 }, { "epoch": 0.12956211448814356, "grad_norm": 0.9423066973686218, "learning_rate": 0.00019984184875113039, "loss": 5.087, "step": 30100 }, { "epoch": 0.1297773339474262, "grad_norm": 3.0054309368133545, "learning_rate": 0.00019984130493435527, "loss": 6.2012, "step": 30150 }, { "epoch": 0.12999255340670882, "grad_norm": 1.743401288986206, "learning_rate": 0.00019984076018494612, "loss": 5.4087, "step": 30200 }, { "epoch": 0.13020777286599144, "grad_norm": 1.3028907775878906, "learning_rate": 0.000199840214502908, "loss": 5.36, "step": 30250 }, { "epoch": 0.13042299232527407, "grad_norm": 1.2445651292800903, "learning_rate": 0.00019983966788824606, "loss": 5.5131, "step": 30300 }, { "epoch": 0.1306382117845567, "grad_norm": 0.42892858386039734, "learning_rate": 0.00019983912034096538, "loss": 5.8252, "step": 30350 }, { "epoch": 0.13085343124383933, "grad_norm": 2.7153801918029785, "learning_rate": 0.00019983857186107112, "loss": 5.498, "step": 30400 }, { "epoch": 0.13106865070312199, "grad_norm": 1.0288418531417847, "learning_rate": 0.00019983802244856832, "loss": 5.5103, "step": 30450 }, { "epoch": 0.13128387016240461, "grad_norm": 0.9950653314590454, "learning_rate": 0.00019983747210346216, "loss": 5.3058, "step": 30500 }, { "epoch": 0.13149908962168724, "grad_norm": 2.4451732635498047, "learning_rate": 0.00019983692082575777, "loss": 5.427, "step": 30550 }, { "epoch": 0.13171430908096987, "grad_norm": 0.27538686990737915, "learning_rate": 0.00019983636861546034, "loss": 5.7099, "step": 30600 }, { "epoch": 0.1319295285402525, "grad_norm": 1.6789100170135498, "learning_rate": 0.00019983581547257498, "loss": 5.2249, "step": 30650 }, { "epoch": 0.13214474799953513, "grad_norm": 0.7159422039985657, "learning_rate": 0.00019983526139710685, "loss": 5.5704, "step": 30700 }, { "epoch": 0.13235996745881776, "grad_norm": 1.239390254020691, "learning_rate": 0.00019983470638906117, "loss": 6.1733, "step": 30750 }, { "epoch": 0.13257518691810039, "grad_norm": 0.3989431858062744, "learning_rate": 0.00019983415044844312, "loss": 5.9651, "step": 30800 }, { "epoch": 0.132790406377383, "grad_norm": 1.3549634218215942, "learning_rate": 0.00019983359357525783, "loss": 5.7785, "step": 30850 }, { "epoch": 0.13300562583666564, "grad_norm": 1.5584107637405396, "learning_rate": 0.00019983303576951057, "loss": 5.5523, "step": 30900 }, { "epoch": 0.13322084529594827, "grad_norm": 0.3714735507965088, "learning_rate": 0.00019983247703120652, "loss": 5.7503, "step": 30950 }, { "epoch": 0.1334360647552309, "grad_norm": 0.8539810180664062, "learning_rate": 0.0001998319173603509, "loss": 5.4318, "step": 31000 }, { "epoch": 0.13365128421451353, "grad_norm": 1.2168965339660645, "learning_rate": 0.00019983135675694897, "loss": 5.8627, "step": 31050 }, { "epoch": 0.13386650367379618, "grad_norm": 1.1346267461776733, "learning_rate": 0.00019983079522100595, "loss": 5.7775, "step": 31100 }, { "epoch": 0.1340817231330788, "grad_norm": 0.8613104224205017, "learning_rate": 0.00019983023275252704, "loss": 5.8492, "step": 31150 }, { "epoch": 0.13429694259236144, "grad_norm": 0.19876469671726227, "learning_rate": 0.00019982966935151755, "loss": 5.7716, "step": 31200 }, { "epoch": 0.13451216205164407, "grad_norm": 1.2013776302337646, "learning_rate": 0.0001998291050179827, "loss": 6.191, "step": 31250 }, { "epoch": 0.1347273815109267, "grad_norm": 2.05531907081604, "learning_rate": 0.00019982853975192782, "loss": 5.8226, "step": 31300 }, { "epoch": 0.13494260097020933, "grad_norm": 1.8567557334899902, "learning_rate": 0.00019982797355335815, "loss": 5.0057, "step": 31350 }, { "epoch": 0.13515782042949195, "grad_norm": 0.9962878227233887, "learning_rate": 0.00019982740642227897, "loss": 4.9752, "step": 31400 }, { "epoch": 0.13537303988877458, "grad_norm": 1.505607008934021, "learning_rate": 0.00019982683835869557, "loss": 5.7071, "step": 31450 }, { "epoch": 0.1355882593480572, "grad_norm": 1.463348150253296, "learning_rate": 0.0001998262693626133, "loss": 5.6676, "step": 31500 }, { "epoch": 0.13580347880733984, "grad_norm": 1.7306724786758423, "learning_rate": 0.00019982569943403745, "loss": 5.8388, "step": 31550 }, { "epoch": 0.13601869826662247, "grad_norm": 1.5143413543701172, "learning_rate": 0.00019982512857297339, "loss": 5.122, "step": 31600 }, { "epoch": 0.1362339177259051, "grad_norm": 1.7233524322509766, "learning_rate": 0.00019982455677942636, "loss": 5.8926, "step": 31650 }, { "epoch": 0.13644913718518772, "grad_norm": 1.497884750366211, "learning_rate": 0.00019982398405340177, "loss": 5.1684, "step": 31700 }, { "epoch": 0.13666435664447035, "grad_norm": 1.3915332555770874, "learning_rate": 0.00019982341039490495, "loss": 5.6456, "step": 31750 }, { "epoch": 0.136879576103753, "grad_norm": 1.213780164718628, "learning_rate": 0.00019982283580394125, "loss": 5.9512, "step": 31800 }, { "epoch": 0.13709479556303564, "grad_norm": 1.0388180017471313, "learning_rate": 0.00019982226028051605, "loss": 5.4373, "step": 31850 }, { "epoch": 0.13731001502231827, "grad_norm": 1.5036028623580933, "learning_rate": 0.00019982168382463473, "loss": 5.6535, "step": 31900 }, { "epoch": 0.1375252344816009, "grad_norm": 1.3470958471298218, "learning_rate": 0.00019982110643630265, "loss": 5.7745, "step": 31950 }, { "epoch": 0.13774045394088352, "grad_norm": 1.090550184249878, "learning_rate": 0.00019982052811552526, "loss": 5.9259, "step": 32000 }, { "epoch": 0.13795567340016615, "grad_norm": 1.0603644847869873, "learning_rate": 0.00019981994886230788, "loss": 5.5144, "step": 32050 }, { "epoch": 0.13817089285944878, "grad_norm": 0.9281832575798035, "learning_rate": 0.000199819368676656, "loss": 5.6133, "step": 32100 }, { "epoch": 0.1383861123187314, "grad_norm": 1.0886186361312866, "learning_rate": 0.00019981878755857496, "loss": 5.6127, "step": 32150 }, { "epoch": 0.13860133177801404, "grad_norm": 1.3918582201004028, "learning_rate": 0.00019981820550807027, "loss": 5.7281, "step": 32200 }, { "epoch": 0.13881655123729666, "grad_norm": 1.7270874977111816, "learning_rate": 0.00019981762252514732, "loss": 5.6989, "step": 32250 }, { "epoch": 0.1390317706965793, "grad_norm": 1.3960040807724, "learning_rate": 0.00019981703860981156, "loss": 5.7557, "step": 32300 }, { "epoch": 0.13924699015586192, "grad_norm": 0.6205440163612366, "learning_rate": 0.00019981645376206846, "loss": 5.7684, "step": 32350 }, { "epoch": 0.13946220961514455, "grad_norm": 1.0669981241226196, "learning_rate": 0.00019981586798192345, "loss": 5.7166, "step": 32400 }, { "epoch": 0.1396774290744272, "grad_norm": 2.1473865509033203, "learning_rate": 0.00019981528126938205, "loss": 5.7392, "step": 32450 }, { "epoch": 0.13989264853370983, "grad_norm": 1.7093976736068726, "learning_rate": 0.0001998146936244497, "loss": 5.4897, "step": 32500 }, { "epoch": 0.14010786799299246, "grad_norm": 0.9719522595405579, "learning_rate": 0.00019981410504713192, "loss": 5.359, "step": 32550 }, { "epoch": 0.1403230874522751, "grad_norm": 2.0018718242645264, "learning_rate": 0.00019981351553743416, "loss": 5.2929, "step": 32600 }, { "epoch": 0.14053830691155772, "grad_norm": 0.37229645252227783, "learning_rate": 0.00019981292509536197, "loss": 5.1639, "step": 32650 }, { "epoch": 0.14075352637084035, "grad_norm": 1.2296282052993774, "learning_rate": 0.00019981233372092088, "loss": 5.7437, "step": 32700 }, { "epoch": 0.14096874583012298, "grad_norm": 2.861159086227417, "learning_rate": 0.0001998117414141164, "loss": 5.3622, "step": 32750 }, { "epoch": 0.1411839652894056, "grad_norm": 1.3920012712478638, "learning_rate": 0.00019981114817495398, "loss": 5.8697, "step": 32800 }, { "epoch": 0.14139918474868823, "grad_norm": 1.0156110525131226, "learning_rate": 0.0001998105540034393, "loss": 5.5156, "step": 32850 }, { "epoch": 0.14161440420797086, "grad_norm": 1.5631990432739258, "learning_rate": 0.0001998099588995778, "loss": 5.4872, "step": 32900 }, { "epoch": 0.1418296236672535, "grad_norm": 1.8582583665847778, "learning_rate": 0.00019980936286337513, "loss": 5.4971, "step": 32950 }, { "epoch": 0.14204484312653612, "grad_norm": 2.29675555229187, "learning_rate": 0.00019980876589483678, "loss": 5.8802, "step": 33000 }, { "epoch": 0.14204484312653612, "eval_loss": 6.09331750869751, "eval_runtime": 49.5644, "eval_samples_per_second": 12.912, "eval_steps_per_second": 6.456, "eval_tts_loss": 6.072868523098193, "step": 33000 }, { "epoch": 0.14226006258581875, "grad_norm": 1.5237661600112915, "learning_rate": 0.00019980816799396834, "loss": 5.8206, "step": 33050 }, { "epoch": 0.1424752820451014, "grad_norm": 1.9801349639892578, "learning_rate": 0.00019980756916077543, "loss": 5.4556, "step": 33100 }, { "epoch": 0.14269050150438403, "grad_norm": 1.5417542457580566, "learning_rate": 0.00019980696939526362, "loss": 5.7238, "step": 33150 }, { "epoch": 0.14290572096366666, "grad_norm": 1.937732458114624, "learning_rate": 0.0001998063686974385, "loss": 5.8057, "step": 33200 }, { "epoch": 0.1431209404229493, "grad_norm": 0.5770896077156067, "learning_rate": 0.00019980576706730573, "loss": 5.4066, "step": 33250 }, { "epoch": 0.14333615988223192, "grad_norm": 1.4596003293991089, "learning_rate": 0.0001998051645048709, "loss": 5.8776, "step": 33300 }, { "epoch": 0.14355137934151455, "grad_norm": 2.772312641143799, "learning_rate": 0.00019980456101013964, "loss": 5.3036, "step": 33350 }, { "epoch": 0.14376659880079717, "grad_norm": 1.4782541990280151, "learning_rate": 0.00019980395658311757, "loss": 6.2243, "step": 33400 }, { "epoch": 0.1439818182600798, "grad_norm": 1.0340211391448975, "learning_rate": 0.00019980335122381034, "loss": 5.7385, "step": 33450 }, { "epoch": 0.14419703771936243, "grad_norm": 0.5317725539207458, "learning_rate": 0.00019980274493222364, "loss": 5.4906, "step": 33500 }, { "epoch": 0.14441225717864506, "grad_norm": 0.5881343483924866, "learning_rate": 0.0001998021377083631, "loss": 5.3772, "step": 33550 }, { "epoch": 0.1446274766379277, "grad_norm": 1.6957509517669678, "learning_rate": 0.00019980152955223439, "loss": 5.399, "step": 33600 }, { "epoch": 0.14484269609721032, "grad_norm": 0.4648636281490326, "learning_rate": 0.00019980092046384323, "loss": 4.9256, "step": 33650 }, { "epoch": 0.14505791555649294, "grad_norm": 0.9518867135047913, "learning_rate": 0.00019980031044319524, "loss": 5.7382, "step": 33700 }, { "epoch": 0.14527313501577557, "grad_norm": 0.2596074640750885, "learning_rate": 0.0001997996994902962, "loss": 5.5104, "step": 33750 }, { "epoch": 0.14548835447505823, "grad_norm": 1.5997047424316406, "learning_rate": 0.00019979908760515175, "loss": 5.8571, "step": 33800 }, { "epoch": 0.14570357393434086, "grad_norm": 1.8062348365783691, "learning_rate": 0.00019979847478776764, "loss": 5.2089, "step": 33850 }, { "epoch": 0.14591879339362349, "grad_norm": 1.623335838317871, "learning_rate": 0.0001997978610381496, "loss": 5.4883, "step": 33900 }, { "epoch": 0.14613401285290611, "grad_norm": 2.0395865440368652, "learning_rate": 0.00019979724635630332, "loss": 5.4164, "step": 33950 }, { "epoch": 0.14634923231218874, "grad_norm": 2.096801996231079, "learning_rate": 0.00019979663074223462, "loss": 5.9087, "step": 34000 }, { "epoch": 0.14656445177147137, "grad_norm": 3.7746353149414062, "learning_rate": 0.00019979601419594915, "loss": 5.8641, "step": 34050 }, { "epoch": 0.146779671230754, "grad_norm": 1.5706762075424194, "learning_rate": 0.00019979539671745276, "loss": 5.5816, "step": 34100 }, { "epoch": 0.14699489069003663, "grad_norm": 1.4360520839691162, "learning_rate": 0.00019979477830675115, "loss": 5.4274, "step": 34150 }, { "epoch": 0.14721011014931926, "grad_norm": 0.9412097334861755, "learning_rate": 0.00019979415896385013, "loss": 4.9226, "step": 34200 }, { "epoch": 0.14742532960860188, "grad_norm": 1.159110426902771, "learning_rate": 0.0001997935386887555, "loss": 5.1073, "step": 34250 }, { "epoch": 0.1476405490678845, "grad_norm": 0.8718621134757996, "learning_rate": 0.00019979291748147305, "loss": 5.2803, "step": 34300 }, { "epoch": 0.14785576852716714, "grad_norm": 2.0006492137908936, "learning_rate": 0.00019979229534200854, "loss": 5.9243, "step": 34350 }, { "epoch": 0.14807098798644977, "grad_norm": 1.357682228088379, "learning_rate": 0.00019979167227036784, "loss": 5.5413, "step": 34400 }, { "epoch": 0.14828620744573243, "grad_norm": 1.542575716972351, "learning_rate": 0.0001997910482665567, "loss": 5.7374, "step": 34450 }, { "epoch": 0.14850142690501505, "grad_norm": 1.928341031074524, "learning_rate": 0.000199790423330581, "loss": 5.4783, "step": 34500 }, { "epoch": 0.14871664636429768, "grad_norm": 1.1719046831130981, "learning_rate": 0.00019978979746244658, "loss": 5.7546, "step": 34550 }, { "epoch": 0.1489318658235803, "grad_norm": 1.0099287033081055, "learning_rate": 0.00019978917066215928, "loss": 5.521, "step": 34600 }, { "epoch": 0.14914708528286294, "grad_norm": 0.9869875311851501, "learning_rate": 0.00019978854292972492, "loss": 5.3479, "step": 34650 }, { "epoch": 0.14936230474214557, "grad_norm": 1.0274266004562378, "learning_rate": 0.00019978791426514942, "loss": 5.0612, "step": 34700 }, { "epoch": 0.1495775242014282, "grad_norm": 2.3290297985076904, "learning_rate": 0.00019978728466843863, "loss": 5.396, "step": 34750 }, { "epoch": 0.14979274366071083, "grad_norm": 1.0535537004470825, "learning_rate": 0.00019978665413959842, "loss": 5.5873, "step": 34800 }, { "epoch": 0.15000796311999345, "grad_norm": 0.9407364726066589, "learning_rate": 0.00019978602267863466, "loss": 5.8635, "step": 34850 }, { "epoch": 0.15022318257927608, "grad_norm": 1.2329208850860596, "learning_rate": 0.0001997853902855533, "loss": 5.7122, "step": 34900 }, { "epoch": 0.1504384020385587, "grad_norm": 1.7497210502624512, "learning_rate": 0.00019978475696036022, "loss": 5.6273, "step": 34950 }, { "epoch": 0.15065362149784134, "grad_norm": 1.5199528932571411, "learning_rate": 0.00019978412270306133, "loss": 5.9487, "step": 35000 }, { "epoch": 0.15086884095712397, "grad_norm": 2.017172336578369, "learning_rate": 0.00019978348751366256, "loss": 5.3335, "step": 35050 }, { "epoch": 0.15108406041640662, "grad_norm": 0.26009300351142883, "learning_rate": 0.00019978285139216984, "loss": 5.8214, "step": 35100 }, { "epoch": 0.15129927987568925, "grad_norm": 0.8831576704978943, "learning_rate": 0.00019978221433858913, "loss": 5.9723, "step": 35150 }, { "epoch": 0.15151449933497188, "grad_norm": 1.0311883687973022, "learning_rate": 0.00019978157635292635, "loss": 5.731, "step": 35200 }, { "epoch": 0.1517297187942545, "grad_norm": 2.034931182861328, "learning_rate": 0.00019978093743518751, "loss": 5.824, "step": 35250 }, { "epoch": 0.15194493825353714, "grad_norm": 1.117568850517273, "learning_rate": 0.00019978029758537853, "loss": 5.6572, "step": 35300 }, { "epoch": 0.15216015771281977, "grad_norm": 0.9784127473831177, "learning_rate": 0.00019977965680350538, "loss": 5.3788, "step": 35350 }, { "epoch": 0.1523753771721024, "grad_norm": 1.6062155961990356, "learning_rate": 0.0001997790150895741, "loss": 5.4571, "step": 35400 }, { "epoch": 0.15259059663138502, "grad_norm": 1.8582569360733032, "learning_rate": 0.00019977837244359065, "loss": 5.839, "step": 35450 }, { "epoch": 0.15280581609066765, "grad_norm": 0.8657233119010925, "learning_rate": 0.00019977772886556103, "loss": 5.3385, "step": 35500 }, { "epoch": 0.15302103554995028, "grad_norm": 1.4130641222000122, "learning_rate": 0.00019977708435549128, "loss": 5.8123, "step": 35550 }, { "epoch": 0.1532362550092329, "grad_norm": 0.5357573628425598, "learning_rate": 0.00019977643891338739, "loss": 5.3959, "step": 35600 }, { "epoch": 0.15345147446851554, "grad_norm": 1.8519198894500732, "learning_rate": 0.00019977579253925538, "loss": 5.81, "step": 35650 }, { "epoch": 0.15366669392779816, "grad_norm": 2.1200778484344482, "learning_rate": 0.0001997751452331013, "loss": 5.8327, "step": 35700 }, { "epoch": 0.1538819133870808, "grad_norm": 1.43500554561615, "learning_rate": 0.00019977449699493121, "loss": 5.8615, "step": 35750 }, { "epoch": 0.15409713284636345, "grad_norm": 1.3869377374649048, "learning_rate": 0.00019977384782475117, "loss": 5.4452, "step": 35800 }, { "epoch": 0.15431235230564608, "grad_norm": 2.8594810962677, "learning_rate": 0.00019977319772256723, "loss": 5.176, "step": 35850 }, { "epoch": 0.1545275717649287, "grad_norm": 0.4655211865901947, "learning_rate": 0.00019977254668838545, "loss": 5.0815, "step": 35900 }, { "epoch": 0.15474279122421133, "grad_norm": 1.1038111448287964, "learning_rate": 0.00019977189472221194, "loss": 5.7572, "step": 35950 }, { "epoch": 0.15495801068349396, "grad_norm": 1.1350911855697632, "learning_rate": 0.00019977124182405276, "loss": 5.792, "step": 36000 }, { "epoch": 0.15495801068349396, "eval_loss": 6.076951026916504, "eval_runtime": 49.7131, "eval_samples_per_second": 12.874, "eval_steps_per_second": 6.437, "eval_tts_loss": 6.026644844018236, "step": 36000 }, { "epoch": 0.1551732301427766, "grad_norm": 0.35101842880249023, "learning_rate": 0.00019977058799391405, "loss": 5.676, "step": 36050 }, { "epoch": 0.15538844960205922, "grad_norm": 1.0075030326843262, "learning_rate": 0.00019976993323180187, "loss": 5.1997, "step": 36100 }, { "epoch": 0.15560366906134185, "grad_norm": 0.7201313972473145, "learning_rate": 0.00019976927753772236, "loss": 5.5899, "step": 36150 }, { "epoch": 0.15581888852062448, "grad_norm": 2.491495370864868, "learning_rate": 0.00019976862091168165, "loss": 5.7507, "step": 36200 }, { "epoch": 0.1560341079799071, "grad_norm": 3.5400898456573486, "learning_rate": 0.00019976796335368589, "loss": 5.5169, "step": 36250 }, { "epoch": 0.15624932743918973, "grad_norm": 0.9274168610572815, "learning_rate": 0.00019976730486374116, "loss": 5.8572, "step": 36300 }, { "epoch": 0.15646454689847236, "grad_norm": 1.3975634574890137, "learning_rate": 0.00019976664544185365, "loss": 5.8779, "step": 36350 }, { "epoch": 0.156679766357755, "grad_norm": 0.9227761626243591, "learning_rate": 0.00019976598508802956, "loss": 5.539, "step": 36400 }, { "epoch": 0.15689498581703765, "grad_norm": 1.708583950996399, "learning_rate": 0.00019976532380227497, "loss": 5.2868, "step": 36450 }, { "epoch": 0.15711020527632027, "grad_norm": 1.3517730236053467, "learning_rate": 0.00019976466158459614, "loss": 5.5348, "step": 36500 }, { "epoch": 0.1573254247356029, "grad_norm": 1.0254641771316528, "learning_rate": 0.00019976399843499922, "loss": 5.4263, "step": 36550 }, { "epoch": 0.15754064419488553, "grad_norm": 2.0401008129119873, "learning_rate": 0.0001997633343534904, "loss": 5.0699, "step": 36600 }, { "epoch": 0.15775586365416816, "grad_norm": 1.2070385217666626, "learning_rate": 0.00019976266934007588, "loss": 5.6586, "step": 36650 }, { "epoch": 0.1579710831134508, "grad_norm": 0.9666840434074402, "learning_rate": 0.0001997620033947619, "loss": 4.8969, "step": 36700 }, { "epoch": 0.15818630257273342, "grad_norm": 0.8969870805740356, "learning_rate": 0.00019976133651755464, "loss": 5.8093, "step": 36750 }, { "epoch": 0.15840152203201605, "grad_norm": 1.4847493171691895, "learning_rate": 0.00019976066870846038, "loss": 5.6451, "step": 36800 }, { "epoch": 0.15861674149129867, "grad_norm": 1.0873966217041016, "learning_rate": 0.00019975999996748532, "loss": 5.5185, "step": 36850 }, { "epoch": 0.1588319609505813, "grad_norm": 0.8979986906051636, "learning_rate": 0.00019975933029463573, "loss": 5.7265, "step": 36900 }, { "epoch": 0.15904718040986393, "grad_norm": 1.5697710514068604, "learning_rate": 0.0001997586596899178, "loss": 5.6779, "step": 36950 }, { "epoch": 0.15926239986914656, "grad_norm": 1.1667039394378662, "learning_rate": 0.0001997579881533379, "loss": 5.0841, "step": 37000 }, { "epoch": 0.1594776193284292, "grad_norm": 1.1385772228240967, "learning_rate": 0.00019975731568490222, "loss": 5.6863, "step": 37050 }, { "epoch": 0.15969283878771182, "grad_norm": 0.504422664642334, "learning_rate": 0.00019975664228461708, "loss": 4.9333, "step": 37100 }, { "epoch": 0.15990805824699447, "grad_norm": 1.9248028993606567, "learning_rate": 0.00019975596795248875, "loss": 5.6364, "step": 37150 }, { "epoch": 0.1601232777062771, "grad_norm": 1.4344807863235474, "learning_rate": 0.00019975529268852357, "loss": 5.4466, "step": 37200 }, { "epoch": 0.16033849716555973, "grad_norm": 0.2080679088830948, "learning_rate": 0.0001997546164927278, "loss": 5.0835, "step": 37250 }, { "epoch": 0.16055371662484236, "grad_norm": 1.6943013668060303, "learning_rate": 0.00019975393936510778, "loss": 5.6713, "step": 37300 }, { "epoch": 0.16076893608412499, "grad_norm": 1.8372087478637695, "learning_rate": 0.0001997532613056698, "loss": 5.4445, "step": 37350 }, { "epoch": 0.16098415554340761, "grad_norm": 3.009153127670288, "learning_rate": 0.00019975258231442027, "loss": 5.495, "step": 37400 }, { "epoch": 0.16119937500269024, "grad_norm": 1.1257492303848267, "learning_rate": 0.00019975190239136545, "loss": 5.7473, "step": 37450 }, { "epoch": 0.16141459446197287, "grad_norm": 0.4329243302345276, "learning_rate": 0.00019975122153651173, "loss": 6.121, "step": 37500 }, { "epoch": 0.1616298139212555, "grad_norm": 1.1361511945724487, "learning_rate": 0.0001997505397498655, "loss": 5.6468, "step": 37550 }, { "epoch": 0.16184503338053813, "grad_norm": 1.1076725721359253, "learning_rate": 0.0001997498570314331, "loss": 5.6562, "step": 37600 }, { "epoch": 0.16206025283982076, "grad_norm": 0.893677294254303, "learning_rate": 0.00019974917338122085, "loss": 5.4809, "step": 37650 }, { "epoch": 0.16227547229910338, "grad_norm": 0.943320631980896, "learning_rate": 0.00019974848879923522, "loss": 5.7874, "step": 37700 }, { "epoch": 0.162490691758386, "grad_norm": 1.5618548393249512, "learning_rate": 0.00019974780328548258, "loss": 5.6303, "step": 37750 }, { "epoch": 0.16270591121766867, "grad_norm": 1.3040452003479004, "learning_rate": 0.00019974711683996935, "loss": 6.0399, "step": 37800 }, { "epoch": 0.1629211306769513, "grad_norm": 0.15198585391044617, "learning_rate": 0.00019974642946270189, "loss": 5.7157, "step": 37850 }, { "epoch": 0.16313635013623393, "grad_norm": 0.8637711405754089, "learning_rate": 0.00019974574115368663, "loss": 5.284, "step": 37900 }, { "epoch": 0.16335156959551655, "grad_norm": 1.1571484804153442, "learning_rate": 0.00019974505191293004, "loss": 5.9375, "step": 37950 }, { "epoch": 0.16356678905479918, "grad_norm": 1.8972350358963013, "learning_rate": 0.00019974436174043857, "loss": 4.8972, "step": 38000 }, { "epoch": 0.1637820085140818, "grad_norm": 0.7007231712341309, "learning_rate": 0.0001997436706362186, "loss": 5.495, "step": 38050 }, { "epoch": 0.16399722797336444, "grad_norm": 1.2033227682113647, "learning_rate": 0.00019974297860027662, "loss": 5.264, "step": 38100 }, { "epoch": 0.16421244743264707, "grad_norm": 1.1766602993011475, "learning_rate": 0.00019974228563261913, "loss": 5.8971, "step": 38150 }, { "epoch": 0.1644276668919297, "grad_norm": 0.5057514905929565, "learning_rate": 0.00019974159173325254, "loss": 5.3752, "step": 38200 }, { "epoch": 0.16464288635121233, "grad_norm": 2.0079345703125, "learning_rate": 0.00019974089690218336, "loss": 5.6291, "step": 38250 }, { "epoch": 0.16485810581049495, "grad_norm": 0.5208863019943237, "learning_rate": 0.0001997402011394181, "loss": 5.466, "step": 38300 }, { "epoch": 0.16507332526977758, "grad_norm": 1.188796043395996, "learning_rate": 0.0001997395044449632, "loss": 5.3487, "step": 38350 }, { "epoch": 0.1652885447290602, "grad_norm": 0.8434421420097351, "learning_rate": 0.00019973880681882524, "loss": 5.4176, "step": 38400 }, { "epoch": 0.16550376418834287, "grad_norm": 0.931089460849762, "learning_rate": 0.0001997381082610107, "loss": 5.6114, "step": 38450 }, { "epoch": 0.1657189836476255, "grad_norm": 0.8881914615631104, "learning_rate": 0.0001997374087715261, "loss": 5.6861, "step": 38500 }, { "epoch": 0.16593420310690812, "grad_norm": 2.0178794860839844, "learning_rate": 0.00019973670835037803, "loss": 5.3612, "step": 38550 }, { "epoch": 0.16614942256619075, "grad_norm": 1.0948411226272583, "learning_rate": 0.00019973600699757296, "loss": 5.2905, "step": 38600 }, { "epoch": 0.16636464202547338, "grad_norm": 1.5716742277145386, "learning_rate": 0.00019973530471311746, "loss": 5.3598, "step": 38650 }, { "epoch": 0.166579861484756, "grad_norm": 1.3890436887741089, "learning_rate": 0.0001997346014970181, "loss": 5.7316, "step": 38700 }, { "epoch": 0.16679508094403864, "grad_norm": 1.4843543767929077, "learning_rate": 0.00019973389734928145, "loss": 5.7356, "step": 38750 }, { "epoch": 0.16701030040332127, "grad_norm": 0.98454350233078, "learning_rate": 0.0001997331922699141, "loss": 5.5709, "step": 38800 }, { "epoch": 0.1672255198626039, "grad_norm": 1.7049065828323364, "learning_rate": 0.00019973248625892263, "loss": 5.5222, "step": 38850 }, { "epoch": 0.16744073932188652, "grad_norm": 0.2527773082256317, "learning_rate": 0.0001997317793163136, "loss": 5.2326, "step": 38900 }, { "epoch": 0.16765595878116915, "grad_norm": 0.536939799785614, "learning_rate": 0.00019973107144209364, "loss": 4.9911, "step": 38950 }, { "epoch": 0.16787117824045178, "grad_norm": 1.2047640085220337, "learning_rate": 0.0001997303626362694, "loss": 5.5245, "step": 39000 }, { "epoch": 0.16787117824045178, "eval_loss": 6.026686668395996, "eval_runtime": 49.5582, "eval_samples_per_second": 12.914, "eval_steps_per_second": 6.457, "eval_tts_loss": 6.081446065124039, "step": 39000 }, { "epoch": 0.1680863976997344, "grad_norm": 1.0895717144012451, "learning_rate": 0.0001997296528988474, "loss": 5.1903, "step": 39050 }, { "epoch": 0.16830161715901704, "grad_norm": 1.4740958213806152, "learning_rate": 0.0001997289422298344, "loss": 5.1494, "step": 39100 }, { "epoch": 0.1685168366182997, "grad_norm": 1.3724490404129028, "learning_rate": 0.0001997282306292369, "loss": 5.5492, "step": 39150 }, { "epoch": 0.16873205607758232, "grad_norm": 0.9725663065910339, "learning_rate": 0.00019972751809706167, "loss": 5.3666, "step": 39200 }, { "epoch": 0.16894727553686495, "grad_norm": 0.9600479006767273, "learning_rate": 0.0001997268046333153, "loss": 5.3816, "step": 39250 }, { "epoch": 0.16916249499614758, "grad_norm": 2.41348934173584, "learning_rate": 0.0001997260902380045, "loss": 5.3915, "step": 39300 }, { "epoch": 0.1693777144554302, "grad_norm": 1.3441704511642456, "learning_rate": 0.00019972537491113587, "loss": 5.3966, "step": 39350 }, { "epoch": 0.16959293391471283, "grad_norm": 0.9384105205535889, "learning_rate": 0.00019972465865271616, "loss": 5.3757, "step": 39400 }, { "epoch": 0.16980815337399546, "grad_norm": 1.0002397298812866, "learning_rate": 0.00019972394146275204, "loss": 5.2036, "step": 39450 }, { "epoch": 0.1700233728332781, "grad_norm": 1.3420881032943726, "learning_rate": 0.00019972322334125018, "loss": 5.5885, "step": 39500 }, { "epoch": 0.17023859229256072, "grad_norm": 1.9879915714263916, "learning_rate": 0.00019972250428821733, "loss": 5.0416, "step": 39550 }, { "epoch": 0.17045381175184335, "grad_norm": 0.9488893747329712, "learning_rate": 0.0001997217843036602, "loss": 5.5289, "step": 39600 }, { "epoch": 0.17066903121112598, "grad_norm": 1.2084654569625854, "learning_rate": 0.0001997210633875855, "loss": 5.6683, "step": 39650 }, { "epoch": 0.1708842506704086, "grad_norm": 0.7972135543823242, "learning_rate": 0.00019972034153999994, "loss": 5.5555, "step": 39700 }, { "epoch": 0.17109947012969123, "grad_norm": 1.1700220108032227, "learning_rate": 0.00019971961876091032, "loss": 5.7032, "step": 39750 }, { "epoch": 0.1713146895889739, "grad_norm": 1.6076688766479492, "learning_rate": 0.00019971889505032334, "loss": 5.4789, "step": 39800 }, { "epoch": 0.17152990904825652, "grad_norm": 1.4435153007507324, "learning_rate": 0.00019971817040824582, "loss": 5.8032, "step": 39850 }, { "epoch": 0.17174512850753915, "grad_norm": 0.3460695743560791, "learning_rate": 0.0001997174448346845, "loss": 4.9887, "step": 39900 }, { "epoch": 0.17196034796682177, "grad_norm": 2.322877883911133, "learning_rate": 0.0001997167183296461, "loss": 5.5289, "step": 39950 }, { "epoch": 0.1721755674261044, "grad_norm": 0.885301947593689, "learning_rate": 0.0001997159908931375, "loss": 5.8789, "step": 40000 }, { "epoch": 0.17239078688538703, "grad_norm": 0.38468772172927856, "learning_rate": 0.00019971526252516545, "loss": 5.2627, "step": 40050 }, { "epoch": 0.17260600634466966, "grad_norm": 0.8022739291191101, "learning_rate": 0.0001997145332257367, "loss": 5.3788, "step": 40100 }, { "epoch": 0.1728212258039523, "grad_norm": 0.94605553150177, "learning_rate": 0.0001997138029948582, "loss": 5.6267, "step": 40150 }, { "epoch": 0.17303644526323492, "grad_norm": 1.0827075242996216, "learning_rate": 0.00019971307183253668, "loss": 5.219, "step": 40200 }, { "epoch": 0.17325166472251755, "grad_norm": 0.9943904280662537, "learning_rate": 0.0001997123397387789, "loss": 5.761, "step": 40250 }, { "epoch": 0.17346688418180017, "grad_norm": 2.0678598880767822, "learning_rate": 0.00019971160671359187, "loss": 5.7093, "step": 40300 }, { "epoch": 0.1736821036410828, "grad_norm": 0.8029616475105286, "learning_rate": 0.0001997108727569823, "loss": 5.3092, "step": 40350 }, { "epoch": 0.17389732310036543, "grad_norm": 1.590114712715149, "learning_rate": 0.0001997101378689571, "loss": 5.7909, "step": 40400 }, { "epoch": 0.1741125425596481, "grad_norm": 0.9241282939910889, "learning_rate": 0.00019970940204952316, "loss": 5.6068, "step": 40450 }, { "epoch": 0.17432776201893072, "grad_norm": 1.995444655418396, "learning_rate": 0.00019970866529868725, "loss": 5.2014, "step": 40500 }, { "epoch": 0.17454298147821334, "grad_norm": 1.0267479419708252, "learning_rate": 0.0001997079276164564, "loss": 5.48, "step": 40550 }, { "epoch": 0.17475820093749597, "grad_norm": 0.6800220608711243, "learning_rate": 0.0001997071890028374, "loss": 5.3011, "step": 40600 }, { "epoch": 0.1749734203967786, "grad_norm": 0.2876277267932892, "learning_rate": 0.00019970644945783712, "loss": 5.0807, "step": 40650 }, { "epoch": 0.17518863985606123, "grad_norm": 0.17046943306922913, "learning_rate": 0.00019970570898146257, "loss": 5.789, "step": 40700 }, { "epoch": 0.17540385931534386, "grad_norm": 1.5328351259231567, "learning_rate": 0.0001997049675737206, "loss": 5.9511, "step": 40750 }, { "epoch": 0.17561907877462649, "grad_norm": 1.3365617990493774, "learning_rate": 0.00019970422523461816, "loss": 5.0479, "step": 40800 }, { "epoch": 0.17583429823390911, "grad_norm": 1.8749881982803345, "learning_rate": 0.00019970348196416218, "loss": 5.2656, "step": 40850 }, { "epoch": 0.17604951769319174, "grad_norm": 0.7998882532119751, "learning_rate": 0.0001997027377623596, "loss": 5.8779, "step": 40900 }, { "epoch": 0.17626473715247437, "grad_norm": 0.22467933595180511, "learning_rate": 0.00019970199262921737, "loss": 4.9465, "step": 40950 }, { "epoch": 0.176479956611757, "grad_norm": 1.4040569067001343, "learning_rate": 0.00019970124656474244, "loss": 5.375, "step": 41000 }, { "epoch": 0.17669517607103963, "grad_norm": 2.196059465408325, "learning_rate": 0.0001997004995689418, "loss": 5.5984, "step": 41050 }, { "epoch": 0.17691039553032226, "grad_norm": 2.592918634414673, "learning_rate": 0.00019969975164182245, "loss": 5.4593, "step": 41100 }, { "epoch": 0.1771256149896049, "grad_norm": 0.9951131343841553, "learning_rate": 0.00019969900278339132, "loss": 5.4199, "step": 41150 }, { "epoch": 0.17734083444888754, "grad_norm": 1.212738037109375, "learning_rate": 0.00019969825299365543, "loss": 5.3039, "step": 41200 }, { "epoch": 0.17755605390817017, "grad_norm": 1.112196683883667, "learning_rate": 0.00019969750227262178, "loss": 5.6228, "step": 41250 }, { "epoch": 0.1777712733674528, "grad_norm": 0.5373335480690002, "learning_rate": 0.0001996967506202974, "loss": 4.992, "step": 41300 }, { "epoch": 0.17798649282673543, "grad_norm": 2.2883248329162598, "learning_rate": 0.0001996959980366893, "loss": 5.5074, "step": 41350 }, { "epoch": 0.17820171228601805, "grad_norm": 0.9419134259223938, "learning_rate": 0.0001996952445218045, "loss": 5.5093, "step": 41400 }, { "epoch": 0.17841693174530068, "grad_norm": 1.423128366470337, "learning_rate": 0.00019969449007565005, "loss": 5.5372, "step": 41450 }, { "epoch": 0.1786321512045833, "grad_norm": 1.0389143228530884, "learning_rate": 0.000199693734698233, "loss": 5.7261, "step": 41500 }, { "epoch": 0.17884737066386594, "grad_norm": 1.5058284997940063, "learning_rate": 0.0001996929783895604, "loss": 5.5092, "step": 41550 }, { "epoch": 0.17906259012314857, "grad_norm": 1.0494542121887207, "learning_rate": 0.0001996922211496393, "loss": 5.5492, "step": 41600 }, { "epoch": 0.1792778095824312, "grad_norm": 1.9194459915161133, "learning_rate": 0.00019969146297847684, "loss": 5.8872, "step": 41650 }, { "epoch": 0.17949302904171383, "grad_norm": 1.9738308191299438, "learning_rate": 0.00019969070387608, "loss": 5.0622, "step": 41700 }, { "epoch": 0.17970824850099645, "grad_norm": 0.8716869354248047, "learning_rate": 0.00019968994384245595, "loss": 5.0345, "step": 41750 }, { "epoch": 0.1799234679602791, "grad_norm": 0.7535789012908936, "learning_rate": 0.00019968918287761173, "loss": 5.6642, "step": 41800 }, { "epoch": 0.18013868741956174, "grad_norm": 0.9864673018455505, "learning_rate": 0.00019968842098155452, "loss": 5.2749, "step": 41850 }, { "epoch": 0.18035390687884437, "grad_norm": 0.67588210105896, "learning_rate": 0.0001996876581542914, "loss": 5.5616, "step": 41900 }, { "epoch": 0.180569126338127, "grad_norm": 1.569879174232483, "learning_rate": 0.00019968689439582945, "loss": 5.6887, "step": 41950 }, { "epoch": 0.18078434579740962, "grad_norm": 0.47104722261428833, "learning_rate": 0.0001996861297061759, "loss": 4.7772, "step": 42000 }, { "epoch": 0.18078434579740962, "eval_loss": 5.99991512298584, "eval_runtime": 49.863, "eval_samples_per_second": 12.835, "eval_steps_per_second": 6.418, "eval_tts_loss": 6.117994822545998, "step": 42000 }, { "epoch": 0.18099956525669225, "grad_norm": 1.1956497430801392, "learning_rate": 0.0001996853640853378, "loss": 5.5143, "step": 42050 }, { "epoch": 0.18121478471597488, "grad_norm": 0.2706930935382843, "learning_rate": 0.00019968459753332235, "loss": 5.1779, "step": 42100 }, { "epoch": 0.1814300041752575, "grad_norm": 1.4756226539611816, "learning_rate": 0.0001996838300501367, "loss": 4.9879, "step": 42150 }, { "epoch": 0.18164522363454014, "grad_norm": 1.7535388469696045, "learning_rate": 0.00019968306163578806, "loss": 5.5615, "step": 42200 }, { "epoch": 0.18186044309382277, "grad_norm": 0.2906237542629242, "learning_rate": 0.00019968229229028355, "loss": 5.1855, "step": 42250 }, { "epoch": 0.1820756625531054, "grad_norm": 1.315151572227478, "learning_rate": 0.00019968152201363035, "loss": 6.0387, "step": 42300 }, { "epoch": 0.18229088201238802, "grad_norm": 0.9953668713569641, "learning_rate": 0.0001996807508058357, "loss": 5.7242, "step": 42350 }, { "epoch": 0.18250610147167065, "grad_norm": 0.46408572793006897, "learning_rate": 0.0001996799786669068, "loss": 5.1885, "step": 42400 }, { "epoch": 0.18272132093095328, "grad_norm": 2.1976325511932373, "learning_rate": 0.00019967920559685082, "loss": 5.0799, "step": 42450 }, { "epoch": 0.18293654039023594, "grad_norm": 1.4606099128723145, "learning_rate": 0.00019967843159567503, "loss": 5.3106, "step": 42500 }, { "epoch": 0.18315175984951856, "grad_norm": 0.8943641781806946, "learning_rate": 0.00019967765666338664, "loss": 5.29, "step": 42550 }, { "epoch": 0.1833669793088012, "grad_norm": 1.5275267362594604, "learning_rate": 0.0001996768807999929, "loss": 5.1727, "step": 42600 }, { "epoch": 0.18358219876808382, "grad_norm": 1.5389777421951294, "learning_rate": 0.00019967610400550102, "loss": 5.3998, "step": 42650 }, { "epoch": 0.18379741822736645, "grad_norm": 0.8801236748695374, "learning_rate": 0.0001996753262799183, "loss": 5.5634, "step": 42700 }, { "epoch": 0.18401263768664908, "grad_norm": 2.4873547554016113, "learning_rate": 0.000199674547623252, "loss": 5.5462, "step": 42750 }, { "epoch": 0.1842278571459317, "grad_norm": 1.4518872499465942, "learning_rate": 0.00019967376803550933, "loss": 5.4495, "step": 42800 }, { "epoch": 0.18444307660521433, "grad_norm": 1.1900620460510254, "learning_rate": 0.00019967298751669767, "loss": 5.1853, "step": 42850 }, { "epoch": 0.18465829606449696, "grad_norm": 1.0711162090301514, "learning_rate": 0.00019967220606682427, "loss": 5.5112, "step": 42900 }, { "epoch": 0.1848735155237796, "grad_norm": 0.9827768802642822, "learning_rate": 0.0001996714236858964, "loss": 5.2163, "step": 42950 }, { "epoch": 0.18508873498306222, "grad_norm": 0.39113849401474, "learning_rate": 0.0001996706403739214, "loss": 5.552, "step": 43000 }, { "epoch": 0.18530395444234485, "grad_norm": 0.5762609243392944, "learning_rate": 0.0001996698561309066, "loss": 5.1091, "step": 43050 }, { "epoch": 0.18551917390162748, "grad_norm": 0.932518720626831, "learning_rate": 0.00019966907095685926, "loss": 5.6003, "step": 43100 }, { "epoch": 0.18573439336091013, "grad_norm": 1.988603949546814, "learning_rate": 0.00019966828485178678, "loss": 5.5967, "step": 43150 }, { "epoch": 0.18594961282019276, "grad_norm": 1.8209559917449951, "learning_rate": 0.00019966749781569649, "loss": 5.3508, "step": 43200 }, { "epoch": 0.1861648322794754, "grad_norm": 1.392592191696167, "learning_rate": 0.0001996667098485957, "loss": 5.3206, "step": 43250 }, { "epoch": 0.18638005173875802, "grad_norm": 1.932367205619812, "learning_rate": 0.00019966592095049186, "loss": 5.8113, "step": 43300 }, { "epoch": 0.18659527119804065, "grad_norm": 0.3724992871284485, "learning_rate": 0.00019966513112139227, "loss": 5.2094, "step": 43350 }, { "epoch": 0.18681049065732327, "grad_norm": 1.0303024053573608, "learning_rate": 0.0001996643403613043, "loss": 5.1263, "step": 43400 }, { "epoch": 0.1870257101166059, "grad_norm": 0.3853681981563568, "learning_rate": 0.00019966354867023537, "loss": 5.7155, "step": 43450 }, { "epoch": 0.18724092957588853, "grad_norm": 0.26003509759902954, "learning_rate": 0.00019966275604819286, "loss": 5.1961, "step": 43500 }, { "epoch": 0.18745614903517116, "grad_norm": 0.2785758674144745, "learning_rate": 0.00019966196249518418, "loss": 5.0258, "step": 43550 }, { "epoch": 0.1876713684944538, "grad_norm": 1.2389429807662964, "learning_rate": 0.00019966116801121677, "loss": 5.3705, "step": 43600 }, { "epoch": 0.18788658795373642, "grad_norm": 1.5231176614761353, "learning_rate": 0.000199660372596298, "loss": 5.8762, "step": 43650 }, { "epoch": 0.18810180741301905, "grad_norm": 1.7266250848770142, "learning_rate": 0.0001996595762504353, "loss": 4.9102, "step": 43700 }, { "epoch": 0.18831702687230167, "grad_norm": 1.4643664360046387, "learning_rate": 0.00019965877897363616, "loss": 5.7471, "step": 43750 }, { "epoch": 0.18853224633158433, "grad_norm": 1.1804057359695435, "learning_rate": 0.00019965798076590797, "loss": 5.1863, "step": 43800 }, { "epoch": 0.18874746579086696, "grad_norm": 1.2418252229690552, "learning_rate": 0.00019965718162725824, "loss": 5.5194, "step": 43850 }, { "epoch": 0.1889626852501496, "grad_norm": 2.8593976497650146, "learning_rate": 0.0001996563815576944, "loss": 5.5533, "step": 43900 }, { "epoch": 0.18917790470943222, "grad_norm": 2.09224796295166, "learning_rate": 0.00019965558055722396, "loss": 5.8581, "step": 43950 }, { "epoch": 0.18939312416871484, "grad_norm": 1.4324254989624023, "learning_rate": 0.00019965477862585434, "loss": 5.1466, "step": 44000 }, { "epoch": 0.18960834362799747, "grad_norm": 1.3727551698684692, "learning_rate": 0.00019965397576359307, "loss": 5.9155, "step": 44050 }, { "epoch": 0.1898235630872801, "grad_norm": 1.2977691888809204, "learning_rate": 0.00019965317197044767, "loss": 5.9675, "step": 44100 }, { "epoch": 0.19003878254656273, "grad_norm": 1.0278069972991943, "learning_rate": 0.0001996523672464256, "loss": 5.6291, "step": 44150 }, { "epoch": 0.19025400200584536, "grad_norm": 1.2213515043258667, "learning_rate": 0.00019965156159153441, "loss": 5.9392, "step": 44200 }, { "epoch": 0.19046922146512799, "grad_norm": 1.6242117881774902, "learning_rate": 0.0001996507550057816, "loss": 5.9912, "step": 44250 }, { "epoch": 0.19068444092441061, "grad_norm": 0.6379998326301575, "learning_rate": 0.00019964994748917475, "loss": 5.4898, "step": 44300 }, { "epoch": 0.19089966038369324, "grad_norm": 0.8886232972145081, "learning_rate": 0.00019964913904172137, "loss": 5.2386, "step": 44350 }, { "epoch": 0.19111487984297587, "grad_norm": 1.1820510625839233, "learning_rate": 0.000199648329663429, "loss": 5.5016, "step": 44400 }, { "epoch": 0.1913300993022585, "grad_norm": 1.0135712623596191, "learning_rate": 0.00019964751935430526, "loss": 5.4797, "step": 44450 }, { "epoch": 0.19154531876154116, "grad_norm": 0.8176568746566772, "learning_rate": 0.00019964670811435764, "loss": 6.0355, "step": 44500 }, { "epoch": 0.19176053822082378, "grad_norm": 1.5005437135696411, "learning_rate": 0.00019964589594359376, "loss": 5.6839, "step": 44550 }, { "epoch": 0.1919757576801064, "grad_norm": 1.17890465259552, "learning_rate": 0.0001996450828420212, "loss": 5.169, "step": 44600 }, { "epoch": 0.19219097713938904, "grad_norm": 1.320347785949707, "learning_rate": 0.00019964426880964754, "loss": 5.5436, "step": 44650 }, { "epoch": 0.19240619659867167, "grad_norm": 0.8962569832801819, "learning_rate": 0.00019964345384648042, "loss": 5.425, "step": 44700 }, { "epoch": 0.1926214160579543, "grad_norm": 0.5257356762886047, "learning_rate": 0.00019964263795252746, "loss": 5.4607, "step": 44750 }, { "epoch": 0.19283663551723693, "grad_norm": 1.59420645236969, "learning_rate": 0.0001996418211277962, "loss": 5.5242, "step": 44800 }, { "epoch": 0.19305185497651955, "grad_norm": 1.7570161819458008, "learning_rate": 0.00019964100337229435, "loss": 5.5545, "step": 44850 }, { "epoch": 0.19326707443580218, "grad_norm": 1.401421070098877, "learning_rate": 0.00019964018468602955, "loss": 5.2291, "step": 44900 }, { "epoch": 0.1934822938950848, "grad_norm": 0.9453854560852051, "learning_rate": 0.00019963936506900938, "loss": 5.8079, "step": 44950 }, { "epoch": 0.19369751335436744, "grad_norm": 1.9389640092849731, "learning_rate": 0.00019963854452124155, "loss": 5.4014, "step": 45000 }, { "epoch": 0.19369751335436744, "eval_loss": 5.988659858703613, "eval_runtime": 49.7399, "eval_samples_per_second": 12.867, "eval_steps_per_second": 6.433, "eval_tts_loss": 6.05616758019919, "step": 45000 }, { "epoch": 0.19391273281365007, "grad_norm": 2.40205717086792, "learning_rate": 0.00019963772304273372, "loss": 5.4876, "step": 45050 }, { "epoch": 0.1941279522729327, "grad_norm": 0.3262667953968048, "learning_rate": 0.00019963690063349354, "loss": 5.6392, "step": 45100 }, { "epoch": 0.19434317173221535, "grad_norm": 0.9905305504798889, "learning_rate": 0.00019963607729352873, "loss": 5.6836, "step": 45150 }, { "epoch": 0.19455839119149798, "grad_norm": 0.9432397484779358, "learning_rate": 0.00019963525302284696, "loss": 5.4935, "step": 45200 }, { "epoch": 0.1947736106507806, "grad_norm": 1.7752591371536255, "learning_rate": 0.0001996344278214559, "loss": 5.1419, "step": 45250 }, { "epoch": 0.19498883011006324, "grad_norm": 1.5263196229934692, "learning_rate": 0.00019963360168936332, "loss": 5.3798, "step": 45300 }, { "epoch": 0.19520404956934587, "grad_norm": 0.8461008071899414, "learning_rate": 0.00019963277462657687, "loss": 5.5759, "step": 45350 }, { "epoch": 0.1954192690286285, "grad_norm": 0.8464626669883728, "learning_rate": 0.00019963194663310434, "loss": 5.2198, "step": 45400 }, { "epoch": 0.19563448848791112, "grad_norm": 1.563346266746521, "learning_rate": 0.00019963111770895342, "loss": 5.6367, "step": 45450 }, { "epoch": 0.19584970794719375, "grad_norm": 1.188881278038025, "learning_rate": 0.00019963028785413185, "loss": 6.1365, "step": 45500 }, { "epoch": 0.19606492740647638, "grad_norm": 1.437353491783142, "learning_rate": 0.00019962945706864744, "loss": 5.2012, "step": 45550 }, { "epoch": 0.196280146865759, "grad_norm": 1.0964992046356201, "learning_rate": 0.00019962862535250788, "loss": 5.4786, "step": 45600 }, { "epoch": 0.19649536632504164, "grad_norm": 1.580206036567688, "learning_rate": 0.00019962779270572096, "loss": 5.3922, "step": 45650 }, { "epoch": 0.19671058578432427, "grad_norm": 0.7344819903373718, "learning_rate": 0.00019962695912829448, "loss": 5.6195, "step": 45700 }, { "epoch": 0.1969258052436069, "grad_norm": 1.0382460355758667, "learning_rate": 0.00019962612462023623, "loss": 5.3143, "step": 45750 }, { "epoch": 0.19714102470288955, "grad_norm": 0.851793110370636, "learning_rate": 0.00019962528918155396, "loss": 5.2774, "step": 45800 }, { "epoch": 0.19735624416217218, "grad_norm": 1.719031572341919, "learning_rate": 0.0001996244528122555, "loss": 5.3567, "step": 45850 }, { "epoch": 0.1975714636214548, "grad_norm": 1.21913743019104, "learning_rate": 0.0001996236155123487, "loss": 4.9464, "step": 45900 }, { "epoch": 0.19778668308073744, "grad_norm": 1.6324721574783325, "learning_rate": 0.00019962277728184132, "loss": 5.147, "step": 45950 }, { "epoch": 0.19800190254002006, "grad_norm": 1.109820008277893, "learning_rate": 0.00019962193812074122, "loss": 5.6524, "step": 46000 }, { "epoch": 0.1982171219993027, "grad_norm": 1.3571628332138062, "learning_rate": 0.00019962109802905623, "loss": 5.5626, "step": 46050 }, { "epoch": 0.19843234145858532, "grad_norm": 0.9243341684341431, "learning_rate": 0.0001996202570067942, "loss": 5.0361, "step": 46100 }, { "epoch": 0.19864756091786795, "grad_norm": 1.4373767375946045, "learning_rate": 0.00019961941505396301, "loss": 5.2323, "step": 46150 }, { "epoch": 0.19886278037715058, "grad_norm": 0.8940660953521729, "learning_rate": 0.00019961857217057048, "loss": 5.4181, "step": 46200 }, { "epoch": 0.1990779998364332, "grad_norm": 1.1559289693832397, "learning_rate": 0.00019961772835662452, "loss": 5.5053, "step": 46250 }, { "epoch": 0.19929321929571583, "grad_norm": 0.45639899373054504, "learning_rate": 0.00019961688361213301, "loss": 5.3154, "step": 46300 }, { "epoch": 0.19950843875499846, "grad_norm": 2.2042691707611084, "learning_rate": 0.00019961603793710385, "loss": 5.3136, "step": 46350 }, { "epoch": 0.1997236582142811, "grad_norm": 0.9833787083625793, "learning_rate": 0.00019961519133154484, "loss": 5.3346, "step": 46400 }, { "epoch": 0.19993887767356372, "grad_norm": 1.2540464401245117, "learning_rate": 0.00019961434379546405, "loss": 5.3664, "step": 46450 }, { "epoch": 0.20015409713284638, "grad_norm": 1.7903770208358765, "learning_rate": 0.00019961349532886927, "loss": 5.7277, "step": 46500 }, { "epoch": 0.200369316592129, "grad_norm": 1.3091398477554321, "learning_rate": 0.00019961264593176847, "loss": 5.9054, "step": 46550 }, { "epoch": 0.20058453605141163, "grad_norm": 2.2425472736358643, "learning_rate": 0.0001996117956041696, "loss": 5.3707, "step": 46600 }, { "epoch": 0.20079975551069426, "grad_norm": 1.6382770538330078, "learning_rate": 0.0001996109443460806, "loss": 5.2643, "step": 46650 }, { "epoch": 0.2010149749699769, "grad_norm": 1.0935531854629517, "learning_rate": 0.00019961009215750937, "loss": 5.5434, "step": 46700 }, { "epoch": 0.20123019442925952, "grad_norm": 1.265020489692688, "learning_rate": 0.00019960923903846397, "loss": 5.3866, "step": 46750 }, { "epoch": 0.20144541388854215, "grad_norm": 2.747368335723877, "learning_rate": 0.00019960838498895225, "loss": 5.3316, "step": 46800 }, { "epoch": 0.20166063334782477, "grad_norm": 0.958810567855835, "learning_rate": 0.00019960753000898226, "loss": 5.4604, "step": 46850 }, { "epoch": 0.2018758528071074, "grad_norm": 1.346915602684021, "learning_rate": 0.000199606674098562, "loss": 5.3334, "step": 46900 }, { "epoch": 0.20209107226639003, "grad_norm": 1.083828091621399, "learning_rate": 0.00019960581725769942, "loss": 5.748, "step": 46950 }, { "epoch": 0.20230629172567266, "grad_norm": 2.008018970489502, "learning_rate": 0.00019960495948640256, "loss": 5.5248, "step": 47000 }, { "epoch": 0.2025215111849553, "grad_norm": 0.9811460375785828, "learning_rate": 0.00019960410078467938, "loss": 4.6768, "step": 47050 }, { "epoch": 0.20273673064423792, "grad_norm": 0.8748836517333984, "learning_rate": 0.00019960324115253795, "loss": 5.4306, "step": 47100 }, { "epoch": 0.20295195010352057, "grad_norm": 1.8672735691070557, "learning_rate": 0.00019960238058998632, "loss": 5.3091, "step": 47150 }, { "epoch": 0.2031671695628032, "grad_norm": 0.7682002186775208, "learning_rate": 0.0001996015190970325, "loss": 5.5424, "step": 47200 }, { "epoch": 0.20338238902208583, "grad_norm": 1.3956269025802612, "learning_rate": 0.00019960065667368445, "loss": 5.6034, "step": 47250 }, { "epoch": 0.20359760848136846, "grad_norm": 0.9481976628303528, "learning_rate": 0.00019959979331995038, "loss": 5.4465, "step": 47300 }, { "epoch": 0.2038128279406511, "grad_norm": 0.8210592865943909, "learning_rate": 0.00019959892903583825, "loss": 5.7286, "step": 47350 }, { "epoch": 0.20402804739993372, "grad_norm": 0.8981212973594666, "learning_rate": 0.0001995980638213562, "loss": 6.0161, "step": 47400 }, { "epoch": 0.20424326685921634, "grad_norm": 0.790116012096405, "learning_rate": 0.00019959719767651226, "loss": 5.2202, "step": 47450 }, { "epoch": 0.20445848631849897, "grad_norm": 0.7612621188163757, "learning_rate": 0.00019959633060131452, "loss": 5.4606, "step": 47500 }, { "epoch": 0.2046737057777816, "grad_norm": 1.2075308561325073, "learning_rate": 0.00019959546259577113, "loss": 5.4151, "step": 47550 }, { "epoch": 0.20488892523706423, "grad_norm": 1.446319580078125, "learning_rate": 0.00019959459365989017, "loss": 5.4488, "step": 47600 }, { "epoch": 0.20510414469634686, "grad_norm": 1.5639781951904297, "learning_rate": 0.00019959372379367975, "loss": 5.5585, "step": 47650 }, { "epoch": 0.20531936415562949, "grad_norm": 0.5640411972999573, "learning_rate": 0.000199592852997148, "loss": 5.5703, "step": 47700 }, { "epoch": 0.20553458361491211, "grad_norm": 1.2129855155944824, "learning_rate": 0.00019959198127030302, "loss": 5.3172, "step": 47750 }, { "epoch": 0.20574980307419474, "grad_norm": 1.2998452186584473, "learning_rate": 0.00019959110861315302, "loss": 5.5114, "step": 47800 }, { "epoch": 0.2059650225334774, "grad_norm": 2.560208559036255, "learning_rate": 0.0001995902350257061, "loss": 5.3824, "step": 47850 }, { "epoch": 0.20618024199276003, "grad_norm": 1.2330033779144287, "learning_rate": 0.00019958936050797047, "loss": 5.7663, "step": 47900 }, { "epoch": 0.20639546145204266, "grad_norm": 1.167486548423767, "learning_rate": 0.00019958848505995423, "loss": 5.5009, "step": 47950 }, { "epoch": 0.20661068091132528, "grad_norm": 2.0923728942871094, "learning_rate": 0.00019958760868166563, "loss": 5.7296, "step": 48000 }, { "epoch": 0.20661068091132528, "eval_loss": 5.96932315826416, "eval_runtime": 49.7182, "eval_samples_per_second": 12.873, "eval_steps_per_second": 6.436, "eval_tts_loss": 6.0562110110587275, "step": 48000 }, { "epoch": 0.2068259003706079, "grad_norm": 1.0311095714569092, "learning_rate": 0.00019958673137311278, "loss": 5.1091, "step": 48050 }, { "epoch": 0.20704111982989054, "grad_norm": 1.3319623470306396, "learning_rate": 0.00019958585313430396, "loss": 5.2096, "step": 48100 }, { "epoch": 0.20725633928917317, "grad_norm": 0.27855992317199707, "learning_rate": 0.0001995849739652473, "loss": 5.0008, "step": 48150 }, { "epoch": 0.2074715587484558, "grad_norm": 0.2503814101219177, "learning_rate": 0.00019958409386595106, "loss": 5.6102, "step": 48200 }, { "epoch": 0.20768677820773843, "grad_norm": 0.2247920036315918, "learning_rate": 0.00019958321283642346, "loss": 5.1267, "step": 48250 }, { "epoch": 0.20790199766702105, "grad_norm": 1.2833117246627808, "learning_rate": 0.00019958233087667267, "loss": 6.0286, "step": 48300 }, { "epoch": 0.20811721712630368, "grad_norm": 1.5172131061553955, "learning_rate": 0.000199581447986707, "loss": 5.7667, "step": 48350 }, { "epoch": 0.2083324365855863, "grad_norm": 1.4411728382110596, "learning_rate": 0.00019958056416653468, "loss": 5.488, "step": 48400 }, { "epoch": 0.20854765604486894, "grad_norm": 1.2462176084518433, "learning_rate": 0.00019957967941616395, "loss": 5.7207, "step": 48450 }, { "epoch": 0.2087628755041516, "grad_norm": 0.5906105041503906, "learning_rate": 0.00019957879373560307, "loss": 5.3533, "step": 48500 }, { "epoch": 0.20897809496343422, "grad_norm": 0.463609904050827, "learning_rate": 0.00019957790712486033, "loss": 5.166, "step": 48550 }, { "epoch": 0.20919331442271685, "grad_norm": 0.9689344167709351, "learning_rate": 0.00019957701958394402, "loss": 5.3691, "step": 48600 }, { "epoch": 0.20940853388199948, "grad_norm": 0.9464945793151855, "learning_rate": 0.00019957613111286238, "loss": 5.7203, "step": 48650 }, { "epoch": 0.2096237533412821, "grad_norm": 0.3608854413032532, "learning_rate": 0.00019957524171162377, "loss": 5.4879, "step": 48700 }, { "epoch": 0.20983897280056474, "grad_norm": 1.4249147176742554, "learning_rate": 0.00019957435138023648, "loss": 5.4722, "step": 48750 }, { "epoch": 0.21005419225984737, "grad_norm": 2.8934414386749268, "learning_rate": 0.0001995734601187088, "loss": 5.3677, "step": 48800 }, { "epoch": 0.21026941171913, "grad_norm": 0.9962794780731201, "learning_rate": 0.00019957256792704908, "loss": 5.779, "step": 48850 }, { "epoch": 0.21048463117841262, "grad_norm": 1.0369484424591064, "learning_rate": 0.00019957167480526565, "loss": 5.2781, "step": 48900 }, { "epoch": 0.21069985063769525, "grad_norm": 1.090420126914978, "learning_rate": 0.0001995707807533669, "loss": 5.3448, "step": 48950 }, { "epoch": 0.21091507009697788, "grad_norm": 1.6538866758346558, "learning_rate": 0.0001995698857713611, "loss": 5.297, "step": 49000 }, { "epoch": 0.2111302895562605, "grad_norm": 0.9921044111251831, "learning_rate": 0.00019956898985925664, "loss": 4.9217, "step": 49050 }, { "epoch": 0.21134550901554314, "grad_norm": 0.9721300005912781, "learning_rate": 0.00019956809301706186, "loss": 5.8114, "step": 49100 }, { "epoch": 0.2115607284748258, "grad_norm": 1.087932825088501, "learning_rate": 0.00019956719524478523, "loss": 5.7786, "step": 49150 }, { "epoch": 0.21177594793410842, "grad_norm": 1.130364179611206, "learning_rate": 0.00019956629654243504, "loss": 5.5718, "step": 49200 }, { "epoch": 0.21199116739339105, "grad_norm": 1.0457923412322998, "learning_rate": 0.00019956539691001973, "loss": 5.411, "step": 49250 }, { "epoch": 0.21220638685267368, "grad_norm": 1.2125331163406372, "learning_rate": 0.0001995644963475477, "loss": 5.5968, "step": 49300 }, { "epoch": 0.2124216063119563, "grad_norm": 0.830668032169342, "learning_rate": 0.00019956359485502736, "loss": 5.6159, "step": 49350 }, { "epoch": 0.21263682577123894, "grad_norm": 1.0496330261230469, "learning_rate": 0.00019956269243246713, "loss": 5.1129, "step": 49400 }, { "epoch": 0.21285204523052156, "grad_norm": 1.529306411743164, "learning_rate": 0.00019956178907987545, "loss": 5.5323, "step": 49450 }, { "epoch": 0.2130672646898042, "grad_norm": 1.6682978868484497, "learning_rate": 0.0001995608847972607, "loss": 5.34, "step": 49500 }, { "epoch": 0.21328248414908682, "grad_norm": 0.898091733455658, "learning_rate": 0.0001995599795846314, "loss": 5.7537, "step": 49550 }, { "epoch": 0.21349770360836945, "grad_norm": 1.6104379892349243, "learning_rate": 0.00019955907344199603, "loss": 5.3177, "step": 49600 }, { "epoch": 0.21371292306765208, "grad_norm": 1.0056496858596802, "learning_rate": 0.00019955816636936296, "loss": 5.6207, "step": 49650 }, { "epoch": 0.2139281425269347, "grad_norm": 1.0176551342010498, "learning_rate": 0.0001995572583667407, "loss": 5.299, "step": 49700 }, { "epoch": 0.21414336198621733, "grad_norm": 0.2221224457025528, "learning_rate": 0.00019955634943413775, "loss": 5.4685, "step": 49750 }, { "epoch": 0.21435858144549996, "grad_norm": 1.6749045848846436, "learning_rate": 0.0001995554395715626, "loss": 5.6135, "step": 49800 }, { "epoch": 0.21457380090478262, "grad_norm": 0.6849638223648071, "learning_rate": 0.00019955452877902372, "loss": 5.1761, "step": 49850 }, { "epoch": 0.21478902036406525, "grad_norm": 1.143099308013916, "learning_rate": 0.00019955361705652967, "loss": 5.0304, "step": 49900 }, { "epoch": 0.21500423982334788, "grad_norm": 0.9740395545959473, "learning_rate": 0.00019955270440408888, "loss": 5.5592, "step": 49950 }, { "epoch": 0.2152194592826305, "grad_norm": 1.3359935283660889, "learning_rate": 0.00019955179082170997, "loss": 5.2623, "step": 50000 }, { "epoch": 0.21543467874191313, "grad_norm": 1.1677852869033813, "learning_rate": 0.00019955087630940143, "loss": 5.1455, "step": 50050 }, { "epoch": 0.21564989820119576, "grad_norm": 0.9470292329788208, "learning_rate": 0.00019954996086717178, "loss": 5.7323, "step": 50100 }, { "epoch": 0.2158651176604784, "grad_norm": 0.6113887429237366, "learning_rate": 0.0001995490444950296, "loss": 5.1541, "step": 50150 }, { "epoch": 0.21608033711976102, "grad_norm": 0.6079469323158264, "learning_rate": 0.00019954812719298346, "loss": 5.3356, "step": 50200 }, { "epoch": 0.21629555657904365, "grad_norm": 1.449463963508606, "learning_rate": 0.00019954720896104192, "loss": 5.4365, "step": 50250 }, { "epoch": 0.21651077603832627, "grad_norm": 0.7095744013786316, "learning_rate": 0.00019954628979921356, "loss": 5.1376, "step": 50300 }, { "epoch": 0.2167259954976089, "grad_norm": 1.6120834350585938, "learning_rate": 0.00019954536970750693, "loss": 5.2021, "step": 50350 }, { "epoch": 0.21694121495689153, "grad_norm": 1.8194729089736938, "learning_rate": 0.00019954444868593068, "loss": 5.5118, "step": 50400 }, { "epoch": 0.21715643441617416, "grad_norm": 1.4188915491104126, "learning_rate": 0.00019954352673449337, "loss": 5.062, "step": 50450 }, { "epoch": 0.21737165387545682, "grad_norm": 1.2952605485916138, "learning_rate": 0.00019954260385320361, "loss": 5.9129, "step": 50500 }, { "epoch": 0.21758687333473944, "grad_norm": 1.6226215362548828, "learning_rate": 0.00019954168004207007, "loss": 5.4048, "step": 50550 }, { "epoch": 0.21780209279402207, "grad_norm": 1.0096547603607178, "learning_rate": 0.00019954075530110137, "loss": 5.7535, "step": 50600 }, { "epoch": 0.2180173122533047, "grad_norm": 0.655689537525177, "learning_rate": 0.0001995398296303061, "loss": 5.1051, "step": 50650 }, { "epoch": 0.21823253171258733, "grad_norm": 1.1218249797821045, "learning_rate": 0.0001995389030296929, "loss": 5.7495, "step": 50700 }, { "epoch": 0.21844775117186996, "grad_norm": 0.9273683428764343, "learning_rate": 0.0001995379754992705, "loss": 5.6025, "step": 50750 }, { "epoch": 0.2186629706311526, "grad_norm": 1.0130168199539185, "learning_rate": 0.00019953704703904753, "loss": 5.4971, "step": 50800 }, { "epoch": 0.21887819009043522, "grad_norm": 0.6400539875030518, "learning_rate": 0.00019953611764903263, "loss": 5.3718, "step": 50850 }, { "epoch": 0.21909340954971784, "grad_norm": 0.8558303713798523, "learning_rate": 0.00019953518732923454, "loss": 5.2819, "step": 50900 }, { "epoch": 0.21930862900900047, "grad_norm": 1.340484619140625, "learning_rate": 0.0001995342560796619, "loss": 5.4645, "step": 50950 }, { "epoch": 0.2195238484682831, "grad_norm": 1.519740104675293, "learning_rate": 0.00019953332390032342, "loss": 5.4591, "step": 51000 }, { "epoch": 0.2195238484682831, "eval_loss": 5.944695949554443, "eval_runtime": 49.9267, "eval_samples_per_second": 12.819, "eval_steps_per_second": 6.409, "eval_tts_loss": 6.116118344604461, "step": 51000 }, { "epoch": 0.21973906792756573, "grad_norm": 1.5892105102539062, "learning_rate": 0.00019953239079122784, "loss": 5.0938, "step": 51050 }, { "epoch": 0.21995428738684836, "grad_norm": 1.673384189605713, "learning_rate": 0.0001995314567523838, "loss": 5.2197, "step": 51100 }, { "epoch": 0.220169506846131, "grad_norm": 0.24516557157039642, "learning_rate": 0.00019953052178380012, "loss": 5.1722, "step": 51150 }, { "epoch": 0.22038472630541364, "grad_norm": 0.9135076403617859, "learning_rate": 0.0001995295858854855, "loss": 5.7343, "step": 51200 }, { "epoch": 0.22059994576469627, "grad_norm": 0.9753496646881104, "learning_rate": 0.00019952864905744863, "loss": 5.158, "step": 51250 }, { "epoch": 0.2208151652239789, "grad_norm": 1.0558239221572876, "learning_rate": 0.00019952771129969833, "loss": 5.1998, "step": 51300 }, { "epoch": 0.22103038468326153, "grad_norm": 2.4674112796783447, "learning_rate": 0.00019952677261224333, "loss": 5.2789, "step": 51350 }, { "epoch": 0.22124560414254416, "grad_norm": 1.2905189990997314, "learning_rate": 0.0001995258329950924, "loss": 5.6504, "step": 51400 }, { "epoch": 0.22146082360182678, "grad_norm": 0.7170586585998535, "learning_rate": 0.00019952489244825432, "loss": 5.4908, "step": 51450 }, { "epoch": 0.2216760430611094, "grad_norm": 1.2904372215270996, "learning_rate": 0.0001995239509717379, "loss": 6.1217, "step": 51500 }, { "epoch": 0.22189126252039204, "grad_norm": 1.2830251455307007, "learning_rate": 0.00019952300856555187, "loss": 5.3131, "step": 51550 }, { "epoch": 0.22210648197967467, "grad_norm": 2.3233048915863037, "learning_rate": 0.0001995220652297051, "loss": 5.0945, "step": 51600 }, { "epoch": 0.2223217014389573, "grad_norm": 2.1694326400756836, "learning_rate": 0.00019952112096420632, "loss": 5.4376, "step": 51650 }, { "epoch": 0.22253692089823993, "grad_norm": 0.8337290287017822, "learning_rate": 0.00019952017576906447, "loss": 5.5018, "step": 51700 }, { "epoch": 0.22275214035752255, "grad_norm": 2.4260177612304688, "learning_rate": 0.00019951922964428828, "loss": 5.3398, "step": 51750 }, { "epoch": 0.22296735981680518, "grad_norm": 1.3624546527862549, "learning_rate": 0.00019951828258988664, "loss": 5.0038, "step": 51800 }, { "epoch": 0.22318257927608784, "grad_norm": 2.266918897628784, "learning_rate": 0.00019951733460586837, "loss": 5.6182, "step": 51850 }, { "epoch": 0.22339779873537047, "grad_norm": 0.9978936314582825, "learning_rate": 0.0001995163856922423, "loss": 5.3711, "step": 51900 }, { "epoch": 0.2236130181946531, "grad_norm": 0.5029309391975403, "learning_rate": 0.0001995154358490174, "loss": 5.2841, "step": 51950 }, { "epoch": 0.22382823765393572, "grad_norm": 0.26915180683135986, "learning_rate": 0.00019951448507620243, "loss": 5.5376, "step": 52000 }, { "epoch": 0.22404345711321835, "grad_norm": 1.940232276916504, "learning_rate": 0.0001995135333738063, "loss": 5.08, "step": 52050 }, { "epoch": 0.22425867657250098, "grad_norm": 1.1586575508117676, "learning_rate": 0.00019951258074183792, "loss": 5.3184, "step": 52100 }, { "epoch": 0.2244738960317836, "grad_norm": 1.1043246984481812, "learning_rate": 0.00019951162718030618, "loss": 5.3389, "step": 52150 }, { "epoch": 0.22468911549106624, "grad_norm": 1.0309019088745117, "learning_rate": 0.00019951067268922002, "loss": 5.491, "step": 52200 }, { "epoch": 0.22490433495034887, "grad_norm": 0.8068641424179077, "learning_rate": 0.0001995097172685883, "loss": 5.6311, "step": 52250 }, { "epoch": 0.2251195544096315, "grad_norm": 1.0859923362731934, "learning_rate": 0.00019950876091841995, "loss": 5.9809, "step": 52300 }, { "epoch": 0.22533477386891412, "grad_norm": 1.0469591617584229, "learning_rate": 0.00019950780363872395, "loss": 5.7609, "step": 52350 }, { "epoch": 0.22554999332819675, "grad_norm": 1.0656195878982544, "learning_rate": 0.00019950684542950918, "loss": 4.7094, "step": 52400 }, { "epoch": 0.22576521278747938, "grad_norm": 1.375428318977356, "learning_rate": 0.00019950588629078463, "loss": 5.2725, "step": 52450 }, { "epoch": 0.22598043224676204, "grad_norm": 1.3102055788040161, "learning_rate": 0.0001995049262225593, "loss": 5.4271, "step": 52500 }, { "epoch": 0.22619565170604466, "grad_norm": 2.097334623336792, "learning_rate": 0.0001995039652248421, "loss": 4.8636, "step": 52550 }, { "epoch": 0.2264108711653273, "grad_norm": 1.5023698806762695, "learning_rate": 0.000199503003297642, "loss": 5.8363, "step": 52600 }, { "epoch": 0.22662609062460992, "grad_norm": 0.5496647953987122, "learning_rate": 0.00019950204044096799, "loss": 5.0711, "step": 52650 }, { "epoch": 0.22684131008389255, "grad_norm": 0.6920158863067627, "learning_rate": 0.0001995010766548291, "loss": 5.224, "step": 52700 }, { "epoch": 0.22705652954317518, "grad_norm": 2.1742353439331055, "learning_rate": 0.0001995001119392343, "loss": 5.4006, "step": 52750 }, { "epoch": 0.2272717490024578, "grad_norm": 1.4097658395767212, "learning_rate": 0.00019949914629419264, "loss": 5.5332, "step": 52800 }, { "epoch": 0.22748696846174044, "grad_norm": 0.8111319541931152, "learning_rate": 0.00019949817971971306, "loss": 5.5721, "step": 52850 }, { "epoch": 0.22770218792102306, "grad_norm": 1.195981502532959, "learning_rate": 0.00019949721221580472, "loss": 5.1212, "step": 52900 }, { "epoch": 0.2279174073803057, "grad_norm": 1.57807195186615, "learning_rate": 0.0001994962437824765, "loss": 5.3336, "step": 52950 }, { "epoch": 0.22813262683958832, "grad_norm": 2.529475688934326, "learning_rate": 0.00019949527441973758, "loss": 5.8824, "step": 53000 }, { "epoch": 0.22834784629887095, "grad_norm": 1.00227952003479, "learning_rate": 0.00019949430412759695, "loss": 5.0921, "step": 53050 }, { "epoch": 0.22856306575815358, "grad_norm": 1.3835440874099731, "learning_rate": 0.00019949333290606366, "loss": 5.1773, "step": 53100 }, { "epoch": 0.22877828521743623, "grad_norm": 1.1759252548217773, "learning_rate": 0.00019949236075514685, "loss": 5.3621, "step": 53150 }, { "epoch": 0.22899350467671886, "grad_norm": 1.039719581604004, "learning_rate": 0.00019949138767485553, "loss": 5.6005, "step": 53200 }, { "epoch": 0.2292087241360015, "grad_norm": 0.5613513588905334, "learning_rate": 0.00019949041366519886, "loss": 5.4723, "step": 53250 }, { "epoch": 0.22942394359528412, "grad_norm": 0.9384167790412903, "learning_rate": 0.00019948943872618585, "loss": 5.6301, "step": 53300 }, { "epoch": 0.22963916305456675, "grad_norm": 1.963740587234497, "learning_rate": 0.00019948846285782566, "loss": 5.226, "step": 53350 }, { "epoch": 0.22985438251384938, "grad_norm": 1.3246345520019531, "learning_rate": 0.0001994874860601274, "loss": 5.6125, "step": 53400 }, { "epoch": 0.230069601973132, "grad_norm": 1.0553481578826904, "learning_rate": 0.00019948650833310022, "loss": 5.871, "step": 53450 }, { "epoch": 0.23028482143241463, "grad_norm": 0.8935693502426147, "learning_rate": 0.00019948552967675317, "loss": 5.5494, "step": 53500 }, { "epoch": 0.23050004089169726, "grad_norm": 2.04744029045105, "learning_rate": 0.0001994845500910955, "loss": 5.2225, "step": 53550 }, { "epoch": 0.2307152603509799, "grad_norm": 3.009582757949829, "learning_rate": 0.0001994835695761363, "loss": 5.7404, "step": 53600 }, { "epoch": 0.23093047981026252, "grad_norm": 0.6760677099227905, "learning_rate": 0.00019948258813188473, "loss": 5.8331, "step": 53650 }, { "epoch": 0.23114569926954515, "grad_norm": 3.136896848678589, "learning_rate": 0.00019948160575834997, "loss": 5.072, "step": 53700 }, { "epoch": 0.23136091872882777, "grad_norm": 0.33145132660865784, "learning_rate": 0.0001994806224555412, "loss": 4.8225, "step": 53750 }, { "epoch": 0.2315761381881104, "grad_norm": 1.6117393970489502, "learning_rate": 0.0001994796382234676, "loss": 5.0981, "step": 53800 }, { "epoch": 0.23179135764739306, "grad_norm": 1.3694429397583008, "learning_rate": 0.00019947865306213833, "loss": 5.5163, "step": 53850 }, { "epoch": 0.2320065771066757, "grad_norm": 1.0630074739456177, "learning_rate": 0.00019947766697156266, "loss": 5.7105, "step": 53900 }, { "epoch": 0.23222179656595832, "grad_norm": 1.1552022695541382, "learning_rate": 0.00019947667995174974, "loss": 5.2232, "step": 53950 }, { "epoch": 0.23243701602524094, "grad_norm": 1.2066733837127686, "learning_rate": 0.00019947569200270883, "loss": 5.4102, "step": 54000 }, { "epoch": 0.23243701602524094, "eval_loss": 5.933196067810059, "eval_runtime": 49.6349, "eval_samples_per_second": 12.894, "eval_steps_per_second": 6.447, "eval_tts_loss": 6.159990210284922, "step": 54000 }, { "epoch": 0.23265223548452357, "grad_norm": 1.1599591970443726, "learning_rate": 0.00019947470312444915, "loss": 5.5433, "step": 54050 }, { "epoch": 0.2328674549438062, "grad_norm": 2.0679221153259277, "learning_rate": 0.00019947371331697993, "loss": 5.5963, "step": 54100 }, { "epoch": 0.23308267440308883, "grad_norm": 1.6526663303375244, "learning_rate": 0.0001994727225803104, "loss": 5.8352, "step": 54150 }, { "epoch": 0.23329789386237146, "grad_norm": 0.4619762599468231, "learning_rate": 0.00019947173091444987, "loss": 5.6489, "step": 54200 }, { "epoch": 0.2335131133216541, "grad_norm": 0.9636098742485046, "learning_rate": 0.00019947073831940752, "loss": 5.0892, "step": 54250 }, { "epoch": 0.23372833278093672, "grad_norm": 1.3923006057739258, "learning_rate": 0.00019946974479519267, "loss": 5.1337, "step": 54300 }, { "epoch": 0.23394355224021934, "grad_norm": 0.9768400192260742, "learning_rate": 0.00019946875034181462, "loss": 5.4833, "step": 54350 }, { "epoch": 0.23415877169950197, "grad_norm": 1.710841178894043, "learning_rate": 0.00019946775495928264, "loss": 6.0097, "step": 54400 }, { "epoch": 0.2343739911587846, "grad_norm": 1.445128083229065, "learning_rate": 0.000199466758647606, "loss": 5.303, "step": 54450 }, { "epoch": 0.23458921061806726, "grad_norm": 1.3129360675811768, "learning_rate": 0.00019946576140679408, "loss": 5.4097, "step": 54500 }, { "epoch": 0.23480443007734988, "grad_norm": 0.8349995017051697, "learning_rate": 0.0001994647632368561, "loss": 5.5871, "step": 54550 }, { "epoch": 0.2350196495366325, "grad_norm": 0.7794367671012878, "learning_rate": 0.00019946376413780144, "loss": 5.347, "step": 54600 }, { "epoch": 0.23523486899591514, "grad_norm": 1.9284659624099731, "learning_rate": 0.00019946276410963944, "loss": 5.6139, "step": 54650 }, { "epoch": 0.23545008845519777, "grad_norm": 1.7804235219955444, "learning_rate": 0.0001994617631523794, "loss": 5.4786, "step": 54700 }, { "epoch": 0.2356653079144804, "grad_norm": 0.6991044282913208, "learning_rate": 0.00019946076126603073, "loss": 5.5059, "step": 54750 }, { "epoch": 0.23588052737376303, "grad_norm": 1.4508239030838013, "learning_rate": 0.00019945975845060275, "loss": 5.4408, "step": 54800 }, { "epoch": 0.23609574683304566, "grad_norm": 0.9658709168434143, "learning_rate": 0.00019945875470610483, "loss": 5.9207, "step": 54850 }, { "epoch": 0.23631096629232828, "grad_norm": 0.5488547682762146, "learning_rate": 0.00019945775003254632, "loss": 5.4852, "step": 54900 }, { "epoch": 0.2365261857516109, "grad_norm": 0.9465630054473877, "learning_rate": 0.00019945674442993665, "loss": 5.4089, "step": 54950 }, { "epoch": 0.23674140521089354, "grad_norm": 1.5555872917175293, "learning_rate": 0.00019945573789828517, "loss": 5.4103, "step": 55000 }, { "epoch": 0.23695662467017617, "grad_norm": 1.0824365615844727, "learning_rate": 0.00019945473043760136, "loss": 5.7769, "step": 55050 }, { "epoch": 0.2371718441294588, "grad_norm": 1.6132992506027222, "learning_rate": 0.00019945372204789455, "loss": 5.3732, "step": 55100 }, { "epoch": 0.23738706358874143, "grad_norm": 3.084467649459839, "learning_rate": 0.0001994527127291742, "loss": 4.9094, "step": 55150 }, { "epoch": 0.23760228304802408, "grad_norm": 1.2095839977264404, "learning_rate": 0.00019945170248144973, "loss": 5.0587, "step": 55200 }, { "epoch": 0.2378175025073067, "grad_norm": 1.8821396827697754, "learning_rate": 0.00019945069130473058, "loss": 5.2084, "step": 55250 }, { "epoch": 0.23803272196658934, "grad_norm": 1.3566553592681885, "learning_rate": 0.00019944967919902618, "loss": 5.4226, "step": 55300 }, { "epoch": 0.23824794142587197, "grad_norm": 1.3181289434432983, "learning_rate": 0.00019944866616434595, "loss": 5.0628, "step": 55350 }, { "epoch": 0.2384631608851546, "grad_norm": 1.0303635597229004, "learning_rate": 0.00019944765220069946, "loss": 5.5708, "step": 55400 }, { "epoch": 0.23867838034443722, "grad_norm": 1.2528727054595947, "learning_rate": 0.0001994466373080961, "loss": 5.5421, "step": 55450 }, { "epoch": 0.23889359980371985, "grad_norm": 1.726357102394104, "learning_rate": 0.00019944562148654536, "loss": 5.6656, "step": 55500 }, { "epoch": 0.23910881926300248, "grad_norm": 0.9170743823051453, "learning_rate": 0.00019944460473605676, "loss": 5.8197, "step": 55550 }, { "epoch": 0.2393240387222851, "grad_norm": 0.8333684802055359, "learning_rate": 0.00019944358705663976, "loss": 5.4304, "step": 55600 }, { "epoch": 0.23953925818156774, "grad_norm": 1.7732478380203247, "learning_rate": 0.00019944256844830384, "loss": 5.7201, "step": 55650 }, { "epoch": 0.23975447764085037, "grad_norm": 0.8470239639282227, "learning_rate": 0.0001994415489110586, "loss": 5.1033, "step": 55700 }, { "epoch": 0.239969697100133, "grad_norm": 0.9307066798210144, "learning_rate": 0.00019944052844491353, "loss": 5.1524, "step": 55750 }, { "epoch": 0.24018491655941562, "grad_norm": 1.613258719444275, "learning_rate": 0.00019943950704987814, "loss": 5.7384, "step": 55800 }, { "epoch": 0.24040013601869828, "grad_norm": 2.1772351264953613, "learning_rate": 0.00019943848472596198, "loss": 5.4521, "step": 55850 }, { "epoch": 0.2406153554779809, "grad_norm": 2.071523904800415, "learning_rate": 0.00019943746147317462, "loss": 5.5099, "step": 55900 }, { "epoch": 0.24083057493726354, "grad_norm": 2.6446337699890137, "learning_rate": 0.00019943643729152557, "loss": 5.3706, "step": 55950 }, { "epoch": 0.24104579439654616, "grad_norm": 1.577038288116455, "learning_rate": 0.00019943541218102445, "loss": 5.3577, "step": 56000 }, { "epoch": 0.2412610138558288, "grad_norm": 1.808347463607788, "learning_rate": 0.00019943438614168082, "loss": 5.6185, "step": 56050 }, { "epoch": 0.24147623331511142, "grad_norm": 1.6489652395248413, "learning_rate": 0.00019943335917350425, "loss": 5.3001, "step": 56100 }, { "epoch": 0.24169145277439405, "grad_norm": 0.39492520689964294, "learning_rate": 0.00019943233127650434, "loss": 5.5671, "step": 56150 }, { "epoch": 0.24190667223367668, "grad_norm": 0.7691733241081238, "learning_rate": 0.00019943130245069068, "loss": 5.2512, "step": 56200 }, { "epoch": 0.2421218916929593, "grad_norm": 1.2769277095794678, "learning_rate": 0.00019943027269607292, "loss": 4.879, "step": 56250 }, { "epoch": 0.24233711115224194, "grad_norm": 1.2992323637008667, "learning_rate": 0.00019942924201266066, "loss": 5.1395, "step": 56300 }, { "epoch": 0.24255233061152456, "grad_norm": 0.34803295135498047, "learning_rate": 0.0001994282104004635, "loss": 5.0497, "step": 56350 }, { "epoch": 0.2427675500708072, "grad_norm": 1.2923216819763184, "learning_rate": 0.0001994271778594911, "loss": 5.132, "step": 56400 }, { "epoch": 0.24298276953008982, "grad_norm": 1.9253233671188354, "learning_rate": 0.0001994261443897531, "loss": 4.9886, "step": 56450 }, { "epoch": 0.24319798898937248, "grad_norm": 0.9916245341300964, "learning_rate": 0.00019942510999125917, "loss": 5.5117, "step": 56500 }, { "epoch": 0.2434132084486551, "grad_norm": 0.9076166152954102, "learning_rate": 0.00019942407466401894, "loss": 5.2923, "step": 56550 }, { "epoch": 0.24362842790793773, "grad_norm": 1.195136308670044, "learning_rate": 0.00019942303840804213, "loss": 5.5561, "step": 56600 }, { "epoch": 0.24384364736722036, "grad_norm": 0.6302568912506104, "learning_rate": 0.00019942200122333837, "loss": 5.4746, "step": 56650 }, { "epoch": 0.244058866826503, "grad_norm": 0.9216848015785217, "learning_rate": 0.00019942096310991735, "loss": 5.7694, "step": 56700 }, { "epoch": 0.24427408628578562, "grad_norm": 1.3539690971374512, "learning_rate": 0.0001994199240677888, "loss": 5.0826, "step": 56750 }, { "epoch": 0.24448930574506825, "grad_norm": 1.4826633930206299, "learning_rate": 0.00019941888409696242, "loss": 5.6182, "step": 56800 }, { "epoch": 0.24470452520435088, "grad_norm": 1.5858474969863892, "learning_rate": 0.00019941784319744792, "loss": 5.0979, "step": 56850 }, { "epoch": 0.2449197446636335, "grad_norm": 3.00496244430542, "learning_rate": 0.00019941680136925503, "loss": 5.4586, "step": 56900 }, { "epoch": 0.24513496412291613, "grad_norm": 0.7692316770553589, "learning_rate": 0.00019941575861239343, "loss": 5.1424, "step": 56950 }, { "epoch": 0.24535018358219876, "grad_norm": 1.3859440088272095, "learning_rate": 0.00019941471492687294, "loss": 5.1116, "step": 57000 }, { "epoch": 0.24535018358219876, "eval_loss": 5.898652076721191, "eval_runtime": 49.7432, "eval_samples_per_second": 12.866, "eval_steps_per_second": 6.433, "eval_tts_loss": 6.118023093788803, "step": 57000 }, { "epoch": 0.2455654030414814, "grad_norm": 0.7647467255592346, "learning_rate": 0.00019941367031270325, "loss": 5.2691, "step": 57050 }, { "epoch": 0.24578062250076402, "grad_norm": 1.147693157196045, "learning_rate": 0.00019941262476989415, "loss": 5.8443, "step": 57100 }, { "epoch": 0.24599584196004665, "grad_norm": 0.29004064202308655, "learning_rate": 0.0001994115782984554, "loss": 5.5052, "step": 57150 }, { "epoch": 0.2462110614193293, "grad_norm": 0.8185487389564514, "learning_rate": 0.00019941053089839677, "loss": 5.6294, "step": 57200 }, { "epoch": 0.24642628087861193, "grad_norm": 0.3498176336288452, "learning_rate": 0.00019940948256972802, "loss": 5.4547, "step": 57250 }, { "epoch": 0.24664150033789456, "grad_norm": 1.4337948560714722, "learning_rate": 0.00019940843331245898, "loss": 5.5073, "step": 57300 }, { "epoch": 0.2468567197971772, "grad_norm": 3.3339781761169434, "learning_rate": 0.00019940738312659945, "loss": 4.9524, "step": 57350 }, { "epoch": 0.24707193925645982, "grad_norm": 1.4038867950439453, "learning_rate": 0.0001994063320121592, "loss": 5.2452, "step": 57400 }, { "epoch": 0.24728715871574244, "grad_norm": 2.6666653156280518, "learning_rate": 0.00019940527996914812, "loss": 5.4803, "step": 57450 }, { "epoch": 0.24750237817502507, "grad_norm": 0.38680300116539, "learning_rate": 0.00019940422699757596, "loss": 5.5608, "step": 57500 }, { "epoch": 0.2477175976343077, "grad_norm": 1.3732534646987915, "learning_rate": 0.0001994031730974526, "loss": 5.1602, "step": 57550 }, { "epoch": 0.24793281709359033, "grad_norm": 1.3537989854812622, "learning_rate": 0.00019940211826878788, "loss": 5.4748, "step": 57600 }, { "epoch": 0.24814803655287296, "grad_norm": 1.6893244981765747, "learning_rate": 0.00019940106251159164, "loss": 5.7819, "step": 57650 }, { "epoch": 0.2483632560121556, "grad_norm": 2.2792913913726807, "learning_rate": 0.00019940000582587377, "loss": 5.3994, "step": 57700 }, { "epoch": 0.24857847547143821, "grad_norm": 0.8679698705673218, "learning_rate": 0.00019939894821164412, "loss": 5.4778, "step": 57750 }, { "epoch": 0.24879369493072084, "grad_norm": 2.538102865219116, "learning_rate": 0.00019939788966891254, "loss": 5.3377, "step": 57800 }, { "epoch": 0.2490089143900035, "grad_norm": 1.2821847200393677, "learning_rate": 0.00019939683019768898, "loss": 5.399, "step": 57850 }, { "epoch": 0.24922413384928613, "grad_norm": 0.6606307625770569, "learning_rate": 0.00019939576979798328, "loss": 4.9964, "step": 57900 }, { "epoch": 0.24943935330856876, "grad_norm": 1.3994174003601074, "learning_rate": 0.0001993947084698054, "loss": 5.6159, "step": 57950 }, { "epoch": 0.24965457276785138, "grad_norm": 1.021496295928955, "learning_rate": 0.00019939364621316522, "loss": 5.4014, "step": 58000 }, { "epoch": 0.249869792227134, "grad_norm": 1.6874817609786987, "learning_rate": 0.00019939258302807266, "loss": 5.6529, "step": 58050 }, { "epoch": 0.2500850116864166, "grad_norm": 0.5173603296279907, "learning_rate": 0.00019939151891453763, "loss": 5.308, "step": 58100 }, { "epoch": 0.25030023114569927, "grad_norm": 0.788709819316864, "learning_rate": 0.00019939045387257012, "loss": 5.6951, "step": 58150 }, { "epoch": 0.2505154506049819, "grad_norm": 1.8218276500701904, "learning_rate": 0.00019938938790218006, "loss": 5.1917, "step": 58200 }, { "epoch": 0.2507306700642645, "grad_norm": 1.3255057334899902, "learning_rate": 0.0001993883210033774, "loss": 5.4549, "step": 58250 }, { "epoch": 0.2509458895235472, "grad_norm": 0.8510144948959351, "learning_rate": 0.00019938725317617212, "loss": 5.1687, "step": 58300 }, { "epoch": 0.2511611089828298, "grad_norm": 1.0843781232833862, "learning_rate": 0.00019938618442057418, "loss": 5.2796, "step": 58350 }, { "epoch": 0.25137632844211244, "grad_norm": 1.3087624311447144, "learning_rate": 0.00019938511473659357, "loss": 5.518, "step": 58400 }, { "epoch": 0.25159154790139504, "grad_norm": 0.7814476490020752, "learning_rate": 0.00019938404412424027, "loss": 5.364, "step": 58450 }, { "epoch": 0.2518067673606777, "grad_norm": 1.4798128604888916, "learning_rate": 0.00019938297258352433, "loss": 5.7872, "step": 58500 }, { "epoch": 0.2520219868199603, "grad_norm": 1.329793930053711, "learning_rate": 0.00019938190011445568, "loss": 5.9459, "step": 58550 }, { "epoch": 0.25223720627924295, "grad_norm": 1.0565863847732544, "learning_rate": 0.0001993808267170444, "loss": 5.5015, "step": 58600 }, { "epoch": 0.25245242573852555, "grad_norm": 1.0509189367294312, "learning_rate": 0.00019937975239130048, "loss": 5.6494, "step": 58650 }, { "epoch": 0.2526676451978082, "grad_norm": 1.2225273847579956, "learning_rate": 0.00019937867713723398, "loss": 5.0232, "step": 58700 }, { "epoch": 0.2528828646570908, "grad_norm": 1.938538670539856, "learning_rate": 0.00019937760095485496, "loss": 5.249, "step": 58750 }, { "epoch": 0.25309808411637347, "grad_norm": 1.1293383836746216, "learning_rate": 0.0001993765238441734, "loss": 5.207, "step": 58800 }, { "epoch": 0.2533133035756561, "grad_norm": 1.1867705583572388, "learning_rate": 0.00019937544580519946, "loss": 5.6551, "step": 58850 }, { "epoch": 0.2535285230349387, "grad_norm": 0.9488784670829773, "learning_rate": 0.00019937436683794315, "loss": 5.4558, "step": 58900 }, { "epoch": 0.2537437424942214, "grad_norm": 1.5962258577346802, "learning_rate": 0.00019937328694241455, "loss": 5.5896, "step": 58950 }, { "epoch": 0.253958961953504, "grad_norm": 0.8867974281311035, "learning_rate": 0.00019937220611862376, "loss": 5.2544, "step": 59000 }, { "epoch": 0.25417418141278664, "grad_norm": 1.0222053527832031, "learning_rate": 0.00019937112436658086, "loss": 5.8059, "step": 59050 }, { "epoch": 0.25438940087206924, "grad_norm": 0.9317276477813721, "learning_rate": 0.00019937004168629597, "loss": 6.0433, "step": 59100 }, { "epoch": 0.2546046203313519, "grad_norm": 0.5588962435722351, "learning_rate": 0.0001993689580777792, "loss": 4.8829, "step": 59150 }, { "epoch": 0.2548198397906345, "grad_norm": 1.3796110153198242, "learning_rate": 0.0001993678735410407, "loss": 5.3568, "step": 59200 }, { "epoch": 0.25503505924991715, "grad_norm": 0.2604403495788574, "learning_rate": 0.00019936678807609055, "loss": 5.0576, "step": 59250 }, { "epoch": 0.25525027870919975, "grad_norm": 0.8437201976776123, "learning_rate": 0.0001993657016829389, "loss": 5.0872, "step": 59300 }, { "epoch": 0.2554654981684824, "grad_norm": 2.668450355529785, "learning_rate": 0.00019936461436159594, "loss": 5.3543, "step": 59350 }, { "epoch": 0.255680717627765, "grad_norm": 1.0612366199493408, "learning_rate": 0.00019936352611207178, "loss": 5.2333, "step": 59400 }, { "epoch": 0.25589593708704766, "grad_norm": 0.9568899869918823, "learning_rate": 0.0001993624369343766, "loss": 5.7243, "step": 59450 }, { "epoch": 0.25611115654633027, "grad_norm": 0.8779672384262085, "learning_rate": 0.0001993613468285206, "loss": 5.714, "step": 59500 }, { "epoch": 0.2563263760056129, "grad_norm": 0.9536569118499756, "learning_rate": 0.00019936025579451395, "loss": 4.908, "step": 59550 }, { "epoch": 0.2565415954648956, "grad_norm": 0.9024091958999634, "learning_rate": 0.0001993591638323668, "loss": 5.1625, "step": 59600 }, { "epoch": 0.2567568149241782, "grad_norm": 2.0752742290496826, "learning_rate": 0.0001993580709420894, "loss": 5.5783, "step": 59650 }, { "epoch": 0.25697203438346083, "grad_norm": 1.188509225845337, "learning_rate": 0.00019935697712369192, "loss": 5.0159, "step": 59700 }, { "epoch": 0.25718725384274344, "grad_norm": 1.3463672399520874, "learning_rate": 0.00019935588237718463, "loss": 5.5025, "step": 59750 }, { "epoch": 0.2574024733020261, "grad_norm": 1.166969656944275, "learning_rate": 0.00019935478670257773, "loss": 5.669, "step": 59800 }, { "epoch": 0.2576176927613087, "grad_norm": 2.373234510421753, "learning_rate": 0.00019935369009988141, "loss": 5.0801, "step": 59850 }, { "epoch": 0.25783291222059135, "grad_norm": 1.5302847623825073, "learning_rate": 0.00019935259256910598, "loss": 5.2618, "step": 59900 }, { "epoch": 0.25804813167987395, "grad_norm": 0.6922273635864258, "learning_rate": 0.00019935149411026167, "loss": 4.7714, "step": 59950 }, { "epoch": 0.2582633511391566, "grad_norm": 1.7331899404525757, "learning_rate": 0.00019935039472335873, "loss": 5.2877, "step": 60000 }, { "epoch": 0.2582633511391566, "eval_loss": 5.879925727844238, "eval_runtime": 49.6178, "eval_samples_per_second": 12.899, "eval_steps_per_second": 6.449, "eval_tts_loss": 6.140848708201845, "step": 60000 }, { "epoch": 0.2584785705984392, "grad_norm": 0.8969081044197083, "learning_rate": 0.00019934929440840743, "loss": 5.1588, "step": 60050 }, { "epoch": 0.25869379005772186, "grad_norm": 1.5806442499160767, "learning_rate": 0.00019934819316541806, "loss": 5.4197, "step": 60100 }, { "epoch": 0.25890900951700446, "grad_norm": 1.5756323337554932, "learning_rate": 0.00019934709099440093, "loss": 5.6955, "step": 60150 }, { "epoch": 0.2591242289762871, "grad_norm": 1.2882046699523926, "learning_rate": 0.00019934598789536628, "loss": 5.3434, "step": 60200 }, { "epoch": 0.2593394484355698, "grad_norm": 2.4296467304229736, "learning_rate": 0.0001993448838683244, "loss": 5.5762, "step": 60250 }, { "epoch": 0.2595546678948524, "grad_norm": 2.0155727863311768, "learning_rate": 0.0001993437789132857, "loss": 5.2229, "step": 60300 }, { "epoch": 0.25976988735413503, "grad_norm": 1.2963393926620483, "learning_rate": 0.00019934267303026046, "loss": 5.3046, "step": 60350 }, { "epoch": 0.25998510681341763, "grad_norm": 1.8880547285079956, "learning_rate": 0.00019934156621925894, "loss": 5.2507, "step": 60400 }, { "epoch": 0.2602003262727003, "grad_norm": 0.9812639355659485, "learning_rate": 0.0001993404584802916, "loss": 5.9492, "step": 60450 }, { "epoch": 0.2604155457319829, "grad_norm": 1.7995789051055908, "learning_rate": 0.00019933934981336866, "loss": 5.1242, "step": 60500 }, { "epoch": 0.26063076519126555, "grad_norm": 1.5030778646469116, "learning_rate": 0.00019933824021850058, "loss": 5.6335, "step": 60550 }, { "epoch": 0.26084598465054815, "grad_norm": 1.8674529790878296, "learning_rate": 0.00019933712969569766, "loss": 6.035, "step": 60600 }, { "epoch": 0.2610612041098308, "grad_norm": 1.4734745025634766, "learning_rate": 0.0001993360182449703, "loss": 5.6836, "step": 60650 }, { "epoch": 0.2612764235691134, "grad_norm": 1.3083159923553467, "learning_rate": 0.0001993349058663289, "loss": 5.2926, "step": 60700 }, { "epoch": 0.26149164302839606, "grad_norm": 1.1803171634674072, "learning_rate": 0.00019933379255978383, "loss": 5.2836, "step": 60750 }, { "epoch": 0.26170686248767866, "grad_norm": 1.1944055557250977, "learning_rate": 0.00019933267832534547, "loss": 4.9428, "step": 60800 }, { "epoch": 0.2619220819469613, "grad_norm": 1.0121811628341675, "learning_rate": 0.00019933156316302426, "loss": 5.0909, "step": 60850 }, { "epoch": 0.26213730140624397, "grad_norm": 0.30350592732429504, "learning_rate": 0.00019933044707283058, "loss": 5.6729, "step": 60900 }, { "epoch": 0.2623525208655266, "grad_norm": 1.5595598220825195, "learning_rate": 0.00019932933005477492, "loss": 5.8932, "step": 60950 }, { "epoch": 0.26256774032480923, "grad_norm": 0.43122929334640503, "learning_rate": 0.00019932821210886766, "loss": 5.5887, "step": 61000 }, { "epoch": 0.26278295978409183, "grad_norm": 0.9367508888244629, "learning_rate": 0.00019932709323511926, "loss": 5.299, "step": 61050 }, { "epoch": 0.2629981792433745, "grad_norm": 1.337235689163208, "learning_rate": 0.00019932597343354016, "loss": 5.7677, "step": 61100 }, { "epoch": 0.2632133987026571, "grad_norm": 1.4722652435302734, "learning_rate": 0.00019932485270414082, "loss": 5.3795, "step": 61150 }, { "epoch": 0.26342861816193974, "grad_norm": 1.6413087844848633, "learning_rate": 0.00019932373104693176, "loss": 5.6048, "step": 61200 }, { "epoch": 0.26364383762122234, "grad_norm": 1.018934726715088, "learning_rate": 0.00019932260846192338, "loss": 5.5471, "step": 61250 }, { "epoch": 0.263859057080505, "grad_norm": 0.9404449462890625, "learning_rate": 0.0001993214849491262, "loss": 5.2828, "step": 61300 }, { "epoch": 0.2640742765397876, "grad_norm": 1.351480484008789, "learning_rate": 0.00019932036050855073, "loss": 5.5083, "step": 61350 }, { "epoch": 0.26428949599907026, "grad_norm": 1.6117326021194458, "learning_rate": 0.00019931923514020748, "loss": 5.051, "step": 61400 }, { "epoch": 0.26450471545835286, "grad_norm": 0.7107943296432495, "learning_rate": 0.0001993181088441069, "loss": 4.9878, "step": 61450 }, { "epoch": 0.2647199349176355, "grad_norm": 0.1900652050971985, "learning_rate": 0.00019931698162025958, "loss": 5.2513, "step": 61500 }, { "epoch": 0.26493515437691817, "grad_norm": 1.1671524047851562, "learning_rate": 0.00019931585346867606, "loss": 4.7672, "step": 61550 }, { "epoch": 0.26515037383620077, "grad_norm": 0.8750630617141724, "learning_rate": 0.0001993147243893668, "loss": 5.6909, "step": 61600 }, { "epoch": 0.2653655932954834, "grad_norm": 0.3895704448223114, "learning_rate": 0.00019931359438234237, "loss": 4.7731, "step": 61650 }, { "epoch": 0.265580812754766, "grad_norm": 1.105689287185669, "learning_rate": 0.0001993124634476134, "loss": 5.0936, "step": 61700 }, { "epoch": 0.2657960322140487, "grad_norm": 1.738852858543396, "learning_rate": 0.00019931133158519035, "loss": 4.8888, "step": 61750 }, { "epoch": 0.2660112516733313, "grad_norm": 1.3927006721496582, "learning_rate": 0.00019931019879508387, "loss": 5.2491, "step": 61800 }, { "epoch": 0.26622647113261394, "grad_norm": 1.0839394330978394, "learning_rate": 0.0001993090650773045, "loss": 5.2186, "step": 61850 }, { "epoch": 0.26644169059189654, "grad_norm": 0.941697359085083, "learning_rate": 0.00019930793043186284, "loss": 5.0865, "step": 61900 }, { "epoch": 0.2666569100511792, "grad_norm": 2.721285820007324, "learning_rate": 0.0001993067948587695, "loss": 5.7166, "step": 61950 }, { "epoch": 0.2668721295104618, "grad_norm": 1.1306333541870117, "learning_rate": 0.00019930565835803507, "loss": 4.991, "step": 62000 }, { "epoch": 0.26708734896974445, "grad_norm": 0.7689560651779175, "learning_rate": 0.00019930452092967017, "loss": 5.5704, "step": 62050 }, { "epoch": 0.26730256842902705, "grad_norm": 2.469508409500122, "learning_rate": 0.00019930338257368547, "loss": 5.6614, "step": 62100 }, { "epoch": 0.2675177878883097, "grad_norm": 1.0509867668151855, "learning_rate": 0.00019930224329009155, "loss": 5.6127, "step": 62150 }, { "epoch": 0.26773300734759237, "grad_norm": 0.800680935382843, "learning_rate": 0.00019930110307889905, "loss": 5.4182, "step": 62200 }, { "epoch": 0.26794822680687497, "grad_norm": 1.2979700565338135, "learning_rate": 0.00019929996194011863, "loss": 5.1867, "step": 62250 }, { "epoch": 0.2681634462661576, "grad_norm": 2.31569504737854, "learning_rate": 0.000199298819873761, "loss": 5.731, "step": 62300 }, { "epoch": 0.2683786657254402, "grad_norm": 1.429870367050171, "learning_rate": 0.00019929767687983674, "loss": 5.1256, "step": 62350 }, { "epoch": 0.2685938851847229, "grad_norm": 0.8851985335350037, "learning_rate": 0.0001992965329583566, "loss": 5.7623, "step": 62400 }, { "epoch": 0.2688091046440055, "grad_norm": 1.2248947620391846, "learning_rate": 0.0001992953881093312, "loss": 5.3508, "step": 62450 }, { "epoch": 0.26902432410328814, "grad_norm": 2.5327236652374268, "learning_rate": 0.00019929424233277131, "loss": 4.9638, "step": 62500 }, { "epoch": 0.26923954356257074, "grad_norm": 1.2113007307052612, "learning_rate": 0.0001992930956286876, "loss": 5.3611, "step": 62550 }, { "epoch": 0.2694547630218534, "grad_norm": 0.7209233045578003, "learning_rate": 0.00019929194799709075, "loss": 5.4988, "step": 62600 }, { "epoch": 0.269669982481136, "grad_norm": 0.5005207061767578, "learning_rate": 0.00019929079943799155, "loss": 5.2196, "step": 62650 }, { "epoch": 0.26988520194041865, "grad_norm": 1.0487269163131714, "learning_rate": 0.00019928964995140068, "loss": 5.2819, "step": 62700 }, { "epoch": 0.27010042139970125, "grad_norm": 1.148982286453247, "learning_rate": 0.00019928849953732884, "loss": 5.3952, "step": 62750 }, { "epoch": 0.2703156408589839, "grad_norm": 1.450569748878479, "learning_rate": 0.00019928734819578682, "loss": 5.0858, "step": 62800 }, { "epoch": 0.27053086031826656, "grad_norm": 0.9858448505401611, "learning_rate": 0.00019928619592678542, "loss": 5.4117, "step": 62850 }, { "epoch": 0.27074607977754916, "grad_norm": 1.1173619031906128, "learning_rate": 0.00019928504273033533, "loss": 5.5443, "step": 62900 }, { "epoch": 0.2709612992368318, "grad_norm": 1.6092243194580078, "learning_rate": 0.00019928388860644735, "loss": 5.8219, "step": 62950 }, { "epoch": 0.2711765186961144, "grad_norm": 1.0376160144805908, "learning_rate": 0.00019928273355513228, "loss": 5.1149, "step": 63000 }, { "epoch": 0.2711765186961144, "eval_loss": 5.871164798736572, "eval_runtime": 49.5446, "eval_samples_per_second": 12.918, "eval_steps_per_second": 6.459, "eval_tts_loss": 6.225132113937378, "step": 63000 }, { "epoch": 0.2713917381553971, "grad_norm": 2.2745652198791504, "learning_rate": 0.00019928157757640087, "loss": 5.4679, "step": 63050 }, { "epoch": 0.2716069576146797, "grad_norm": 1.3299380540847778, "learning_rate": 0.00019928042067026395, "loss": 5.6661, "step": 63100 }, { "epoch": 0.27182217707396233, "grad_norm": 1.4722543954849243, "learning_rate": 0.0001992792628367323, "loss": 5.1782, "step": 63150 }, { "epoch": 0.27203739653324494, "grad_norm": 1.5097934007644653, "learning_rate": 0.00019927810407581676, "loss": 5.4424, "step": 63200 }, { "epoch": 0.2722526159925276, "grad_norm": 1.5726457834243774, "learning_rate": 0.00019927694438752816, "loss": 5.4876, "step": 63250 }, { "epoch": 0.2724678354518102, "grad_norm": 2.3015873432159424, "learning_rate": 0.0001992757837718773, "loss": 5.1819, "step": 63300 }, { "epoch": 0.27268305491109285, "grad_norm": 1.0681145191192627, "learning_rate": 0.00019927462222887505, "loss": 5.2295, "step": 63350 }, { "epoch": 0.27289827437037545, "grad_norm": 1.8812118768692017, "learning_rate": 0.00019927345975853224, "loss": 5.3446, "step": 63400 }, { "epoch": 0.2731134938296581, "grad_norm": 1.0138641595840454, "learning_rate": 0.00019927229636085975, "loss": 5.2657, "step": 63450 }, { "epoch": 0.2733287132889407, "grad_norm": 1.4204459190368652, "learning_rate": 0.00019927113203586843, "loss": 5.4383, "step": 63500 }, { "epoch": 0.27354393274822336, "grad_norm": 1.4382017850875854, "learning_rate": 0.00019926996678356918, "loss": 5.6152, "step": 63550 }, { "epoch": 0.273759152207506, "grad_norm": 1.3196011781692505, "learning_rate": 0.00019926880060397282, "loss": 5.4973, "step": 63600 }, { "epoch": 0.2739743716667886, "grad_norm": 1.122327446937561, "learning_rate": 0.00019926763349709033, "loss": 5.5598, "step": 63650 }, { "epoch": 0.2741895911260713, "grad_norm": 0.7965094447135925, "learning_rate": 0.00019926646546293255, "loss": 5.0762, "step": 63700 }, { "epoch": 0.2744048105853539, "grad_norm": 1.2317302227020264, "learning_rate": 0.00019926529650151045, "loss": 5.7043, "step": 63750 }, { "epoch": 0.27462003004463653, "grad_norm": 0.5376330614089966, "learning_rate": 0.00019926412661283488, "loss": 4.8935, "step": 63800 }, { "epoch": 0.27483524950391913, "grad_norm": 1.5474131107330322, "learning_rate": 0.0001992629557969168, "loss": 5.2955, "step": 63850 }, { "epoch": 0.2750504689632018, "grad_norm": 0.5069549679756165, "learning_rate": 0.00019926178405376717, "loss": 5.2711, "step": 63900 }, { "epoch": 0.2752656884224844, "grad_norm": 1.6454486846923828, "learning_rate": 0.00019926061138339688, "loss": 5.3964, "step": 63950 }, { "epoch": 0.27548090788176705, "grad_norm": 1.4306119680404663, "learning_rate": 0.00019925943778581698, "loss": 5.3654, "step": 64000 }, { "epoch": 0.27569612734104965, "grad_norm": 0.9777541756629944, "learning_rate": 0.0001992582632610383, "loss": 5.4766, "step": 64050 }, { "epoch": 0.2759113468003323, "grad_norm": 0.4216323494911194, "learning_rate": 0.00019925708780907194, "loss": 4.9182, "step": 64100 }, { "epoch": 0.2761265662596149, "grad_norm": 0.40271201729774475, "learning_rate": 0.00019925591142992878, "loss": 5.5076, "step": 64150 }, { "epoch": 0.27634178571889756, "grad_norm": 1.1948846578598022, "learning_rate": 0.00019925473412361986, "loss": 5.2592, "step": 64200 }, { "epoch": 0.2765570051781802, "grad_norm": 1.0532360076904297, "learning_rate": 0.00019925355589015618, "loss": 5.3759, "step": 64250 }, { "epoch": 0.2767722246374628, "grad_norm": 0.8662710189819336, "learning_rate": 0.00019925237672954873, "loss": 5.786, "step": 64300 }, { "epoch": 0.27698744409674547, "grad_norm": 1.199186086654663, "learning_rate": 0.0001992511966418085, "loss": 4.8603, "step": 64350 }, { "epoch": 0.2772026635560281, "grad_norm": 1.65153968334198, "learning_rate": 0.0001992500156269466, "loss": 5.206, "step": 64400 }, { "epoch": 0.27741788301531073, "grad_norm": 1.5576038360595703, "learning_rate": 0.00019924883368497396, "loss": 5.7383, "step": 64450 }, { "epoch": 0.27763310247459333, "grad_norm": 2.0579190254211426, "learning_rate": 0.00019924765081590166, "loss": 5.5931, "step": 64500 }, { "epoch": 0.277848321933876, "grad_norm": 1.4515509605407715, "learning_rate": 0.0001992464670197408, "loss": 5.7746, "step": 64550 }, { "epoch": 0.2780635413931586, "grad_norm": 0.9352480173110962, "learning_rate": 0.00019924528229650236, "loss": 5.7478, "step": 64600 }, { "epoch": 0.27827876085244124, "grad_norm": 3.7729220390319824, "learning_rate": 0.00019924409664619743, "loss": 5.5028, "step": 64650 }, { "epoch": 0.27849398031172384, "grad_norm": 0.6527139544487, "learning_rate": 0.0001992429100688371, "loss": 5.0194, "step": 64700 }, { "epoch": 0.2787091997710065, "grad_norm": 1.1825965642929077, "learning_rate": 0.00019924172256443248, "loss": 5.344, "step": 64750 }, { "epoch": 0.2789244192302891, "grad_norm": 1.0752168893814087, "learning_rate": 0.00019924053413299459, "loss": 5.4503, "step": 64800 }, { "epoch": 0.27913963868957176, "grad_norm": 2.8413245677948, "learning_rate": 0.0001992393447745346, "loss": 5.5325, "step": 64850 }, { "epoch": 0.2793548581488544, "grad_norm": 1.1719008684158325, "learning_rate": 0.00019923815448906357, "loss": 4.778, "step": 64900 }, { "epoch": 0.279570077608137, "grad_norm": 1.0549101829528809, "learning_rate": 0.00019923696327659265, "loss": 5.5458, "step": 64950 }, { "epoch": 0.27978529706741967, "grad_norm": 1.0189428329467773, "learning_rate": 0.00019923577113713296, "loss": 5.0601, "step": 65000 }, { "epoch": 0.28000051652670227, "grad_norm": 1.6529736518859863, "learning_rate": 0.00019923457807069562, "loss": 5.4416, "step": 65050 }, { "epoch": 0.2802157359859849, "grad_norm": 1.0104727745056152, "learning_rate": 0.00019923338407729183, "loss": 5.0473, "step": 65100 }, { "epoch": 0.2804309554452675, "grad_norm": 0.9842342734336853, "learning_rate": 0.00019923218915693266, "loss": 5.4548, "step": 65150 }, { "epoch": 0.2806461749045502, "grad_norm": 0.8941143751144409, "learning_rate": 0.00019923099330962932, "loss": 5.449, "step": 65200 }, { "epoch": 0.2808613943638328, "grad_norm": 0.6821762323379517, "learning_rate": 0.000199229796535393, "loss": 5.1712, "step": 65250 }, { "epoch": 0.28107661382311544, "grad_norm": 0.9237731695175171, "learning_rate": 0.00019922859883423483, "loss": 5.0035, "step": 65300 }, { "epoch": 0.28129183328239804, "grad_norm": 0.9364696741104126, "learning_rate": 0.00019922740020616604, "loss": 5.3977, "step": 65350 }, { "epoch": 0.2815070527416807, "grad_norm": 1.2968800067901611, "learning_rate": 0.00019922620065119783, "loss": 5.7444, "step": 65400 }, { "epoch": 0.2817222722009633, "grad_norm": 0.9561204314231873, "learning_rate": 0.00019922500016934132, "loss": 4.9451, "step": 65450 }, { "epoch": 0.28193749166024595, "grad_norm": 1.071677327156067, "learning_rate": 0.00019922379876060784, "loss": 5.4164, "step": 65500 }, { "epoch": 0.2821527111195286, "grad_norm": 0.5360262989997864, "learning_rate": 0.00019922259642500854, "loss": 5.6898, "step": 65550 }, { "epoch": 0.2823679305788112, "grad_norm": 1.108843445777893, "learning_rate": 0.00019922139316255465, "loss": 5.2025, "step": 65600 }, { "epoch": 0.28258315003809387, "grad_norm": 1.212461233139038, "learning_rate": 0.00019922018897325749, "loss": 5.2528, "step": 65650 }, { "epoch": 0.28279836949737647, "grad_norm": 0.3431418836116791, "learning_rate": 0.0001992189838571282, "loss": 5.0032, "step": 65700 }, { "epoch": 0.2830135889566591, "grad_norm": 1.0608516931533813, "learning_rate": 0.00019921777781417812, "loss": 5.5792, "step": 65750 }, { "epoch": 0.2832288084159417, "grad_norm": 1.5319684743881226, "learning_rate": 0.00019921657084441845, "loss": 5.279, "step": 65800 }, { "epoch": 0.2834440278752244, "grad_norm": 1.9544752836227417, "learning_rate": 0.0001992153629478605, "loss": 5.1969, "step": 65850 }, { "epoch": 0.283659247334507, "grad_norm": 2.0416131019592285, "learning_rate": 0.00019921415412451555, "loss": 4.9177, "step": 65900 }, { "epoch": 0.28387446679378964, "grad_norm": 1.2180432081222534, "learning_rate": 0.00019921294437439492, "loss": 5.5838, "step": 65950 }, { "epoch": 0.28408968625307224, "grad_norm": 0.9794612526893616, "learning_rate": 0.00019921173369750987, "loss": 5.5642, "step": 66000 }, { "epoch": 0.28408968625307224, "eval_loss": 5.867137432098389, "eval_runtime": 49.5692, "eval_samples_per_second": 12.911, "eval_steps_per_second": 6.456, "eval_tts_loss": 6.095993421173494, "step": 66000 }, { "epoch": 0.2843049057123549, "grad_norm": 0.3153873682022095, "learning_rate": 0.00019921052209387173, "loss": 5.2204, "step": 66050 }, { "epoch": 0.2845201251716375, "grad_norm": 1.4601600170135498, "learning_rate": 0.0001992093095634918, "loss": 5.3236, "step": 66100 }, { "epoch": 0.28473534463092015, "grad_norm": 0.9067076444625854, "learning_rate": 0.0001992080961063814, "loss": 5.0876, "step": 66150 }, { "epoch": 0.2849505640902028, "grad_norm": 1.346899151802063, "learning_rate": 0.0001992068817225519, "loss": 5.4078, "step": 66200 }, { "epoch": 0.2851657835494854, "grad_norm": 1.466066598892212, "learning_rate": 0.00019920566641201464, "loss": 4.9349, "step": 66250 }, { "epoch": 0.28538100300876806, "grad_norm": 2.6139328479766846, "learning_rate": 0.00019920445017478094, "loss": 5.2166, "step": 66300 }, { "epoch": 0.28559622246805066, "grad_norm": 0.45421701669692993, "learning_rate": 0.00019920323301086216, "loss": 5.3102, "step": 66350 }, { "epoch": 0.2858114419273333, "grad_norm": 1.0745230913162231, "learning_rate": 0.0001992020149202697, "loss": 4.9543, "step": 66400 }, { "epoch": 0.2860266613866159, "grad_norm": 1.5027798414230347, "learning_rate": 0.00019920079590301495, "loss": 5.2679, "step": 66450 }, { "epoch": 0.2862418808458986, "grad_norm": 1.837969422340393, "learning_rate": 0.00019919957595910926, "loss": 5.7917, "step": 66500 }, { "epoch": 0.2864571003051812, "grad_norm": 1.0262552499771118, "learning_rate": 0.00019919835508856403, "loss": 5.1007, "step": 66550 }, { "epoch": 0.28667231976446383, "grad_norm": 0.7781627178192139, "learning_rate": 0.0001991971332913907, "loss": 5.5501, "step": 66600 }, { "epoch": 0.28688753922374644, "grad_norm": 1.3165550231933594, "learning_rate": 0.00019919591056760064, "loss": 5.2566, "step": 66650 }, { "epoch": 0.2871027586830291, "grad_norm": 0.9946284890174866, "learning_rate": 0.00019919468691720528, "loss": 5.0408, "step": 66700 }, { "epoch": 0.2873179781423117, "grad_norm": 1.508659839630127, "learning_rate": 0.00019919346234021604, "loss": 5.0736, "step": 66750 }, { "epoch": 0.28753319760159435, "grad_norm": 1.76619291305542, "learning_rate": 0.00019919223683664443, "loss": 5.2131, "step": 66800 }, { "epoch": 0.28774841706087695, "grad_norm": 1.1459413766860962, "learning_rate": 0.00019919101040650182, "loss": 4.9412, "step": 66850 }, { "epoch": 0.2879636365201596, "grad_norm": 0.9297530651092529, "learning_rate": 0.00019918978304979968, "loss": 5.328, "step": 66900 }, { "epoch": 0.28817885597944226, "grad_norm": 1.198630928993225, "learning_rate": 0.0001991885547665495, "loss": 5.5296, "step": 66950 }, { "epoch": 0.28839407543872486, "grad_norm": 1.6198254823684692, "learning_rate": 0.0001991873255567627, "loss": 4.896, "step": 67000 }, { "epoch": 0.2886092948980075, "grad_norm": 1.4500781297683716, "learning_rate": 0.00019918609542045088, "loss": 5.6998, "step": 67050 }, { "epoch": 0.2888245143572901, "grad_norm": 0.27062666416168213, "learning_rate": 0.0001991848643576254, "loss": 5.3941, "step": 67100 }, { "epoch": 0.2890397338165728, "grad_norm": 1.6398937702178955, "learning_rate": 0.00019918363236829783, "loss": 5.1178, "step": 67150 }, { "epoch": 0.2892549532758554, "grad_norm": 1.2050538063049316, "learning_rate": 0.00019918239945247964, "loss": 5.4526, "step": 67200 }, { "epoch": 0.28947017273513803, "grad_norm": 2.30238938331604, "learning_rate": 0.00019918116561018238, "loss": 6.0935, "step": 67250 }, { "epoch": 0.28968539219442063, "grad_norm": 0.5447923541069031, "learning_rate": 0.0001991799308414176, "loss": 4.9143, "step": 67300 }, { "epoch": 0.2899006116537033, "grad_norm": 1.0323342084884644, "learning_rate": 0.00019917869514619673, "loss": 4.8164, "step": 67350 }, { "epoch": 0.2901158311129859, "grad_norm": 1.455588936805725, "learning_rate": 0.00019917745852453145, "loss": 4.9863, "step": 67400 }, { "epoch": 0.29033105057226855, "grad_norm": 2.6993448734283447, "learning_rate": 0.0001991762209764332, "loss": 5.3733, "step": 67450 }, { "epoch": 0.29054627003155115, "grad_norm": 0.9622714519500732, "learning_rate": 0.00019917498250191358, "loss": 5.2846, "step": 67500 }, { "epoch": 0.2907614894908338, "grad_norm": 0.44370925426483154, "learning_rate": 0.00019917374310098417, "loss": 5.1386, "step": 67550 }, { "epoch": 0.29097670895011646, "grad_norm": 1.1750051975250244, "learning_rate": 0.00019917250277365654, "loss": 5.2063, "step": 67600 }, { "epoch": 0.29119192840939906, "grad_norm": 2.2386646270751953, "learning_rate": 0.00019917126151994228, "loss": 5.4903, "step": 67650 }, { "epoch": 0.2914071478686817, "grad_norm": 0.7769112586975098, "learning_rate": 0.00019917001933985297, "loss": 5.7998, "step": 67700 }, { "epoch": 0.2916223673279643, "grad_norm": 1.4545053243637085, "learning_rate": 0.00019916877623340026, "loss": 5.2068, "step": 67750 }, { "epoch": 0.29183758678724697, "grad_norm": 0.6497049927711487, "learning_rate": 0.00019916753220059564, "loss": 5.1041, "step": 67800 }, { "epoch": 0.2920528062465296, "grad_norm": 1.3967607021331787, "learning_rate": 0.0001991662872414509, "loss": 5.6838, "step": 67850 }, { "epoch": 0.29226802570581223, "grad_norm": 0.49085384607315063, "learning_rate": 0.00019916504135597754, "loss": 5.5457, "step": 67900 }, { "epoch": 0.29248324516509483, "grad_norm": 1.2447017431259155, "learning_rate": 0.00019916379454418726, "loss": 5.6187, "step": 67950 }, { "epoch": 0.2926984646243775, "grad_norm": 1.264901041984558, "learning_rate": 0.00019916254680609169, "loss": 5.4268, "step": 68000 }, { "epoch": 0.2929136840836601, "grad_norm": 1.2325254678726196, "learning_rate": 0.00019916129814170246, "loss": 5.3412, "step": 68050 }, { "epoch": 0.29312890354294274, "grad_norm": 1.146125078201294, "learning_rate": 0.0001991600485510313, "loss": 5.3804, "step": 68100 }, { "epoch": 0.29334412300222534, "grad_norm": 0.5340553522109985, "learning_rate": 0.0001991587980340898, "loss": 5.8074, "step": 68150 }, { "epoch": 0.293559342461508, "grad_norm": 1.0179150104522705, "learning_rate": 0.0001991575465908897, "loss": 5.4257, "step": 68200 }, { "epoch": 0.29377456192079066, "grad_norm": 1.0674803256988525, "learning_rate": 0.00019915629422144268, "loss": 5.6134, "step": 68250 }, { "epoch": 0.29398978138007326, "grad_norm": 1.142561435699463, "learning_rate": 0.00019915504092576043, "loss": 5.1788, "step": 68300 }, { "epoch": 0.2942050008393559, "grad_norm": 0.9329453706741333, "learning_rate": 0.00019915378670385466, "loss": 5.5133, "step": 68350 }, { "epoch": 0.2944202202986385, "grad_norm": 0.4120321273803711, "learning_rate": 0.00019915253155573708, "loss": 5.0717, "step": 68400 }, { "epoch": 0.29463543975792117, "grad_norm": 1.1016050577163696, "learning_rate": 0.00019915127548141942, "loss": 5.412, "step": 68450 }, { "epoch": 0.29485065921720377, "grad_norm": 0.9888213872909546, "learning_rate": 0.0001991500184809134, "loss": 5.2103, "step": 68500 }, { "epoch": 0.2950658786764864, "grad_norm": 0.9642549753189087, "learning_rate": 0.00019914876055423077, "loss": 5.4229, "step": 68550 }, { "epoch": 0.295281098135769, "grad_norm": 0.903765082359314, "learning_rate": 0.0001991475017013833, "loss": 5.0622, "step": 68600 }, { "epoch": 0.2954963175950517, "grad_norm": 0.6423918008804321, "learning_rate": 0.00019914624192238274, "loss": 5.5674, "step": 68650 }, { "epoch": 0.2957115370543343, "grad_norm": 1.531118631362915, "learning_rate": 0.00019914498121724084, "loss": 5.3773, "step": 68700 }, { "epoch": 0.29592675651361694, "grad_norm": 1.3013007640838623, "learning_rate": 0.0001991437195859694, "loss": 5.3104, "step": 68750 }, { "epoch": 0.29614197597289954, "grad_norm": 0.5730806589126587, "learning_rate": 0.00019914245702858015, "loss": 4.839, "step": 68800 }, { "epoch": 0.2963571954321822, "grad_norm": 2.352799415588379, "learning_rate": 0.000199141193545085, "loss": 5.7131, "step": 68850 }, { "epoch": 0.29657241489146485, "grad_norm": 1.3988633155822754, "learning_rate": 0.00019913992913549563, "loss": 5.0396, "step": 68900 }, { "epoch": 0.29678763435074745, "grad_norm": 1.1648283004760742, "learning_rate": 0.0001991386637998239, "loss": 5.6184, "step": 68950 }, { "epoch": 0.2970028538100301, "grad_norm": 1.0608463287353516, "learning_rate": 0.00019913739753808164, "loss": 4.5106, "step": 69000 }, { "epoch": 0.2970028538100301, "eval_loss": 5.836971282958984, "eval_runtime": 49.9413, "eval_samples_per_second": 12.815, "eval_steps_per_second": 6.408, "eval_tts_loss": 6.208672539491006, "step": 69000 }, { "epoch": 0.2972180732693127, "grad_norm": 1.5558115243911743, "learning_rate": 0.00019913613035028068, "loss": 5.1781, "step": 69050 }, { "epoch": 0.29743329272859537, "grad_norm": 0.3375048339366913, "learning_rate": 0.00019913486223643284, "loss": 5.4338, "step": 69100 }, { "epoch": 0.29764851218787797, "grad_norm": 0.4728853106498718, "learning_rate": 0.00019913359319654995, "loss": 5.144, "step": 69150 }, { "epoch": 0.2978637316471606, "grad_norm": 0.970389723777771, "learning_rate": 0.00019913232323064391, "loss": 5.3348, "step": 69200 }, { "epoch": 0.2980789511064432, "grad_norm": 1.6435494422912598, "learning_rate": 0.00019913105233872655, "loss": 4.722, "step": 69250 }, { "epoch": 0.2982941705657259, "grad_norm": 2.8756000995635986, "learning_rate": 0.00019912978052080975, "loss": 5.4815, "step": 69300 }, { "epoch": 0.2985093900250085, "grad_norm": 1.4299538135528564, "learning_rate": 0.0001991285077769054, "loss": 5.6095, "step": 69350 }, { "epoch": 0.29872460948429114, "grad_norm": 1.6154749393463135, "learning_rate": 0.00019912723410702533, "loss": 5.2131, "step": 69400 }, { "epoch": 0.29893982894357374, "grad_norm": 1.3093351125717163, "learning_rate": 0.00019912595951118153, "loss": 5.0966, "step": 69450 }, { "epoch": 0.2991550484028564, "grad_norm": 1.9883438348770142, "learning_rate": 0.0001991246839893859, "loss": 5.3842, "step": 69500 }, { "epoch": 0.29937026786213905, "grad_norm": 1.8229937553405762, "learning_rate": 0.00019912340754165025, "loss": 5.2032, "step": 69550 }, { "epoch": 0.29958548732142165, "grad_norm": 0.989456295967102, "learning_rate": 0.0001991221301679866, "loss": 5.0253, "step": 69600 }, { "epoch": 0.2998007067807043, "grad_norm": 1.143552303314209, "learning_rate": 0.00019912085186840685, "loss": 5.3272, "step": 69650 }, { "epoch": 0.3000159262399869, "grad_norm": 1.1889044046401978, "learning_rate": 0.00019911957264292292, "loss": 5.1066, "step": 69700 }, { "epoch": 0.30023114569926956, "grad_norm": 1.308305263519287, "learning_rate": 0.0001991182924915468, "loss": 5.303, "step": 69750 }, { "epoch": 0.30044636515855216, "grad_norm": 2.666020393371582, "learning_rate": 0.00019911701141429045, "loss": 4.8754, "step": 69800 }, { "epoch": 0.3006615846178348, "grad_norm": 0.8066595792770386, "learning_rate": 0.0001991157294111658, "loss": 5.4602, "step": 69850 }, { "epoch": 0.3008768040771174, "grad_norm": 0.46715113520622253, "learning_rate": 0.00019911444648218487, "loss": 5.4691, "step": 69900 }, { "epoch": 0.3010920235364001, "grad_norm": 1.0550122261047363, "learning_rate": 0.0001991131626273596, "loss": 5.5134, "step": 69950 }, { "epoch": 0.3013072429956827, "grad_norm": 0.9917054176330566, "learning_rate": 0.000199111877846702, "loss": 5.3549, "step": 70000 }, { "epoch": 0.30152246245496533, "grad_norm": 0.9414690732955933, "learning_rate": 0.00019911059214022404, "loss": 5.8014, "step": 70050 }, { "epoch": 0.30173768191424793, "grad_norm": 0.782961905002594, "learning_rate": 0.00019910930550793782, "loss": 5.0287, "step": 70100 }, { "epoch": 0.3019529013735306, "grad_norm": 2.059298515319824, "learning_rate": 0.00019910801794985528, "loss": 5.8265, "step": 70150 }, { "epoch": 0.30216812083281325, "grad_norm": 2.2751901149749756, "learning_rate": 0.00019910672946598846, "loss": 5.4375, "step": 70200 }, { "epoch": 0.30238334029209585, "grad_norm": 1.4261114597320557, "learning_rate": 0.0001991054400563494, "loss": 5.4288, "step": 70250 }, { "epoch": 0.3025985597513785, "grad_norm": 0.9379740357398987, "learning_rate": 0.00019910414972095015, "loss": 5.72, "step": 70300 }, { "epoch": 0.3028137792106611, "grad_norm": 1.7472569942474365, "learning_rate": 0.00019910285845980277, "loss": 5.0566, "step": 70350 }, { "epoch": 0.30302899866994376, "grad_norm": 1.6992313861846924, "learning_rate": 0.0001991015662729193, "loss": 5.4258, "step": 70400 }, { "epoch": 0.30324421812922636, "grad_norm": 0.48861002922058105, "learning_rate": 0.00019910027316031182, "loss": 5.3051, "step": 70450 }, { "epoch": 0.303459437588509, "grad_norm": 1.1619253158569336, "learning_rate": 0.00019909897912199242, "loss": 5.6029, "step": 70500 }, { "epoch": 0.3036746570477916, "grad_norm": 1.0039682388305664, "learning_rate": 0.00019909768415797322, "loss": 5.8169, "step": 70550 }, { "epoch": 0.3038898765070743, "grad_norm": 1.6701710224151611, "learning_rate": 0.00019909638826826623, "loss": 5.1364, "step": 70600 }, { "epoch": 0.3041050959663569, "grad_norm": 0.9190298318862915, "learning_rate": 0.0001990950914528836, "loss": 5.1043, "step": 70650 }, { "epoch": 0.30432031542563953, "grad_norm": 1.1085485219955444, "learning_rate": 0.00019909379371183748, "loss": 4.8433, "step": 70700 }, { "epoch": 0.30453553488492213, "grad_norm": 1.5493112802505493, "learning_rate": 0.00019909249504513994, "loss": 5.2927, "step": 70750 }, { "epoch": 0.3047507543442048, "grad_norm": 2.11024808883667, "learning_rate": 0.00019909119545280312, "loss": 5.0528, "step": 70800 }, { "epoch": 0.3049659738034874, "grad_norm": 0.6498634219169617, "learning_rate": 0.0001990898949348392, "loss": 5.3103, "step": 70850 }, { "epoch": 0.30518119326277005, "grad_norm": 1.1049177646636963, "learning_rate": 0.00019908859349126025, "loss": 5.4782, "step": 70900 }, { "epoch": 0.3053964127220527, "grad_norm": 1.4137567281723022, "learning_rate": 0.0001990872911220785, "loss": 5.311, "step": 70950 }, { "epoch": 0.3056116321813353, "grad_norm": 1.2929673194885254, "learning_rate": 0.00019908598782730611, "loss": 5.7224, "step": 71000 }, { "epoch": 0.30582685164061796, "grad_norm": 0.5374126434326172, "learning_rate": 0.00019908468360695523, "loss": 5.4514, "step": 71050 }, { "epoch": 0.30604207109990056, "grad_norm": 0.6860604286193848, "learning_rate": 0.00019908337846103803, "loss": 5.2789, "step": 71100 }, { "epoch": 0.3062572905591832, "grad_norm": 1.859605073928833, "learning_rate": 0.00019908207238956673, "loss": 4.8399, "step": 71150 }, { "epoch": 0.3064725100184658, "grad_norm": 2.2693068981170654, "learning_rate": 0.0001990807653925535, "loss": 5.3613, "step": 71200 }, { "epoch": 0.30668772947774847, "grad_norm": 0.8788193464279175, "learning_rate": 0.0001990794574700106, "loss": 5.4697, "step": 71250 }, { "epoch": 0.3069029489370311, "grad_norm": 2.0328993797302246, "learning_rate": 0.0001990781486219502, "loss": 5.1864, "step": 71300 }, { "epoch": 0.30711816839631373, "grad_norm": 1.2383195161819458, "learning_rate": 0.00019907683884838454, "loss": 5.0245, "step": 71350 }, { "epoch": 0.30733338785559633, "grad_norm": 1.8877040147781372, "learning_rate": 0.00019907552814932587, "loss": 5.3304, "step": 71400 }, { "epoch": 0.307548607314879, "grad_norm": 1.3530504703521729, "learning_rate": 0.0001990742165247864, "loss": 5.0305, "step": 71450 }, { "epoch": 0.3077638267741616, "grad_norm": 1.0530447959899902, "learning_rate": 0.00019907290397477842, "loss": 5.3821, "step": 71500 }, { "epoch": 0.30797904623344424, "grad_norm": 0.45578014850616455, "learning_rate": 0.00019907159049931415, "loss": 5.2246, "step": 71550 }, { "epoch": 0.3081942656927269, "grad_norm": 0.22908712923526764, "learning_rate": 0.00019907027609840591, "loss": 5.276, "step": 71600 }, { "epoch": 0.3084094851520095, "grad_norm": 1.1186048984527588, "learning_rate": 0.00019906896077206594, "loss": 5.8003, "step": 71650 }, { "epoch": 0.30862470461129216, "grad_norm": 0.8725970387458801, "learning_rate": 0.00019906764452030652, "loss": 5.0288, "step": 71700 }, { "epoch": 0.30883992407057476, "grad_norm": 1.2545086145401, "learning_rate": 0.00019906632734314, "loss": 5.5399, "step": 71750 }, { "epoch": 0.3090551435298574, "grad_norm": 1.9843069314956665, "learning_rate": 0.00019906500924057863, "loss": 5.3769, "step": 71800 }, { "epoch": 0.30927036298914, "grad_norm": 1.5703312158584595, "learning_rate": 0.00019906369021263474, "loss": 5.5671, "step": 71850 }, { "epoch": 0.30948558244842267, "grad_norm": 1.1905301809310913, "learning_rate": 0.00019906237025932064, "loss": 5.4668, "step": 71900 }, { "epoch": 0.30970080190770527, "grad_norm": 0.9840947389602661, "learning_rate": 0.00019906104938064866, "loss": 5.327, "step": 71950 }, { "epoch": 0.3099160213669879, "grad_norm": 0.9461138248443604, "learning_rate": 0.00019905972757663118, "loss": 5.5259, "step": 72000 }, { "epoch": 0.3099160213669879, "eval_loss": 5.835556983947754, "eval_runtime": 49.7553, "eval_samples_per_second": 12.863, "eval_steps_per_second": 6.431, "eval_tts_loss": 6.216718823243762, "step": 72000 }, { "epoch": 0.3101312408262705, "grad_norm": 0.5248814821243286, "learning_rate": 0.00019905840484728051, "loss": 5.4628, "step": 72050 }, { "epoch": 0.3103464602855532, "grad_norm": 1.2956351041793823, "learning_rate": 0.000199057081192609, "loss": 5.1591, "step": 72100 }, { "epoch": 0.3105616797448358, "grad_norm": 0.9425171613693237, "learning_rate": 0.00019905575661262903, "loss": 5.6221, "step": 72150 }, { "epoch": 0.31077689920411844, "grad_norm": 0.7613900899887085, "learning_rate": 0.00019905443110735296, "loss": 5.4065, "step": 72200 }, { "epoch": 0.3109921186634011, "grad_norm": 1.600230097770691, "learning_rate": 0.0001990531046767932, "loss": 5.5943, "step": 72250 }, { "epoch": 0.3112073381226837, "grad_norm": 1.6966335773468018, "learning_rate": 0.0001990517773209621, "loss": 5.5542, "step": 72300 }, { "epoch": 0.31142255758196635, "grad_norm": 0.9809229373931885, "learning_rate": 0.00019905044903987209, "loss": 5.4873, "step": 72350 }, { "epoch": 0.31163777704124895, "grad_norm": 0.8278667330741882, "learning_rate": 0.0001990491198335356, "loss": 5.4586, "step": 72400 }, { "epoch": 0.3118529965005316, "grad_norm": 0.5459896922111511, "learning_rate": 0.00019904778970196495, "loss": 5.1817, "step": 72450 }, { "epoch": 0.3120682159598142, "grad_norm": 1.3841564655303955, "learning_rate": 0.00019904645864517268, "loss": 5.0582, "step": 72500 }, { "epoch": 0.31228343541909687, "grad_norm": 1.2816241979599, "learning_rate": 0.00019904512666317118, "loss": 4.9344, "step": 72550 }, { "epoch": 0.31249865487837947, "grad_norm": 1.3908047676086426, "learning_rate": 0.00019904379375597285, "loss": 5.5705, "step": 72600 }, { "epoch": 0.3127138743376621, "grad_norm": 1.5982807874679565, "learning_rate": 0.00019904245992359018, "loss": 5.0999, "step": 72650 }, { "epoch": 0.3129290937969447, "grad_norm": 0.9752349257469177, "learning_rate": 0.00019904112516603567, "loss": 5.2963, "step": 72700 }, { "epoch": 0.3131443132562274, "grad_norm": 1.459230661392212, "learning_rate": 0.00019903978948332168, "loss": 5.356, "step": 72750 }, { "epoch": 0.31335953271551, "grad_norm": 1.9065794944763184, "learning_rate": 0.00019903845287546078, "loss": 5.1222, "step": 72800 }, { "epoch": 0.31357475217479264, "grad_norm": 0.9093602299690247, "learning_rate": 0.00019903711534246545, "loss": 5.2066, "step": 72850 }, { "epoch": 0.3137899716340753, "grad_norm": 1.3751813173294067, "learning_rate": 0.00019903577688434813, "loss": 5.2927, "step": 72900 }, { "epoch": 0.3140051910933579, "grad_norm": 1.275325894355774, "learning_rate": 0.00019903443750112136, "loss": 5.6133, "step": 72950 }, { "epoch": 0.31422041055264055, "grad_norm": 0.895286500453949, "learning_rate": 0.00019903309719279766, "loss": 5.0992, "step": 73000 }, { "epoch": 0.31443563001192315, "grad_norm": 0.2969270348548889, "learning_rate": 0.00019903175595938953, "loss": 5.3337, "step": 73050 }, { "epoch": 0.3146508494712058, "grad_norm": 1.0043972730636597, "learning_rate": 0.0001990304138009095, "loss": 5.2786, "step": 73100 }, { "epoch": 0.3148660689304884, "grad_norm": 1.221260666847229, "learning_rate": 0.0001990290707173701, "loss": 5.2692, "step": 73150 }, { "epoch": 0.31508128838977106, "grad_norm": 1.3537753820419312, "learning_rate": 0.0001990277267087839, "loss": 5.3309, "step": 73200 }, { "epoch": 0.31529650784905366, "grad_norm": 1.6285384893417358, "learning_rate": 0.00019902638177516345, "loss": 4.9684, "step": 73250 }, { "epoch": 0.3155117273083363, "grad_norm": 2.759366750717163, "learning_rate": 0.00019902503591652128, "loss": 5.3012, "step": 73300 }, { "epoch": 0.3157269467676189, "grad_norm": 1.528834581375122, "learning_rate": 0.00019902368913286999, "loss": 5.7567, "step": 73350 }, { "epoch": 0.3159421662269016, "grad_norm": 1.1157690286636353, "learning_rate": 0.00019902234142422218, "loss": 5.4389, "step": 73400 }, { "epoch": 0.3161573856861842, "grad_norm": 2.367074489593506, "learning_rate": 0.00019902099279059038, "loss": 5.7612, "step": 73450 }, { "epoch": 0.31637260514546683, "grad_norm": 1.0102823972702026, "learning_rate": 0.00019901964323198725, "loss": 5.3272, "step": 73500 }, { "epoch": 0.3165878246047495, "grad_norm": 0.2647194564342499, "learning_rate": 0.00019901829274842538, "loss": 5.2546, "step": 73550 }, { "epoch": 0.3168030440640321, "grad_norm": 1.3755005598068237, "learning_rate": 0.00019901694133991736, "loss": 5.3446, "step": 73600 }, { "epoch": 0.31701826352331475, "grad_norm": 1.3309435844421387, "learning_rate": 0.00019901558900647584, "loss": 4.9041, "step": 73650 }, { "epoch": 0.31723348298259735, "grad_norm": 1.0189489126205444, "learning_rate": 0.00019901423574811343, "loss": 5.1729, "step": 73700 }, { "epoch": 0.31744870244188, "grad_norm": 0.34407445788383484, "learning_rate": 0.00019901288156484278, "loss": 5.618, "step": 73750 }, { "epoch": 0.3176639219011626, "grad_norm": 0.9659144282341003, "learning_rate": 0.00019901152645667656, "loss": 4.6486, "step": 73800 }, { "epoch": 0.31787914136044526, "grad_norm": 0.9739236831665039, "learning_rate": 0.0001990101704236274, "loss": 5.232, "step": 73850 }, { "epoch": 0.31809436081972786, "grad_norm": 1.206861972808838, "learning_rate": 0.00019900881346570796, "loss": 5.31, "step": 73900 }, { "epoch": 0.3183095802790105, "grad_norm": 1.6683642864227295, "learning_rate": 0.00019900745558293096, "loss": 5.432, "step": 73950 }, { "epoch": 0.3185247997382931, "grad_norm": 0.7589787244796753, "learning_rate": 0.00019900609677530904, "loss": 5.0736, "step": 74000 }, { "epoch": 0.3187400191975758, "grad_norm": 1.1359760761260986, "learning_rate": 0.00019900473704285494, "loss": 5.5529, "step": 74050 }, { "epoch": 0.3189552386568584, "grad_norm": 0.9875409603118896, "learning_rate": 0.0001990033763855813, "loss": 5.0495, "step": 74100 }, { "epoch": 0.31917045811614103, "grad_norm": 1.465036392211914, "learning_rate": 0.00019900201480350086, "loss": 5.5387, "step": 74150 }, { "epoch": 0.31938567757542363, "grad_norm": 2.0794029235839844, "learning_rate": 0.00019900065229662637, "loss": 4.787, "step": 74200 }, { "epoch": 0.3196008970347063, "grad_norm": 1.0492929220199585, "learning_rate": 0.00019899928886497051, "loss": 5.6444, "step": 74250 }, { "epoch": 0.31981611649398894, "grad_norm": 1.0311580896377563, "learning_rate": 0.00019899792450854605, "loss": 5.15, "step": 74300 }, { "epoch": 0.32003133595327155, "grad_norm": 0.2868709862232208, "learning_rate": 0.0001989965592273657, "loss": 5.3279, "step": 74350 }, { "epoch": 0.3202465554125542, "grad_norm": 1.454851508140564, "learning_rate": 0.00019899519302144221, "loss": 5.1285, "step": 74400 }, { "epoch": 0.3204617748718368, "grad_norm": 1.5956199169158936, "learning_rate": 0.0001989938258907884, "loss": 4.8558, "step": 74450 }, { "epoch": 0.32067699433111946, "grad_norm": 0.23580290377140045, "learning_rate": 0.000198992457835417, "loss": 5.8293, "step": 74500 }, { "epoch": 0.32089221379040206, "grad_norm": 2.3093059062957764, "learning_rate": 0.0001989910888553408, "loss": 5.9197, "step": 74550 }, { "epoch": 0.3211074332496847, "grad_norm": 0.8805928230285645, "learning_rate": 0.00019898971895057255, "loss": 5.2191, "step": 74600 }, { "epoch": 0.3213226527089673, "grad_norm": 1.3612765073776245, "learning_rate": 0.0001989883481211251, "loss": 5.3218, "step": 74650 }, { "epoch": 0.32153787216824997, "grad_norm": 1.3770984411239624, "learning_rate": 0.00019898697636701123, "loss": 5.6784, "step": 74700 }, { "epoch": 0.3217530916275326, "grad_norm": 2.7420105934143066, "learning_rate": 0.00019898560368824378, "loss": 5.5778, "step": 74750 }, { "epoch": 0.32196831108681523, "grad_norm": 1.3744019269943237, "learning_rate": 0.00019898423008483552, "loss": 5.258, "step": 74800 }, { "epoch": 0.32218353054609783, "grad_norm": 1.6315791606903076, "learning_rate": 0.00019898285555679932, "loss": 4.9771, "step": 74850 }, { "epoch": 0.3223987500053805, "grad_norm": 1.471250057220459, "learning_rate": 0.000198981480104148, "loss": 5.3815, "step": 74900 }, { "epoch": 0.32261396946466314, "grad_norm": 1.4421088695526123, "learning_rate": 0.00019898010372689445, "loss": 5.2223, "step": 74950 }, { "epoch": 0.32282918892394574, "grad_norm": 0.3042413890361786, "learning_rate": 0.00019897872642505148, "loss": 5.3932, "step": 75000 }, { "epoch": 0.32282918892394574, "eval_loss": 5.816632270812988, "eval_runtime": 49.6563, "eval_samples_per_second": 12.889, "eval_steps_per_second": 6.444, "eval_tts_loss": 6.256568662669459, "step": 75000 }, { "epoch": 0.3230444083832284, "grad_norm": 2.6437158584594727, "learning_rate": 0.00019897734819863193, "loss": 5.1403, "step": 75050 }, { "epoch": 0.323259627842511, "grad_norm": 0.5232779383659363, "learning_rate": 0.00019897596904764874, "loss": 4.9948, "step": 75100 }, { "epoch": 0.32347484730179366, "grad_norm": 1.8792979717254639, "learning_rate": 0.0001989745889721148, "loss": 5.7217, "step": 75150 }, { "epoch": 0.32369006676107626, "grad_norm": 1.3090003728866577, "learning_rate": 0.00019897320797204296, "loss": 5.5208, "step": 75200 }, { "epoch": 0.3239052862203589, "grad_norm": 1.058024287223816, "learning_rate": 0.00019897182604744612, "loss": 5.6674, "step": 75250 }, { "epoch": 0.3241205056796415, "grad_norm": 0.43548765778541565, "learning_rate": 0.0001989704431983372, "loss": 4.8276, "step": 75300 }, { "epoch": 0.32433572513892417, "grad_norm": 1.1836903095245361, "learning_rate": 0.0001989690594247291, "loss": 5.0045, "step": 75350 }, { "epoch": 0.32455094459820677, "grad_norm": 0.9054325222969055, "learning_rate": 0.0001989676747266348, "loss": 5.3277, "step": 75400 }, { "epoch": 0.3247661640574894, "grad_norm": 0.4670887887477875, "learning_rate": 0.00019896628910406715, "loss": 5.1967, "step": 75450 }, { "epoch": 0.324981383516772, "grad_norm": 1.1862126588821411, "learning_rate": 0.00019896490255703916, "loss": 4.8897, "step": 75500 }, { "epoch": 0.3251966029760547, "grad_norm": 0.958124041557312, "learning_rate": 0.00019896351508556377, "loss": 5.2202, "step": 75550 }, { "epoch": 0.32541182243533734, "grad_norm": 1.1534645557403564, "learning_rate": 0.00019896212668965392, "loss": 4.7881, "step": 75600 }, { "epoch": 0.32562704189461994, "grad_norm": 1.666911244392395, "learning_rate": 0.00019896073736932262, "loss": 5.5406, "step": 75650 }, { "epoch": 0.3258422613539026, "grad_norm": 1.023339033126831, "learning_rate": 0.00019895934712458278, "loss": 5.1366, "step": 75700 }, { "epoch": 0.3260574808131852, "grad_norm": 0.6236042976379395, "learning_rate": 0.00019895795595544744, "loss": 5.5929, "step": 75750 }, { "epoch": 0.32627270027246785, "grad_norm": 0.4430462419986725, "learning_rate": 0.0001989565638619296, "loss": 5.0457, "step": 75800 }, { "epoch": 0.32648791973175045, "grad_norm": 1.50547456741333, "learning_rate": 0.00019895517084404224, "loss": 5.4122, "step": 75850 }, { "epoch": 0.3267031391910331, "grad_norm": 2.3983607292175293, "learning_rate": 0.00019895377690179835, "loss": 5.3723, "step": 75900 }, { "epoch": 0.3269183586503157, "grad_norm": 2.03287935256958, "learning_rate": 0.00019895238203521102, "loss": 5.5812, "step": 75950 }, { "epoch": 0.32713357810959837, "grad_norm": 1.187690019607544, "learning_rate": 0.00019895098624429322, "loss": 5.5911, "step": 76000 }, { "epoch": 0.32734879756888097, "grad_norm": 1.7694103717803955, "learning_rate": 0.00019894958952905797, "loss": 5.4778, "step": 76050 }, { "epoch": 0.3275640170281636, "grad_norm": 1.4373817443847656, "learning_rate": 0.00019894819188951841, "loss": 5.3045, "step": 76100 }, { "epoch": 0.3277792364874462, "grad_norm": 0.5848255157470703, "learning_rate": 0.00019894679332568753, "loss": 5.6104, "step": 76150 }, { "epoch": 0.3279944559467289, "grad_norm": 0.5865252017974854, "learning_rate": 0.0001989453938375784, "loss": 5.706, "step": 76200 }, { "epoch": 0.32820967540601154, "grad_norm": 2.230865240097046, "learning_rate": 0.00019894399342520406, "loss": 4.8305, "step": 76250 }, { "epoch": 0.32842489486529414, "grad_norm": 0.3542001247406006, "learning_rate": 0.00019894259208857768, "loss": 5.4017, "step": 76300 }, { "epoch": 0.3286401143245768, "grad_norm": 0.6679815649986267, "learning_rate": 0.00019894118982771226, "loss": 5.5258, "step": 76350 }, { "epoch": 0.3288553337838594, "grad_norm": 1.4102795124053955, "learning_rate": 0.00019893978664262094, "loss": 5.7301, "step": 76400 }, { "epoch": 0.32907055324314205, "grad_norm": 0.4218766689300537, "learning_rate": 0.00019893838253331686, "loss": 5.5317, "step": 76450 }, { "epoch": 0.32928577270242465, "grad_norm": 1.5690826177597046, "learning_rate": 0.00019893697749981306, "loss": 5.8285, "step": 76500 }, { "epoch": 0.3295009921617073, "grad_norm": 1.4050378799438477, "learning_rate": 0.00019893557154212273, "loss": 5.867, "step": 76550 }, { "epoch": 0.3297162116209899, "grad_norm": 1.1503032445907593, "learning_rate": 0.00019893416466025896, "loss": 4.8233, "step": 76600 }, { "epoch": 0.32993143108027256, "grad_norm": 1.0819889307022095, "learning_rate": 0.0001989327568542349, "loss": 5.8181, "step": 76650 }, { "epoch": 0.33014665053955516, "grad_norm": 0.8860617280006409, "learning_rate": 0.00019893134812406377, "loss": 5.5732, "step": 76700 }, { "epoch": 0.3303618699988378, "grad_norm": 1.2157230377197266, "learning_rate": 0.00019892993846975862, "loss": 5.0914, "step": 76750 }, { "epoch": 0.3305770894581204, "grad_norm": 0.38327622413635254, "learning_rate": 0.00019892852789133266, "loss": 4.8117, "step": 76800 }, { "epoch": 0.3307923089174031, "grad_norm": 0.9518935680389404, "learning_rate": 0.00019892711638879907, "loss": 5.263, "step": 76850 }, { "epoch": 0.33100752837668573, "grad_norm": 0.9119991064071655, "learning_rate": 0.00019892570396217108, "loss": 5.2406, "step": 76900 }, { "epoch": 0.33122274783596833, "grad_norm": 0.2483982890844345, "learning_rate": 0.00019892429061146182, "loss": 4.8422, "step": 76950 }, { "epoch": 0.331437967295251, "grad_norm": 0.23411157727241516, "learning_rate": 0.0001989228763366845, "loss": 5.5495, "step": 77000 }, { "epoch": 0.3316531867545336, "grad_norm": 1.3146731853485107, "learning_rate": 0.0001989214611378524, "loss": 5.2221, "step": 77050 }, { "epoch": 0.33186840621381625, "grad_norm": 1.3578884601593018, "learning_rate": 0.00019892004501497864, "loss": 5.6299, "step": 77100 }, { "epoch": 0.33208362567309885, "grad_norm": 1.0364763736724854, "learning_rate": 0.0001989186279680765, "loss": 5.3774, "step": 77150 }, { "epoch": 0.3322988451323815, "grad_norm": 0.9951731562614441, "learning_rate": 0.00019891720999715918, "loss": 5.3616, "step": 77200 }, { "epoch": 0.3325140645916641, "grad_norm": 1.1129027605056763, "learning_rate": 0.00019891579110224, "loss": 5.1177, "step": 77250 }, { "epoch": 0.33272928405094676, "grad_norm": 1.152578353881836, "learning_rate": 0.00019891437128333216, "loss": 5.2366, "step": 77300 }, { "epoch": 0.33294450351022936, "grad_norm": 0.33281660079956055, "learning_rate": 0.00019891295054044896, "loss": 5.4432, "step": 77350 }, { "epoch": 0.333159722969512, "grad_norm": 0.8074372410774231, "learning_rate": 0.0001989115288736036, "loss": 4.9115, "step": 77400 }, { "epoch": 0.3333749424287946, "grad_norm": 0.47893890738487244, "learning_rate": 0.0001989101062828094, "loss": 5.2889, "step": 77450 }, { "epoch": 0.3335901618880773, "grad_norm": 1.1877598762512207, "learning_rate": 0.0001989086827680797, "loss": 5.2092, "step": 77500 }, { "epoch": 0.3338053813473599, "grad_norm": 1.4649574756622314, "learning_rate": 0.0001989072583294277, "loss": 5.4696, "step": 77550 }, { "epoch": 0.33402060080664253, "grad_norm": 0.8255696296691895, "learning_rate": 0.0001989058329668668, "loss": 5.4626, "step": 77600 }, { "epoch": 0.3342358202659252, "grad_norm": 1.0854380130767822, "learning_rate": 0.00019890440668041026, "loss": 5.2252, "step": 77650 }, { "epoch": 0.3344510397252078, "grad_norm": 1.1636431217193604, "learning_rate": 0.00019890297947007142, "loss": 5.0263, "step": 77700 }, { "epoch": 0.33466625918449044, "grad_norm": 1.3154423236846924, "learning_rate": 0.0001989015513358636, "loss": 5.0496, "step": 77750 }, { "epoch": 0.33488147864377305, "grad_norm": 2.9022109508514404, "learning_rate": 0.00019890012227780012, "loss": 5.5079, "step": 77800 }, { "epoch": 0.3350966981030557, "grad_norm": 0.7472476363182068, "learning_rate": 0.00019889869229589436, "loss": 5.1069, "step": 77850 }, { "epoch": 0.3353119175623383, "grad_norm": 0.5201330780982971, "learning_rate": 0.00019889726139015972, "loss": 5.3747, "step": 77900 }, { "epoch": 0.33552713702162096, "grad_norm": 0.5293427109718323, "learning_rate": 0.00019889582956060948, "loss": 5.3466, "step": 77950 }, { "epoch": 0.33574235648090356, "grad_norm": 1.4955304861068726, "learning_rate": 0.00019889439680725705, "loss": 4.7429, "step": 78000 }, { "epoch": 0.33574235648090356, "eval_loss": 5.812685012817383, "eval_runtime": 49.6036, "eval_samples_per_second": 12.902, "eval_steps_per_second": 6.451, "eval_tts_loss": 6.2458004377945935, "step": 78000 }, { "epoch": 0.3359575759401862, "grad_norm": 0.19656507670879364, "learning_rate": 0.00019889296313011583, "loss": 5.5726, "step": 78050 }, { "epoch": 0.3361727953994688, "grad_norm": 0.9900782704353333, "learning_rate": 0.00019889152852919922, "loss": 5.477, "step": 78100 }, { "epoch": 0.33638801485875147, "grad_norm": 0.9490571022033691, "learning_rate": 0.00019889009300452058, "loss": 4.7794, "step": 78150 }, { "epoch": 0.3366032343180341, "grad_norm": 1.2709683179855347, "learning_rate": 0.00019888865655609333, "loss": 5.7921, "step": 78200 }, { "epoch": 0.33681845377731673, "grad_norm": 0.9170878529548645, "learning_rate": 0.00019888721918393093, "loss": 4.8374, "step": 78250 }, { "epoch": 0.3370336732365994, "grad_norm": 1.9614874124526978, "learning_rate": 0.00019888578088804673, "loss": 5.5752, "step": 78300 }, { "epoch": 0.337248892695882, "grad_norm": 0.5722851157188416, "learning_rate": 0.00019888434166845427, "loss": 5.9601, "step": 78350 }, { "epoch": 0.33746411215516464, "grad_norm": 1.3677841424942017, "learning_rate": 0.00019888290152516688, "loss": 5.3591, "step": 78400 }, { "epoch": 0.33767933161444724, "grad_norm": 1.6835966110229492, "learning_rate": 0.00019888146045819808, "loss": 5.4093, "step": 78450 }, { "epoch": 0.3378945510737299, "grad_norm": 0.2911199927330017, "learning_rate": 0.00019888001846756135, "loss": 5.4754, "step": 78500 }, { "epoch": 0.3381097705330125, "grad_norm": 0.8818338513374329, "learning_rate": 0.0001988785755532701, "loss": 5.3987, "step": 78550 }, { "epoch": 0.33832498999229516, "grad_norm": 0.7543797492980957, "learning_rate": 0.00019887713171533783, "loss": 5.2546, "step": 78600 }, { "epoch": 0.33854020945157776, "grad_norm": 2.1869378089904785, "learning_rate": 0.00019887568695377806, "loss": 5.1854, "step": 78650 }, { "epoch": 0.3387554289108604, "grad_norm": 1.7009363174438477, "learning_rate": 0.00019887424126860425, "loss": 5.323, "step": 78700 }, { "epoch": 0.338970648370143, "grad_norm": 1.490432620048523, "learning_rate": 0.0001988727946598299, "loss": 5.0676, "step": 78750 }, { "epoch": 0.33918586782942567, "grad_norm": 0.7909417152404785, "learning_rate": 0.00019887134712746855, "loss": 5.1364, "step": 78800 }, { "epoch": 0.33940108728870827, "grad_norm": 0.9895051121711731, "learning_rate": 0.00019886989867153368, "loss": 4.804, "step": 78850 }, { "epoch": 0.3396163067479909, "grad_norm": 1.5644705295562744, "learning_rate": 0.00019886844929203892, "loss": 5.0059, "step": 78900 }, { "epoch": 0.3398315262072736, "grad_norm": 1.5916138887405396, "learning_rate": 0.0001988669989889977, "loss": 5.2689, "step": 78950 }, { "epoch": 0.3400467456665562, "grad_norm": 1.377152919769287, "learning_rate": 0.00019886554776242358, "loss": 5.1674, "step": 79000 }, { "epoch": 0.34026196512583884, "grad_norm": 1.0791736841201782, "learning_rate": 0.00019886409561233016, "loss": 5.31, "step": 79050 }, { "epoch": 0.34047718458512144, "grad_norm": 0.9436035752296448, "learning_rate": 0.000198862642538731, "loss": 5.4436, "step": 79100 }, { "epoch": 0.3406924040444041, "grad_norm": 1.0207070112228394, "learning_rate": 0.00019886118854163965, "loss": 5.4245, "step": 79150 }, { "epoch": 0.3409076235036867, "grad_norm": 1.4442112445831299, "learning_rate": 0.0001988597336210697, "loss": 5.503, "step": 79200 }, { "epoch": 0.34112284296296935, "grad_norm": 1.301574468612671, "learning_rate": 0.00019885827777703474, "loss": 5.3756, "step": 79250 }, { "epoch": 0.34133806242225195, "grad_norm": 0.9998663663864136, "learning_rate": 0.00019885682100954837, "loss": 4.8899, "step": 79300 }, { "epoch": 0.3415532818815346, "grad_norm": 2.571669816970825, "learning_rate": 0.0001988553633186242, "loss": 5.3501, "step": 79350 }, { "epoch": 0.3417685013408172, "grad_norm": 1.5996116399765015, "learning_rate": 0.00019885390470427585, "loss": 5.5571, "step": 79400 }, { "epoch": 0.34198372080009987, "grad_norm": 2.027496576309204, "learning_rate": 0.00019885244516651694, "loss": 5.4141, "step": 79450 }, { "epoch": 0.34219894025938247, "grad_norm": 1.2688251733779907, "learning_rate": 0.0001988509847053611, "loss": 5.0072, "step": 79500 }, { "epoch": 0.3424141597186651, "grad_norm": 1.2283700704574585, "learning_rate": 0.00019884952332082197, "loss": 5.0198, "step": 79550 }, { "epoch": 0.3426293791779478, "grad_norm": 2.8826253414154053, "learning_rate": 0.00019884806101291323, "loss": 5.3383, "step": 79600 }, { "epoch": 0.3428445986372304, "grad_norm": 0.9892476201057434, "learning_rate": 0.00019884659778164854, "loss": 5.0929, "step": 79650 }, { "epoch": 0.34305981809651304, "grad_norm": 0.9513550400733948, "learning_rate": 0.00019884513362704152, "loss": 4.8331, "step": 79700 }, { "epoch": 0.34327503755579564, "grad_norm": 1.2987762689590454, "learning_rate": 0.00019884366854910587, "loss": 5.3314, "step": 79750 }, { "epoch": 0.3434902570150783, "grad_norm": 2.2528154850006104, "learning_rate": 0.00019884220254785527, "loss": 5.7539, "step": 79800 }, { "epoch": 0.3437054764743609, "grad_norm": 1.5998704433441162, "learning_rate": 0.00019884073562330344, "loss": 5.1843, "step": 79850 }, { "epoch": 0.34392069593364355, "grad_norm": 0.7340489029884338, "learning_rate": 0.00019883926777546407, "loss": 5.041, "step": 79900 }, { "epoch": 0.34413591539292615, "grad_norm": 1.3127961158752441, "learning_rate": 0.00019883779900435085, "loss": 5.6249, "step": 79950 }, { "epoch": 0.3443511348522088, "grad_norm": 1.5402268171310425, "learning_rate": 0.0001988363293099775, "loss": 5.4642, "step": 80000 }, { "epoch": 0.3445663543114914, "grad_norm": 1.0891565084457397, "learning_rate": 0.0001988348586923578, "loss": 4.9332, "step": 80050 }, { "epoch": 0.34478157377077406, "grad_norm": 0.5498387217521667, "learning_rate": 0.00019883338715150545, "loss": 5.1546, "step": 80100 }, { "epoch": 0.34499679323005666, "grad_norm": 1.309228777885437, "learning_rate": 0.00019883191468743417, "loss": 5.3846, "step": 80150 }, { "epoch": 0.3452120126893393, "grad_norm": 1.2727851867675781, "learning_rate": 0.00019883044130015779, "loss": 5.9296, "step": 80200 }, { "epoch": 0.345427232148622, "grad_norm": 1.5175766944885254, "learning_rate": 0.00019882896698969, "loss": 4.9837, "step": 80250 }, { "epoch": 0.3456424516079046, "grad_norm": 1.1583060026168823, "learning_rate": 0.0001988274917560446, "loss": 5.6468, "step": 80300 }, { "epoch": 0.34585767106718723, "grad_norm": 2.0517542362213135, "learning_rate": 0.00019882601559923537, "loss": 5.5199, "step": 80350 }, { "epoch": 0.34607289052646983, "grad_norm": 0.3956255614757538, "learning_rate": 0.00019882453851927608, "loss": 5.3339, "step": 80400 }, { "epoch": 0.3462881099857525, "grad_norm": 1.3976761102676392, "learning_rate": 0.00019882306051618057, "loss": 5.4501, "step": 80450 }, { "epoch": 0.3465033294450351, "grad_norm": 1.1188108921051025, "learning_rate": 0.00019882158158996264, "loss": 5.4496, "step": 80500 }, { "epoch": 0.34671854890431775, "grad_norm": 1.0449341535568237, "learning_rate": 0.00019882010174063607, "loss": 5.158, "step": 80550 }, { "epoch": 0.34693376836360035, "grad_norm": 2.001939296722412, "learning_rate": 0.0001988186209682147, "loss": 5.1911, "step": 80600 }, { "epoch": 0.347148987822883, "grad_norm": 1.5066605806350708, "learning_rate": 0.00019881713927271234, "loss": 5.563, "step": 80650 }, { "epoch": 0.3473642072821656, "grad_norm": 1.6057406663894653, "learning_rate": 0.00019881565665414287, "loss": 5.028, "step": 80700 }, { "epoch": 0.34757942674144826, "grad_norm": 3.2972354888916016, "learning_rate": 0.00019881417311252013, "loss": 5.3692, "step": 80750 }, { "epoch": 0.34779464620073086, "grad_norm": 2.110369920730591, "learning_rate": 0.00019881268864785798, "loss": 5.0777, "step": 80800 }, { "epoch": 0.3480098656600135, "grad_norm": 0.6217010617256165, "learning_rate": 0.00019881120326017024, "loss": 5.0255, "step": 80850 }, { "epoch": 0.3482250851192962, "grad_norm": 0.8570384979248047, "learning_rate": 0.00019880971694947083, "loss": 5.1002, "step": 80900 }, { "epoch": 0.3484403045785788, "grad_norm": 0.39642226696014404, "learning_rate": 0.00019880822971577366, "loss": 5.1268, "step": 80950 }, { "epoch": 0.34865552403786143, "grad_norm": 1.0293327569961548, "learning_rate": 0.00019880674155909258, "loss": 5.2959, "step": 81000 }, { "epoch": 0.34865552403786143, "eval_loss": 5.802739143371582, "eval_runtime": 49.5632, "eval_samples_per_second": 12.913, "eval_steps_per_second": 6.456, "eval_tts_loss": 6.204205649483613, "step": 81000 }, { "epoch": 0.34887074349714403, "grad_norm": 0.6336168646812439, "learning_rate": 0.0001988052524794415, "loss": 5.1059, "step": 81050 }, { "epoch": 0.3490859629564267, "grad_norm": 1.0020567178726196, "learning_rate": 0.0001988037624768343, "loss": 4.7416, "step": 81100 }, { "epoch": 0.3493011824157093, "grad_norm": 1.2143259048461914, "learning_rate": 0.00019880227155128494, "loss": 5.5135, "step": 81150 }, { "epoch": 0.34951640187499194, "grad_norm": 1.3719422817230225, "learning_rate": 0.00019880077970280734, "loss": 5.4911, "step": 81200 }, { "epoch": 0.34973162133427455, "grad_norm": 1.6934672594070435, "learning_rate": 0.00019879928693141544, "loss": 5.5816, "step": 81250 }, { "epoch": 0.3499468407935572, "grad_norm": 2.084852933883667, "learning_rate": 0.00019879779323712316, "loss": 5.5137, "step": 81300 }, { "epoch": 0.3501620602528398, "grad_norm": 1.038326382637024, "learning_rate": 0.0001987962986199445, "loss": 5.1539, "step": 81350 }, { "epoch": 0.35037727971212246, "grad_norm": 0.9970792531967163, "learning_rate": 0.00019879480307989335, "loss": 5.7858, "step": 81400 }, { "epoch": 0.35059249917140506, "grad_norm": 0.904610276222229, "learning_rate": 0.00019879330661698373, "loss": 5.1276, "step": 81450 }, { "epoch": 0.3508077186306877, "grad_norm": 0.9890577793121338, "learning_rate": 0.0001987918092312296, "loss": 5.3929, "step": 81500 }, { "epoch": 0.3510229380899703, "grad_norm": 1.0181537866592407, "learning_rate": 0.00019879031092264498, "loss": 5.5945, "step": 81550 }, { "epoch": 0.35123815754925297, "grad_norm": 0.8964313864707947, "learning_rate": 0.00019878881169124385, "loss": 5.5202, "step": 81600 }, { "epoch": 0.35145337700853563, "grad_norm": 1.2868396043777466, "learning_rate": 0.0001987873115370402, "loss": 5.3432, "step": 81650 }, { "epoch": 0.35166859646781823, "grad_norm": 1.9279711246490479, "learning_rate": 0.000198785810460048, "loss": 5.7124, "step": 81700 }, { "epoch": 0.3518838159271009, "grad_norm": 1.3642297983169556, "learning_rate": 0.00019878430846028138, "loss": 5.9333, "step": 81750 }, { "epoch": 0.3520990353863835, "grad_norm": 0.38019654154777527, "learning_rate": 0.00019878280553775428, "loss": 5.5511, "step": 81800 }, { "epoch": 0.35231425484566614, "grad_norm": 1.8021266460418701, "learning_rate": 0.0001987813016924808, "loss": 5.4629, "step": 81850 }, { "epoch": 0.35252947430494874, "grad_norm": 0.3747234344482422, "learning_rate": 0.0001987797969244749, "loss": 4.9989, "step": 81900 }, { "epoch": 0.3527446937642314, "grad_norm": 0.7816826105117798, "learning_rate": 0.00019877829123375075, "loss": 5.0252, "step": 81950 }, { "epoch": 0.352959913223514, "grad_norm": 0.6312609314918518, "learning_rate": 0.00019877678462032234, "loss": 5.5962, "step": 82000 }, { "epoch": 0.35317513268279666, "grad_norm": 0.5077404975891113, "learning_rate": 0.00019877527708420375, "loss": 5.2144, "step": 82050 }, { "epoch": 0.35339035214207926, "grad_norm": 3.282327890396118, "learning_rate": 0.0001987737686254091, "loss": 5.5642, "step": 82100 }, { "epoch": 0.3536055716013619, "grad_norm": 2.282127618789673, "learning_rate": 0.00019877225924395245, "loss": 5.0603, "step": 82150 }, { "epoch": 0.3538207910606445, "grad_norm": 1.2372082471847534, "learning_rate": 0.00019877074893984788, "loss": 5.2277, "step": 82200 }, { "epoch": 0.35403601051992717, "grad_norm": 1.0760921239852905, "learning_rate": 0.00019876923771310954, "loss": 4.92, "step": 82250 }, { "epoch": 0.3542512299792098, "grad_norm": 1.6167240142822266, "learning_rate": 0.00019876772556375153, "loss": 5.1765, "step": 82300 }, { "epoch": 0.3544664494384924, "grad_norm": 1.5011370182037354, "learning_rate": 0.000198766212491788, "loss": 5.2255, "step": 82350 }, { "epoch": 0.3546816688977751, "grad_norm": 0.7011114954948425, "learning_rate": 0.00019876469849723302, "loss": 5.4789, "step": 82400 }, { "epoch": 0.3548968883570577, "grad_norm": 1.3472721576690674, "learning_rate": 0.0001987631835801008, "loss": 5.393, "step": 82450 }, { "epoch": 0.35511210781634034, "grad_norm": 0.9625145792961121, "learning_rate": 0.00019876166774040544, "loss": 5.5952, "step": 82500 }, { "epoch": 0.35532732727562294, "grad_norm": 1.7273434400558472, "learning_rate": 0.00019876015097816113, "loss": 5.1195, "step": 82550 }, { "epoch": 0.3555425467349056, "grad_norm": 1.3163423538208008, "learning_rate": 0.00019875863329338203, "loss": 5.0556, "step": 82600 }, { "epoch": 0.3557577661941882, "grad_norm": 1.539316177368164, "learning_rate": 0.0001987571146860823, "loss": 5.4569, "step": 82650 }, { "epoch": 0.35597298565347085, "grad_norm": 3.000507354736328, "learning_rate": 0.00019875559515627615, "loss": 4.7816, "step": 82700 }, { "epoch": 0.35618820511275345, "grad_norm": 1.2645989656448364, "learning_rate": 0.0001987540747039778, "loss": 5.2665, "step": 82750 }, { "epoch": 0.3564034245720361, "grad_norm": 0.9509237408638, "learning_rate": 0.0001987525533292014, "loss": 5.1757, "step": 82800 }, { "epoch": 0.3566186440313187, "grad_norm": 0.48791977763175964, "learning_rate": 0.00019875103103196118, "loss": 5.1434, "step": 82850 }, { "epoch": 0.35683386349060137, "grad_norm": 0.512896716594696, "learning_rate": 0.00019874950781227136, "loss": 5.2316, "step": 82900 }, { "epoch": 0.357049082949884, "grad_norm": 1.4114935398101807, "learning_rate": 0.00019874798367014616, "loss": 4.8794, "step": 82950 }, { "epoch": 0.3572643024091666, "grad_norm": 1.7547246217727661, "learning_rate": 0.00019874645860559984, "loss": 5.1932, "step": 83000 }, { "epoch": 0.3574795218684493, "grad_norm": 3.2740697860717773, "learning_rate": 0.0001987449326186466, "loss": 5.8483, "step": 83050 }, { "epoch": 0.3576947413277319, "grad_norm": 0.7520732283592224, "learning_rate": 0.0001987434057093008, "loss": 4.7314, "step": 83100 }, { "epoch": 0.35790996078701454, "grad_norm": 1.4581164121627808, "learning_rate": 0.00019874187787757657, "loss": 4.9343, "step": 83150 }, { "epoch": 0.35812518024629714, "grad_norm": 0.21867212653160095, "learning_rate": 0.00019874034912348827, "loss": 4.6431, "step": 83200 }, { "epoch": 0.3583403997055798, "grad_norm": 2.651803970336914, "learning_rate": 0.00019873881944705013, "loss": 5.0749, "step": 83250 }, { "epoch": 0.3585556191648624, "grad_norm": 1.7629032135009766, "learning_rate": 0.0001987372888482765, "loss": 5.1961, "step": 83300 }, { "epoch": 0.35877083862414505, "grad_norm": 1.6579959392547607, "learning_rate": 0.0001987357573271816, "loss": 5.4684, "step": 83350 }, { "epoch": 0.35898605808342765, "grad_norm": 0.4342198371887207, "learning_rate": 0.0001987342248837798, "loss": 4.9655, "step": 83400 }, { "epoch": 0.3592012775427103, "grad_norm": 1.3197637796401978, "learning_rate": 0.00019873269151808538, "loss": 4.8386, "step": 83450 }, { "epoch": 0.3594164970019929, "grad_norm": 1.1278932094573975, "learning_rate": 0.00019873115723011267, "loss": 5.4091, "step": 83500 }, { "epoch": 0.35963171646127556, "grad_norm": 1.0648422241210938, "learning_rate": 0.00019872962201987598, "loss": 5.7734, "step": 83550 }, { "epoch": 0.3598469359205582, "grad_norm": 0.9673904180526733, "learning_rate": 0.00019872808588738974, "loss": 5.3745, "step": 83600 }, { "epoch": 0.3600621553798408, "grad_norm": 2.3647608757019043, "learning_rate": 0.0001987265488326682, "loss": 5.133, "step": 83650 }, { "epoch": 0.3602773748391235, "grad_norm": 0.5540787577629089, "learning_rate": 0.00019872501085572573, "loss": 5.5415, "step": 83700 }, { "epoch": 0.3604925942984061, "grad_norm": 2.5640785694122314, "learning_rate": 0.00019872347195657678, "loss": 5.4496, "step": 83750 }, { "epoch": 0.36070781375768873, "grad_norm": 1.504962682723999, "learning_rate": 0.0001987219321352356, "loss": 5.1915, "step": 83800 }, { "epoch": 0.36092303321697133, "grad_norm": 3.2269387245178223, "learning_rate": 0.0001987203913917167, "loss": 5.0964, "step": 83850 }, { "epoch": 0.361138252676254, "grad_norm": 1.049228310585022, "learning_rate": 0.00019871884972603437, "loss": 5.3079, "step": 83900 }, { "epoch": 0.3613534721355366, "grad_norm": 1.5873708724975586, "learning_rate": 0.00019871730713820306, "loss": 5.3396, "step": 83950 }, { "epoch": 0.36156869159481925, "grad_norm": 0.7775790691375732, "learning_rate": 0.00019871576362823717, "loss": 5.4325, "step": 84000 }, { "epoch": 0.36156869159481925, "eval_loss": 5.815018653869629, "eval_runtime": 49.9773, "eval_samples_per_second": 12.806, "eval_steps_per_second": 6.403, "eval_tts_loss": 6.25183512322865, "step": 84000 }, { "epoch": 0.36178391105410185, "grad_norm": 0.9280009865760803, "learning_rate": 0.00019871421919615112, "loss": 4.966, "step": 84050 }, { "epoch": 0.3619991305133845, "grad_norm": 1.0812287330627441, "learning_rate": 0.00019871267384195934, "loss": 5.4469, "step": 84100 }, { "epoch": 0.3622143499726671, "grad_norm": 1.5501861572265625, "learning_rate": 0.00019871112756567627, "loss": 5.4037, "step": 84150 }, { "epoch": 0.36242956943194976, "grad_norm": 1.2483713626861572, "learning_rate": 0.00019870958036731634, "loss": 5.2635, "step": 84200 }, { "epoch": 0.3626447888912324, "grad_norm": 0.39273199439048767, "learning_rate": 0.000198708032246894, "loss": 5.1066, "step": 84250 }, { "epoch": 0.362860008350515, "grad_norm": 1.9892148971557617, "learning_rate": 0.00019870648320442373, "loss": 5.6587, "step": 84300 }, { "epoch": 0.3630752278097977, "grad_norm": 1.3120931386947632, "learning_rate": 0.00019870493323991998, "loss": 5.2391, "step": 84350 }, { "epoch": 0.3632904472690803, "grad_norm": 0.9528974890708923, "learning_rate": 0.00019870338235339726, "loss": 5.0484, "step": 84400 }, { "epoch": 0.36350566672836293, "grad_norm": 1.3743690252304077, "learning_rate": 0.00019870183054487001, "loss": 5.4868, "step": 84450 }, { "epoch": 0.36372088618764553, "grad_norm": 0.5074523091316223, "learning_rate": 0.0001987002778143528, "loss": 4.9961, "step": 84500 }, { "epoch": 0.3639361056469282, "grad_norm": 1.2393244504928589, "learning_rate": 0.00019869872416186, "loss": 5.5561, "step": 84550 }, { "epoch": 0.3641513251062108, "grad_norm": 0.6982135772705078, "learning_rate": 0.00019869716958740628, "loss": 5.6771, "step": 84600 }, { "epoch": 0.36436654456549344, "grad_norm": 2.371898651123047, "learning_rate": 0.00019869561409100608, "loss": 5.2362, "step": 84650 }, { "epoch": 0.36458176402477604, "grad_norm": 0.27971023321151733, "learning_rate": 0.0001986940576726739, "loss": 5.5869, "step": 84700 }, { "epoch": 0.3647969834840587, "grad_norm": 1.28279447555542, "learning_rate": 0.00019869250033242435, "loss": 5.8469, "step": 84750 }, { "epoch": 0.3650122029433413, "grad_norm": 0.7584916949272156, "learning_rate": 0.00019869094207027195, "loss": 5.1446, "step": 84800 }, { "epoch": 0.36522742240262396, "grad_norm": 1.0596297979354858, "learning_rate": 0.00019868938288623123, "loss": 5.1689, "step": 84850 }, { "epoch": 0.36544264186190656, "grad_norm": 1.6439939737319946, "learning_rate": 0.0001986878227803168, "loss": 5.3055, "step": 84900 }, { "epoch": 0.3656578613211892, "grad_norm": 0.6399978995323181, "learning_rate": 0.00019868626175254323, "loss": 5.2631, "step": 84950 }, { "epoch": 0.36587308078047187, "grad_norm": 1.3927440643310547, "learning_rate": 0.00019868469980292502, "loss": 5.4153, "step": 85000 }, { "epoch": 0.36608830023975447, "grad_norm": 2.4580836296081543, "learning_rate": 0.00019868313693147685, "loss": 5.1207, "step": 85050 }, { "epoch": 0.36630351969903713, "grad_norm": 1.5194587707519531, "learning_rate": 0.0001986815731382133, "loss": 5.714, "step": 85100 }, { "epoch": 0.36651873915831973, "grad_norm": 1.755123257637024, "learning_rate": 0.00019868000842314897, "loss": 5.3388, "step": 85150 }, { "epoch": 0.3667339586176024, "grad_norm": 2.799065351486206, "learning_rate": 0.00019867844278629847, "loss": 4.7973, "step": 85200 }, { "epoch": 0.366949178076885, "grad_norm": 1.2038650512695312, "learning_rate": 0.00019867687622767641, "loss": 5.347, "step": 85250 }, { "epoch": 0.36716439753616764, "grad_norm": 1.3695787191390991, "learning_rate": 0.00019867530874729747, "loss": 5.0932, "step": 85300 }, { "epoch": 0.36737961699545024, "grad_norm": 1.084961175918579, "learning_rate": 0.00019867374034517626, "loss": 5.5226, "step": 85350 }, { "epoch": 0.3675948364547329, "grad_norm": 1.9511195421218872, "learning_rate": 0.00019867217102132741, "loss": 5.2208, "step": 85400 }, { "epoch": 0.3678100559140155, "grad_norm": 1.1633294820785522, "learning_rate": 0.00019867060077576566, "loss": 5.6474, "step": 85450 }, { "epoch": 0.36802527537329816, "grad_norm": 2.406092643737793, "learning_rate": 0.0001986690296085056, "loss": 5.6959, "step": 85500 }, { "epoch": 0.36824049483258076, "grad_norm": 1.630139708518982, "learning_rate": 0.0001986674575195619, "loss": 5.3374, "step": 85550 }, { "epoch": 0.3684557142918634, "grad_norm": 1.0205661058425903, "learning_rate": 0.0001986658845089493, "loss": 5.4706, "step": 85600 }, { "epoch": 0.36867093375114607, "grad_norm": 1.1999229192733765, "learning_rate": 0.00019866431057668246, "loss": 5.0163, "step": 85650 }, { "epoch": 0.36888615321042867, "grad_norm": 1.4434999227523804, "learning_rate": 0.0001986627357227761, "loss": 5.54, "step": 85700 }, { "epoch": 0.3691013726697113, "grad_norm": 1.8499248027801514, "learning_rate": 0.00019866115994724494, "loss": 5.1054, "step": 85750 }, { "epoch": 0.3693165921289939, "grad_norm": 1.3774948120117188, "learning_rate": 0.00019865958325010366, "loss": 4.8827, "step": 85800 }, { "epoch": 0.3695318115882766, "grad_norm": 1.1609528064727783, "learning_rate": 0.00019865800563136703, "loss": 5.2023, "step": 85850 }, { "epoch": 0.3697470310475592, "grad_norm": 1.4930899143218994, "learning_rate": 0.00019865642709104972, "loss": 5.733, "step": 85900 }, { "epoch": 0.36996225050684184, "grad_norm": 1.0417994260787964, "learning_rate": 0.00019865484762916658, "loss": 5.1301, "step": 85950 }, { "epoch": 0.37017746996612444, "grad_norm": 0.38525938987731934, "learning_rate": 0.00019865326724573227, "loss": 5.4178, "step": 86000 }, { "epoch": 0.3703926894254071, "grad_norm": 1.2566380500793457, "learning_rate": 0.00019865168594076157, "loss": 5.0028, "step": 86050 }, { "epoch": 0.3706079088846897, "grad_norm": 1.8241444826126099, "learning_rate": 0.0001986501037142693, "loss": 5.2333, "step": 86100 }, { "epoch": 0.37082312834397235, "grad_norm": 1.8255606889724731, "learning_rate": 0.0001986485205662702, "loss": 5.3665, "step": 86150 }, { "epoch": 0.37103834780325495, "grad_norm": 0.8046995401382446, "learning_rate": 0.0001986469364967791, "loss": 5.448, "step": 86200 }, { "epoch": 0.3712535672625376, "grad_norm": 1.0548220872879028, "learning_rate": 0.00019864535150581074, "loss": 4.8449, "step": 86250 }, { "epoch": 0.37146878672182027, "grad_norm": 2.7417585849761963, "learning_rate": 0.00019864376559337995, "loss": 5.4419, "step": 86300 }, { "epoch": 0.37168400618110287, "grad_norm": 0.9974085092544556, "learning_rate": 0.00019864217875950155, "loss": 5.4818, "step": 86350 }, { "epoch": 0.3718992256403855, "grad_norm": 0.5262095332145691, "learning_rate": 0.00019864059100419037, "loss": 5.2257, "step": 86400 }, { "epoch": 0.3721144450996681, "grad_norm": 1.7141706943511963, "learning_rate": 0.00019863900232746122, "loss": 4.8888, "step": 86450 }, { "epoch": 0.3723296645589508, "grad_norm": 0.6931888461112976, "learning_rate": 0.00019863741272932895, "loss": 4.9885, "step": 86500 }, { "epoch": 0.3725448840182334, "grad_norm": 2.0557987689971924, "learning_rate": 0.0001986358222098084, "loss": 5.4093, "step": 86550 }, { "epoch": 0.37276010347751604, "grad_norm": 1.6285730600357056, "learning_rate": 0.00019863423076891442, "loss": 5.6818, "step": 86600 }, { "epoch": 0.37297532293679864, "grad_norm": 1.543258786201477, "learning_rate": 0.00019863263840666192, "loss": 5.5539, "step": 86650 }, { "epoch": 0.3731905423960813, "grad_norm": 1.466125249862671, "learning_rate": 0.00019863104512306574, "loss": 4.9495, "step": 86700 }, { "epoch": 0.3734057618553639, "grad_norm": 1.3224036693572998, "learning_rate": 0.00019862945091814077, "loss": 5.4994, "step": 86750 }, { "epoch": 0.37362098131464655, "grad_norm": 0.5728740096092224, "learning_rate": 0.00019862785579190188, "loss": 5.3691, "step": 86800 }, { "epoch": 0.37383620077392915, "grad_norm": 2.102623462677002, "learning_rate": 0.00019862625974436402, "loss": 4.9775, "step": 86850 }, { "epoch": 0.3740514202332118, "grad_norm": 0.5567491054534912, "learning_rate": 0.00019862466277554205, "loss": 5.5919, "step": 86900 }, { "epoch": 0.37426663969249446, "grad_norm": 1.1038466691970825, "learning_rate": 0.00019862306488545092, "loss": 5.3283, "step": 86950 }, { "epoch": 0.37448185915177706, "grad_norm": 1.3939763307571411, "learning_rate": 0.00019862146607410553, "loss": 5.344, "step": 87000 }, { "epoch": 0.37448185915177706, "eval_loss": 5.786531925201416, "eval_runtime": 49.7502, "eval_samples_per_second": 12.864, "eval_steps_per_second": 6.432, "eval_tts_loss": 6.215621896016398, "step": 87000 }, { "epoch": 0.3746970786110597, "grad_norm": 1.3509036302566528, "learning_rate": 0.00019861986634152085, "loss": 5.0609, "step": 87050 }, { "epoch": 0.3749122980703423, "grad_norm": 0.9164344668388367, "learning_rate": 0.0001986182656877118, "loss": 4.9144, "step": 87100 }, { "epoch": 0.375127517529625, "grad_norm": 1.201137661933899, "learning_rate": 0.00019861666411269332, "loss": 5.3846, "step": 87150 }, { "epoch": 0.3753427369889076, "grad_norm": 1.6279445886611938, "learning_rate": 0.0001986150616164804, "loss": 5.2109, "step": 87200 }, { "epoch": 0.37555795644819023, "grad_norm": 1.3204104900360107, "learning_rate": 0.00019861345819908798, "loss": 5.2844, "step": 87250 }, { "epoch": 0.37577317590747283, "grad_norm": 1.2050327062606812, "learning_rate": 0.00019861185386053104, "loss": 5.1339, "step": 87300 }, { "epoch": 0.3759883953667555, "grad_norm": 1.0506649017333984, "learning_rate": 0.00019861024860082463, "loss": 5.0036, "step": 87350 }, { "epoch": 0.3762036148260381, "grad_norm": 1.3199291229248047, "learning_rate": 0.00019860864241998366, "loss": 5.1437, "step": 87400 }, { "epoch": 0.37641883428532075, "grad_norm": 1.1201114654541016, "learning_rate": 0.00019860703531802317, "loss": 5.1614, "step": 87450 }, { "epoch": 0.37663405374460335, "grad_norm": 0.7332944273948669, "learning_rate": 0.00019860542729495815, "loss": 5.0, "step": 87500 }, { "epoch": 0.376849273203886, "grad_norm": 0.8768817186355591, "learning_rate": 0.00019860381835080365, "loss": 5.5138, "step": 87550 }, { "epoch": 0.37706449266316866, "grad_norm": 1.3065227270126343, "learning_rate": 0.0001986022084855747, "loss": 5.3533, "step": 87600 }, { "epoch": 0.37727971212245126, "grad_norm": 1.1327372789382935, "learning_rate": 0.0001986005976992863, "loss": 4.8194, "step": 87650 }, { "epoch": 0.3774949315817339, "grad_norm": 1.3736509084701538, "learning_rate": 0.00019859898599195356, "loss": 5.0752, "step": 87700 }, { "epoch": 0.3777101510410165, "grad_norm": 0.885672390460968, "learning_rate": 0.00019859737336359146, "loss": 5.4836, "step": 87750 }, { "epoch": 0.3779253705002992, "grad_norm": 0.9914229512214661, "learning_rate": 0.00019859575981421513, "loss": 5.0773, "step": 87800 }, { "epoch": 0.3781405899595818, "grad_norm": 1.4045931100845337, "learning_rate": 0.00019859414534383963, "loss": 5.1732, "step": 87850 }, { "epoch": 0.37835580941886443, "grad_norm": 1.6671243906021118, "learning_rate": 0.00019859252995248002, "loss": 5.7191, "step": 87900 }, { "epoch": 0.37857102887814703, "grad_norm": 1.5633172988891602, "learning_rate": 0.00019859091364015136, "loss": 5.5926, "step": 87950 }, { "epoch": 0.3787862483374297, "grad_norm": 1.577291488647461, "learning_rate": 0.0001985892964068688, "loss": 5.4711, "step": 88000 }, { "epoch": 0.3790014677967123, "grad_norm": 1.385677456855774, "learning_rate": 0.00019858767825264745, "loss": 4.9225, "step": 88050 }, { "epoch": 0.37921668725599494, "grad_norm": 0.12608899176120758, "learning_rate": 0.0001985860591775024, "loss": 5.1063, "step": 88100 }, { "epoch": 0.37943190671527754, "grad_norm": 0.8715876340866089, "learning_rate": 0.00019858443918144877, "loss": 5.5403, "step": 88150 }, { "epoch": 0.3796471261745602, "grad_norm": 1.1905145645141602, "learning_rate": 0.00019858281826450172, "loss": 5.1944, "step": 88200 }, { "epoch": 0.37986234563384286, "grad_norm": 1.3014360666275024, "learning_rate": 0.00019858119642667633, "loss": 5.4062, "step": 88250 }, { "epoch": 0.38007756509312546, "grad_norm": 1.9780232906341553, "learning_rate": 0.00019857957366798784, "loss": 5.3489, "step": 88300 }, { "epoch": 0.3802927845524081, "grad_norm": 1.465547800064087, "learning_rate": 0.00019857794998845134, "loss": 5.4978, "step": 88350 }, { "epoch": 0.3805080040116907, "grad_norm": 1.7695120573043823, "learning_rate": 0.00019857632538808204, "loss": 5.675, "step": 88400 }, { "epoch": 0.38072322347097337, "grad_norm": 1.1777067184448242, "learning_rate": 0.0001985746998668951, "loss": 5.51, "step": 88450 }, { "epoch": 0.38093844293025597, "grad_norm": 1.3394852876663208, "learning_rate": 0.00019857307342490568, "loss": 5.5087, "step": 88500 }, { "epoch": 0.38115366238953863, "grad_norm": 1.1689295768737793, "learning_rate": 0.00019857144606212897, "loss": 5.3344, "step": 88550 }, { "epoch": 0.38136888184882123, "grad_norm": 1.2763595581054688, "learning_rate": 0.00019856981777858023, "loss": 5.3674, "step": 88600 }, { "epoch": 0.3815841013081039, "grad_norm": 0.2509107291698456, "learning_rate": 0.00019856818857427464, "loss": 5.2817, "step": 88650 }, { "epoch": 0.3817993207673865, "grad_norm": 0.7023130655288696, "learning_rate": 0.00019856655844922738, "loss": 5.3943, "step": 88700 }, { "epoch": 0.38201454022666914, "grad_norm": 1.617323398590088, "learning_rate": 0.00019856492740345373, "loss": 5.1865, "step": 88750 }, { "epoch": 0.38222975968595174, "grad_norm": 2.008674383163452, "learning_rate": 0.0001985632954369689, "loss": 4.9676, "step": 88800 }, { "epoch": 0.3824449791452344, "grad_norm": 1.4900872707366943, "learning_rate": 0.00019856166254978813, "loss": 5.5932, "step": 88850 }, { "epoch": 0.382660198604517, "grad_norm": 2.1086366176605225, "learning_rate": 0.00019856002874192672, "loss": 5.3097, "step": 88900 }, { "epoch": 0.38287541806379966, "grad_norm": 1.4192103147506714, "learning_rate": 0.0001985583940133999, "loss": 4.9362, "step": 88950 }, { "epoch": 0.3830906375230823, "grad_norm": 1.3249175548553467, "learning_rate": 0.0001985567583642229, "loss": 5.1286, "step": 89000 }, { "epoch": 0.3833058569823649, "grad_norm": 2.6378602981567383, "learning_rate": 0.00019855512179441104, "loss": 5.6369, "step": 89050 }, { "epoch": 0.38352107644164757, "grad_norm": 0.30357474088668823, "learning_rate": 0.00019855348430397956, "loss": 4.9013, "step": 89100 }, { "epoch": 0.38373629590093017, "grad_norm": 2.528350353240967, "learning_rate": 0.00019855184589294387, "loss": 5.7461, "step": 89150 }, { "epoch": 0.3839515153602128, "grad_norm": 1.3142184019088745, "learning_rate": 0.00019855020656131918, "loss": 5.0624, "step": 89200 }, { "epoch": 0.3841667348194954, "grad_norm": 0.30306756496429443, "learning_rate": 0.0001985485663091208, "loss": 4.794, "step": 89250 }, { "epoch": 0.3843819542787781, "grad_norm": 1.8944971561431885, "learning_rate": 0.0001985469251363641, "loss": 5.6903, "step": 89300 }, { "epoch": 0.3845971737380607, "grad_norm": 1.5686907768249512, "learning_rate": 0.00019854528304306439, "loss": 5.4387, "step": 89350 }, { "epoch": 0.38481239319734334, "grad_norm": 2.231105327606201, "learning_rate": 0.000198543640029237, "loss": 5.4196, "step": 89400 }, { "epoch": 0.38502761265662594, "grad_norm": 1.5708229541778564, "learning_rate": 0.0001985419960948973, "loss": 4.7853, "step": 89450 }, { "epoch": 0.3852428321159086, "grad_norm": 1.8484828472137451, "learning_rate": 0.00019854035124006062, "loss": 5.3987, "step": 89500 }, { "epoch": 0.3854580515751912, "grad_norm": 0.6372663974761963, "learning_rate": 0.00019853870546474232, "loss": 5.1019, "step": 89550 }, { "epoch": 0.38567327103447385, "grad_norm": 1.118377685546875, "learning_rate": 0.0001985370587689578, "loss": 5.2804, "step": 89600 }, { "epoch": 0.3858884904937565, "grad_norm": 0.9574761986732483, "learning_rate": 0.00019853541115272242, "loss": 4.6118, "step": 89650 }, { "epoch": 0.3861037099530391, "grad_norm": 1.2139641046524048, "learning_rate": 0.00019853376261605158, "loss": 5.1853, "step": 89700 }, { "epoch": 0.38631892941232177, "grad_norm": 0.8361402153968811, "learning_rate": 0.0001985321131589607, "loss": 5.0532, "step": 89750 }, { "epoch": 0.38653414887160437, "grad_norm": 1.1094739437103271, "learning_rate": 0.00019853046278146518, "loss": 5.2794, "step": 89800 }, { "epoch": 0.386749368330887, "grad_norm": 1.9436157941818237, "learning_rate": 0.00019852881148358043, "loss": 5.7607, "step": 89850 }, { "epoch": 0.3869645877901696, "grad_norm": 2.2569940090179443, "learning_rate": 0.00019852715926532184, "loss": 5.5093, "step": 89900 }, { "epoch": 0.3871798072494523, "grad_norm": 1.5790894031524658, "learning_rate": 0.00019852550612670488, "loss": 5.3146, "step": 89950 }, { "epoch": 0.3873950267087349, "grad_norm": 0.5068002343177795, "learning_rate": 0.00019852385206774498, "loss": 5.2954, "step": 90000 }, { "epoch": 0.3873950267087349, "eval_loss": 5.76200008392334, "eval_runtime": 49.9545, "eval_samples_per_second": 12.812, "eval_steps_per_second": 6.406, "eval_tts_loss": 6.257961057228831, "step": 90000 }, { "epoch": 0.38761024616801754, "grad_norm": 0.8411883115768433, "learning_rate": 0.00019852219708845764, "loss": 5.3367, "step": 90050 }, { "epoch": 0.38782546562730014, "grad_norm": 1.7344359159469604, "learning_rate": 0.00019852054118885823, "loss": 5.4851, "step": 90100 }, { "epoch": 0.3880406850865828, "grad_norm": 1.0844186544418335, "learning_rate": 0.0001985188843689623, "loss": 5.2311, "step": 90150 }, { "epoch": 0.3882559045458654, "grad_norm": 0.9388478398323059, "learning_rate": 0.00019851722662878526, "loss": 5.2811, "step": 90200 }, { "epoch": 0.38847112400514805, "grad_norm": 0.31950509548187256, "learning_rate": 0.00019851556796834262, "loss": 4.9145, "step": 90250 }, { "epoch": 0.3886863434644307, "grad_norm": 1.7691386938095093, "learning_rate": 0.00019851390838764992, "loss": 5.4081, "step": 90300 }, { "epoch": 0.3889015629237133, "grad_norm": 1.0105483531951904, "learning_rate": 0.0001985122478867226, "loss": 4.9604, "step": 90350 }, { "epoch": 0.38911678238299596, "grad_norm": 1.0391011238098145, "learning_rate": 0.0001985105864655762, "loss": 5.306, "step": 90400 }, { "epoch": 0.38933200184227856, "grad_norm": 1.247122883796692, "learning_rate": 0.00019850892412422622, "loss": 5.093, "step": 90450 }, { "epoch": 0.3895472213015612, "grad_norm": 0.9270398616790771, "learning_rate": 0.0001985072608626882, "loss": 4.8877, "step": 90500 }, { "epoch": 0.3897624407608438, "grad_norm": 1.1012998819351196, "learning_rate": 0.0001985055966809777, "loss": 4.8702, "step": 90550 }, { "epoch": 0.3899776602201265, "grad_norm": 1.0820469856262207, "learning_rate": 0.00019850393157911023, "loss": 5.8878, "step": 90600 }, { "epoch": 0.3901928796794091, "grad_norm": 1.4381848573684692, "learning_rate": 0.00019850226555710138, "loss": 5.1598, "step": 90650 }, { "epoch": 0.39040809913869173, "grad_norm": 0.46012648940086365, "learning_rate": 0.00019850059861496665, "loss": 5.3686, "step": 90700 }, { "epoch": 0.39062331859797433, "grad_norm": 0.7692629098892212, "learning_rate": 0.0001984989307527217, "loss": 5.1854, "step": 90750 }, { "epoch": 0.390838538057257, "grad_norm": 1.0443968772888184, "learning_rate": 0.00019849726197038203, "loss": 5.389, "step": 90800 }, { "epoch": 0.3910537575165396, "grad_norm": 0.7662686109542847, "learning_rate": 0.00019849559226796327, "loss": 5.3429, "step": 90850 }, { "epoch": 0.39126897697582225, "grad_norm": 0.9735128879547119, "learning_rate": 0.000198493921645481, "loss": 5.6797, "step": 90900 }, { "epoch": 0.3914841964351049, "grad_norm": 1.0142021179199219, "learning_rate": 0.00019849225010295084, "loss": 4.5849, "step": 90950 }, { "epoch": 0.3916994158943875, "grad_norm": 2.39231538772583, "learning_rate": 0.00019849057764038838, "loss": 4.7412, "step": 91000 }, { "epoch": 0.39191463535367016, "grad_norm": 1.378099799156189, "learning_rate": 0.00019848890425780926, "loss": 5.1388, "step": 91050 }, { "epoch": 0.39212985481295276, "grad_norm": 1.6604862213134766, "learning_rate": 0.00019848722995522913, "loss": 5.5353, "step": 91100 }, { "epoch": 0.3923450742722354, "grad_norm": 1.0201205015182495, "learning_rate": 0.0001984855547326636, "loss": 5.0028, "step": 91150 }, { "epoch": 0.392560293731518, "grad_norm": 1.3570383787155151, "learning_rate": 0.0001984838785901283, "loss": 5.0597, "step": 91200 }, { "epoch": 0.3927755131908007, "grad_norm": 0.7101498246192932, "learning_rate": 0.00019848220152763897, "loss": 5.5275, "step": 91250 }, { "epoch": 0.3929907326500833, "grad_norm": 1.7676786184310913, "learning_rate": 0.00019848052354521118, "loss": 5.3742, "step": 91300 }, { "epoch": 0.39320595210936593, "grad_norm": 1.0614954233169556, "learning_rate": 0.00019847884464286065, "loss": 5.182, "step": 91350 }, { "epoch": 0.39342117156864853, "grad_norm": 1.9872487783432007, "learning_rate": 0.0001984771648206031, "loss": 5.4288, "step": 91400 }, { "epoch": 0.3936363910279312, "grad_norm": 1.7502387762069702, "learning_rate": 0.00019847548407845415, "loss": 5.4804, "step": 91450 }, { "epoch": 0.3938516104872138, "grad_norm": 1.0443567037582397, "learning_rate": 0.00019847380241642953, "loss": 5.5361, "step": 91500 }, { "epoch": 0.39406682994649644, "grad_norm": 1.6611417531967163, "learning_rate": 0.00019847211983454495, "loss": 5.4697, "step": 91550 }, { "epoch": 0.3942820494057791, "grad_norm": 0.9315400719642639, "learning_rate": 0.0001984704363328161, "loss": 4.7591, "step": 91600 }, { "epoch": 0.3944972688650617, "grad_norm": 1.2008951902389526, "learning_rate": 0.0001984687519112588, "loss": 5.7361, "step": 91650 }, { "epoch": 0.39471248832434436, "grad_norm": 1.1882729530334473, "learning_rate": 0.00019846706656988868, "loss": 5.2096, "step": 91700 }, { "epoch": 0.39492770778362696, "grad_norm": 0.5675370693206787, "learning_rate": 0.00019846538030872152, "loss": 5.2686, "step": 91750 }, { "epoch": 0.3951429272429096, "grad_norm": 1.339798092842102, "learning_rate": 0.00019846369312777307, "loss": 5.497, "step": 91800 }, { "epoch": 0.3953581467021922, "grad_norm": 1.4954688549041748, "learning_rate": 0.0001984620050270591, "loss": 5.6591, "step": 91850 }, { "epoch": 0.39557336616147487, "grad_norm": 1.3909872770309448, "learning_rate": 0.00019846031600659534, "loss": 5.5412, "step": 91900 }, { "epoch": 0.39578858562075747, "grad_norm": 1.7089918851852417, "learning_rate": 0.00019845862606639763, "loss": 5.4463, "step": 91950 }, { "epoch": 0.3960038050800401, "grad_norm": 0.9484457969665527, "learning_rate": 0.0001984569352064817, "loss": 5.396, "step": 92000 }, { "epoch": 0.39621902453932273, "grad_norm": 1.8879566192626953, "learning_rate": 0.00019845524342686338, "loss": 5.6921, "step": 92050 }, { "epoch": 0.3964342439986054, "grad_norm": 0.8068963885307312, "learning_rate": 0.0001984535507275585, "loss": 5.2532, "step": 92100 }, { "epoch": 0.396649463457888, "grad_norm": 0.6368489861488342, "learning_rate": 0.00019845185710858277, "loss": 4.6286, "step": 92150 }, { "epoch": 0.39686468291717064, "grad_norm": 0.9673624038696289, "learning_rate": 0.0001984501625699521, "loss": 5.2076, "step": 92200 }, { "epoch": 0.39707990237645324, "grad_norm": 1.3644472360610962, "learning_rate": 0.00019844846711168234, "loss": 5.4988, "step": 92250 }, { "epoch": 0.3972951218357359, "grad_norm": 0.5469921827316284, "learning_rate": 0.0001984467707337892, "loss": 5.5781, "step": 92300 }, { "epoch": 0.39751034129501855, "grad_norm": 0.5544052720069885, "learning_rate": 0.00019844507343628865, "loss": 4.7167, "step": 92350 }, { "epoch": 0.39772556075430116, "grad_norm": 1.490154504776001, "learning_rate": 0.0001984433752191965, "loss": 5.679, "step": 92400 }, { "epoch": 0.3979407802135838, "grad_norm": 1.0462696552276611, "learning_rate": 0.0001984416760825286, "loss": 5.0232, "step": 92450 }, { "epoch": 0.3981559996728664, "grad_norm": 1.511855125427246, "learning_rate": 0.00019843997602630084, "loss": 5.0028, "step": 92500 }, { "epoch": 0.39837121913214907, "grad_norm": 1.7351406812667847, "learning_rate": 0.00019843827505052912, "loss": 5.7227, "step": 92550 }, { "epoch": 0.39858643859143167, "grad_norm": 1.4883391857147217, "learning_rate": 0.0001984365731552293, "loss": 5.5027, "step": 92600 }, { "epoch": 0.3988016580507143, "grad_norm": 1.0307890176773071, "learning_rate": 0.00019843487034041727, "loss": 5.3153, "step": 92650 }, { "epoch": 0.3990168775099969, "grad_norm": 0.5970161557197571, "learning_rate": 0.00019843316660610896, "loss": 5.2929, "step": 92700 }, { "epoch": 0.3992320969692796, "grad_norm": 1.6669255495071411, "learning_rate": 0.00019843146195232025, "loss": 5.2752, "step": 92750 }, { "epoch": 0.3994473164285622, "grad_norm": 1.9138835668563843, "learning_rate": 0.00019842975637906712, "loss": 5.4733, "step": 92800 }, { "epoch": 0.39966253588784484, "grad_norm": 1.2873516082763672, "learning_rate": 0.00019842804988636544, "loss": 5.4485, "step": 92850 }, { "epoch": 0.39987775534712744, "grad_norm": 0.630450963973999, "learning_rate": 0.00019842634247423119, "loss": 5.3367, "step": 92900 }, { "epoch": 0.4000929748064101, "grad_norm": 1.817284345626831, "learning_rate": 0.00019842463414268037, "loss": 4.8944, "step": 92950 }, { "epoch": 0.40030819426569275, "grad_norm": 1.9982460737228394, "learning_rate": 0.0001984229248917288, "loss": 4.985, "step": 93000 }, { "epoch": 0.40030819426569275, "eval_loss": 5.757879734039307, "eval_runtime": 49.8997, "eval_samples_per_second": 12.826, "eval_steps_per_second": 6.413, "eval_tts_loss": 6.247412407351318, "step": 93000 }, { "epoch": 0.40052341372497535, "grad_norm": 2.709087610244751, "learning_rate": 0.00019842121472139255, "loss": 5.3518, "step": 93050 }, { "epoch": 0.400738633184258, "grad_norm": 1.2494267225265503, "learning_rate": 0.00019841950363168757, "loss": 5.3404, "step": 93100 }, { "epoch": 0.4009538526435406, "grad_norm": 1.251767635345459, "learning_rate": 0.00019841779162262983, "loss": 5.3215, "step": 93150 }, { "epoch": 0.40116907210282327, "grad_norm": 0.9008392691612244, "learning_rate": 0.00019841607869423538, "loss": 5.1668, "step": 93200 }, { "epoch": 0.40138429156210587, "grad_norm": 1.280427098274231, "learning_rate": 0.00019841436484652013, "loss": 5.1142, "step": 93250 }, { "epoch": 0.4015995110213885, "grad_norm": 1.226851224899292, "learning_rate": 0.00019841265007950016, "loss": 4.9875, "step": 93300 }, { "epoch": 0.4018147304806711, "grad_norm": 1.703042984008789, "learning_rate": 0.00019841093439319145, "loss": 4.9458, "step": 93350 }, { "epoch": 0.4020299499399538, "grad_norm": 0.5830041170120239, "learning_rate": 0.00019840921778761005, "loss": 5.4162, "step": 93400 }, { "epoch": 0.4022451693992364, "grad_norm": 1.7768830060958862, "learning_rate": 0.00019840750026277196, "loss": 5.3202, "step": 93450 }, { "epoch": 0.40246038885851904, "grad_norm": 0.9061640501022339, "learning_rate": 0.0001984057818186933, "loss": 4.5312, "step": 93500 }, { "epoch": 0.40267560831780164, "grad_norm": 0.6009230017662048, "learning_rate": 0.00019840406245539001, "loss": 5.2821, "step": 93550 }, { "epoch": 0.4028908277770843, "grad_norm": 0.8450963497161865, "learning_rate": 0.00019840234217287828, "loss": 5.1254, "step": 93600 }, { "epoch": 0.40310604723636695, "grad_norm": 1.0173791646957397, "learning_rate": 0.00019840062097117406, "loss": 4.9021, "step": 93650 }, { "epoch": 0.40332126669564955, "grad_norm": 1.6597051620483398, "learning_rate": 0.00019839889885029348, "loss": 5.174, "step": 93700 }, { "epoch": 0.4035364861549322, "grad_norm": 1.3861052989959717, "learning_rate": 0.00019839717581025265, "loss": 5.2947, "step": 93750 }, { "epoch": 0.4037517056142148, "grad_norm": 1.9175605773925781, "learning_rate": 0.0001983954518510676, "loss": 5.4682, "step": 93800 }, { "epoch": 0.40396692507349746, "grad_norm": 1.630787968635559, "learning_rate": 0.00019839372697275453, "loss": 5.1559, "step": 93850 }, { "epoch": 0.40418214453278006, "grad_norm": 1.8695857524871826, "learning_rate": 0.00019839200117532948, "loss": 5.345, "step": 93900 }, { "epoch": 0.4043973639920627, "grad_norm": 1.2207072973251343, "learning_rate": 0.00019839027445880855, "loss": 5.4687, "step": 93950 }, { "epoch": 0.4046125834513453, "grad_norm": 1.3172318935394287, "learning_rate": 0.00019838854682320794, "loss": 4.9673, "step": 94000 }, { "epoch": 0.404827802910628, "grad_norm": 1.14497709274292, "learning_rate": 0.00019838681826854376, "loss": 4.7402, "step": 94050 }, { "epoch": 0.4050430223699106, "grad_norm": 0.6153233051300049, "learning_rate": 0.00019838508879483214, "loss": 4.9634, "step": 94100 }, { "epoch": 0.40525824182919323, "grad_norm": 2.505521535873413, "learning_rate": 0.00019838335840208926, "loss": 5.2494, "step": 94150 }, { "epoch": 0.40547346128847583, "grad_norm": 1.539318323135376, "learning_rate": 0.0001983816270903313, "loss": 4.7165, "step": 94200 }, { "epoch": 0.4056886807477585, "grad_norm": 2.1183629035949707, "learning_rate": 0.00019837989485957435, "loss": 5.1789, "step": 94250 }, { "epoch": 0.40590390020704115, "grad_norm": 0.9830565452575684, "learning_rate": 0.00019837816170983463, "loss": 5.4011, "step": 94300 }, { "epoch": 0.40611911966632375, "grad_norm": 1.3257405757904053, "learning_rate": 0.0001983764276411284, "loss": 5.6187, "step": 94350 }, { "epoch": 0.4063343391256064, "grad_norm": 1.3475052118301392, "learning_rate": 0.00019837469265347176, "loss": 4.8378, "step": 94400 }, { "epoch": 0.406549558584889, "grad_norm": 1.608943223953247, "learning_rate": 0.000198372956746881, "loss": 5.1095, "step": 94450 }, { "epoch": 0.40676477804417166, "grad_norm": 1.3941752910614014, "learning_rate": 0.00019837121992137227, "loss": 4.9316, "step": 94500 }, { "epoch": 0.40697999750345426, "grad_norm": 0.7446410655975342, "learning_rate": 0.00019836948217696182, "loss": 5.0324, "step": 94550 }, { "epoch": 0.4071952169627369, "grad_norm": 1.8726286888122559, "learning_rate": 0.00019836774351366588, "loss": 5.673, "step": 94600 }, { "epoch": 0.4074104364220195, "grad_norm": 1.6451877355575562, "learning_rate": 0.00019836600393150075, "loss": 5.059, "step": 94650 }, { "epoch": 0.4076256558813022, "grad_norm": 1.0001126527786255, "learning_rate": 0.00019836426343048253, "loss": 4.9492, "step": 94700 }, { "epoch": 0.4078408753405848, "grad_norm": 0.43162086606025696, "learning_rate": 0.00019836252201062764, "loss": 5.8655, "step": 94750 }, { "epoch": 0.40805609479986743, "grad_norm": 1.131543517112732, "learning_rate": 0.00019836077967195224, "loss": 4.9929, "step": 94800 }, { "epoch": 0.40827131425915003, "grad_norm": 1.0589274168014526, "learning_rate": 0.00019835903641447265, "loss": 5.142, "step": 94850 }, { "epoch": 0.4084865337184327, "grad_norm": 0.6269965767860413, "learning_rate": 0.00019835729223820515, "loss": 5.1349, "step": 94900 }, { "epoch": 0.40870175317771534, "grad_norm": 1.6212940216064453, "learning_rate": 0.00019835554714316605, "loss": 5.7592, "step": 94950 }, { "epoch": 0.40891697263699794, "grad_norm": 1.453898549079895, "learning_rate": 0.0001983538011293716, "loss": 4.9383, "step": 95000 }, { "epoch": 0.4091321920962806, "grad_norm": 1.0058389902114868, "learning_rate": 0.00019835205419683815, "loss": 5.7329, "step": 95050 }, { "epoch": 0.4093474115555632, "grad_norm": 2.2823047637939453, "learning_rate": 0.00019835030634558204, "loss": 5.6481, "step": 95100 }, { "epoch": 0.40956263101484586, "grad_norm": 0.8001925945281982, "learning_rate": 0.0001983485575756195, "loss": 4.9838, "step": 95150 }, { "epoch": 0.40977785047412846, "grad_norm": 2.2790889739990234, "learning_rate": 0.000198346807886967, "loss": 5.2238, "step": 95200 }, { "epoch": 0.4099930699334111, "grad_norm": 1.4880340099334717, "learning_rate": 0.0001983450572796408, "loss": 5.3247, "step": 95250 }, { "epoch": 0.4102082893926937, "grad_norm": 2.2464723587036133, "learning_rate": 0.00019834330575365725, "loss": 5.2516, "step": 95300 }, { "epoch": 0.41042350885197637, "grad_norm": 0.9757848381996155, "learning_rate": 0.00019834155330903275, "loss": 5.4749, "step": 95350 }, { "epoch": 0.41063872831125897, "grad_norm": 1.3296558856964111, "learning_rate": 0.00019833979994578362, "loss": 4.9809, "step": 95400 }, { "epoch": 0.4108539477705416, "grad_norm": 1.256693720817566, "learning_rate": 0.00019833804566392628, "loss": 5.5948, "step": 95450 }, { "epoch": 0.41106916722982423, "grad_norm": 0.37737587094306946, "learning_rate": 0.00019833629046347713, "loss": 5.0967, "step": 95500 }, { "epoch": 0.4112843866891069, "grad_norm": 1.051485300064087, "learning_rate": 0.00019833453434445252, "loss": 5.1936, "step": 95550 }, { "epoch": 0.4114996061483895, "grad_norm": 1.0475132465362549, "learning_rate": 0.00019833277730686887, "loss": 5.2217, "step": 95600 }, { "epoch": 0.41171482560767214, "grad_norm": 0.18857957422733307, "learning_rate": 0.0001983310193507426, "loss": 5.026, "step": 95650 }, { "epoch": 0.4119300450669548, "grad_norm": 1.0519349575042725, "learning_rate": 0.00019832926047609015, "loss": 5.6942, "step": 95700 }, { "epoch": 0.4121452645262374, "grad_norm": 0.8312272429466248, "learning_rate": 0.00019832750068292788, "loss": 5.2717, "step": 95750 }, { "epoch": 0.41236048398552005, "grad_norm": 1.7180200815200806, "learning_rate": 0.00019832573997127236, "loss": 5.3271, "step": 95800 }, { "epoch": 0.41257570344480266, "grad_norm": 0.5454203486442566, "learning_rate": 0.00019832397834113988, "loss": 5.2302, "step": 95850 }, { "epoch": 0.4127909229040853, "grad_norm": 1.2990691661834717, "learning_rate": 0.00019832221579254702, "loss": 5.2785, "step": 95900 }, { "epoch": 0.4130061423633679, "grad_norm": 1.3484632968902588, "learning_rate": 0.00019832045232551015, "loss": 5.2882, "step": 95950 }, { "epoch": 0.41322136182265057, "grad_norm": 2.1265485286712646, "learning_rate": 0.0001983186879400458, "loss": 5.3539, "step": 96000 }, { "epoch": 0.41322136182265057, "eval_loss": 5.756744384765625, "eval_runtime": 49.7907, "eval_samples_per_second": 12.854, "eval_steps_per_second": 6.427, "eval_tts_loss": 6.238794485812143, "step": 96000 }, { "epoch": 0.41343658128193317, "grad_norm": 1.3300528526306152, "learning_rate": 0.00019831692263617048, "loss": 5.0169, "step": 96050 }, { "epoch": 0.4136518007412158, "grad_norm": 1.795886754989624, "learning_rate": 0.0001983151564139006, "loss": 5.5792, "step": 96100 }, { "epoch": 0.4138670202004984, "grad_norm": 1.4187487363815308, "learning_rate": 0.00019831338927325272, "loss": 5.0908, "step": 96150 }, { "epoch": 0.4140822396597811, "grad_norm": 2.9060065746307373, "learning_rate": 0.0001983116212142433, "loss": 5.4422, "step": 96200 }, { "epoch": 0.4142974591190637, "grad_norm": 1.4156203269958496, "learning_rate": 0.00019830985223688886, "loss": 5.6101, "step": 96250 }, { "epoch": 0.41451267857834634, "grad_norm": 1.1357344388961792, "learning_rate": 0.00019830808234120602, "loss": 5.0402, "step": 96300 }, { "epoch": 0.414727898037629, "grad_norm": 0.6349567770957947, "learning_rate": 0.00019830631152721118, "loss": 5.0623, "step": 96350 }, { "epoch": 0.4149431174969116, "grad_norm": 0.23770560324192047, "learning_rate": 0.00019830453979492095, "loss": 5.514, "step": 96400 }, { "epoch": 0.41515833695619425, "grad_norm": 1.4053065776824951, "learning_rate": 0.00019830276714435187, "loss": 5.4123, "step": 96450 }, { "epoch": 0.41537355641547685, "grad_norm": 1.155515193939209, "learning_rate": 0.0001983009935755205, "loss": 4.9672, "step": 96500 }, { "epoch": 0.4155887758747595, "grad_norm": 0.5635316371917725, "learning_rate": 0.0001982992190884434, "loss": 4.8091, "step": 96550 }, { "epoch": 0.4158039953340421, "grad_norm": 1.134390950202942, "learning_rate": 0.00019829744368313716, "loss": 5.4412, "step": 96600 }, { "epoch": 0.41601921479332477, "grad_norm": 1.4817259311676025, "learning_rate": 0.00019829566735961836, "loss": 4.8746, "step": 96650 }, { "epoch": 0.41623443425260737, "grad_norm": 1.0903801918029785, "learning_rate": 0.00019829389011790359, "loss": 4.9235, "step": 96700 }, { "epoch": 0.41644965371189, "grad_norm": 1.2789667844772339, "learning_rate": 0.00019829211195800944, "loss": 5.0041, "step": 96750 }, { "epoch": 0.4166648731711726, "grad_norm": 1.2438594102859497, "learning_rate": 0.0001982903328799525, "loss": 5.46, "step": 96800 }, { "epoch": 0.4168800926304553, "grad_norm": 0.32034504413604736, "learning_rate": 0.00019828855288374947, "loss": 5.0711, "step": 96850 }, { "epoch": 0.4170953120897379, "grad_norm": 0.647940993309021, "learning_rate": 0.0001982867719694169, "loss": 5.0782, "step": 96900 }, { "epoch": 0.41731053154902054, "grad_norm": 1.758195400238037, "learning_rate": 0.00019828499013697145, "loss": 5.0605, "step": 96950 }, { "epoch": 0.4175257510083032, "grad_norm": 1.2064770460128784, "learning_rate": 0.00019828320738642975, "loss": 4.9637, "step": 97000 }, { "epoch": 0.4177409704675858, "grad_norm": 1.6847350597381592, "learning_rate": 0.0001982814237178085, "loss": 5.2653, "step": 97050 }, { "epoch": 0.41795618992686845, "grad_norm": 1.478193759918213, "learning_rate": 0.00019827963913112432, "loss": 4.8579, "step": 97100 }, { "epoch": 0.41817140938615105, "grad_norm": 1.945642113685608, "learning_rate": 0.0001982778536263939, "loss": 5.4091, "step": 97150 }, { "epoch": 0.4183866288454337, "grad_norm": 2.5969762802124023, "learning_rate": 0.0001982760672036339, "loss": 5.5433, "step": 97200 }, { "epoch": 0.4186018483047163, "grad_norm": 0.9743031859397888, "learning_rate": 0.00019827427986286101, "loss": 5.4795, "step": 97250 }, { "epoch": 0.41881706776399896, "grad_norm": 3.587322235107422, "learning_rate": 0.00019827249160409195, "loss": 5.4387, "step": 97300 }, { "epoch": 0.41903228722328156, "grad_norm": 2.7431259155273438, "learning_rate": 0.0001982707024273434, "loss": 5.0007, "step": 97350 }, { "epoch": 0.4192475066825642, "grad_norm": 0.6481873989105225, "learning_rate": 0.00019826891233263206, "loss": 5.4692, "step": 97400 }, { "epoch": 0.4194627261418468, "grad_norm": 0.5444102883338928, "learning_rate": 0.00019826712131997467, "loss": 4.9071, "step": 97450 }, { "epoch": 0.4196779456011295, "grad_norm": 0.2223920375108719, "learning_rate": 0.000198265329389388, "loss": 5.4011, "step": 97500 }, { "epoch": 0.4198931650604121, "grad_norm": 2.17893648147583, "learning_rate": 0.0001982635365408887, "loss": 4.9828, "step": 97550 }, { "epoch": 0.42010838451969473, "grad_norm": 1.0711735486984253, "learning_rate": 0.0001982617427744936, "loss": 5.6546, "step": 97600 }, { "epoch": 0.4203236039789774, "grad_norm": 1.3348705768585205, "learning_rate": 0.00019825994809021942, "loss": 5.4035, "step": 97650 }, { "epoch": 0.42053882343826, "grad_norm": 1.9088832139968872, "learning_rate": 0.00019825815248808294, "loss": 5.3402, "step": 97700 }, { "epoch": 0.42075404289754265, "grad_norm": 0.35627007484436035, "learning_rate": 0.0001982563559681009, "loss": 5.4091, "step": 97750 }, { "epoch": 0.42096926235682525, "grad_norm": 1.6408166885375977, "learning_rate": 0.0001982545585302901, "loss": 4.7911, "step": 97800 }, { "epoch": 0.4211844818161079, "grad_norm": 0.9826110005378723, "learning_rate": 0.00019825276017466735, "loss": 5.8576, "step": 97850 }, { "epoch": 0.4213997012753905, "grad_norm": 1.4764680862426758, "learning_rate": 0.0001982509609012494, "loss": 5.5049, "step": 97900 }, { "epoch": 0.42161492073467316, "grad_norm": 1.5818370580673218, "learning_rate": 0.00019824916071005313, "loss": 5.0049, "step": 97950 }, { "epoch": 0.42183014019395576, "grad_norm": 1.2040057182312012, "learning_rate": 0.0001982473596010953, "loss": 5.0086, "step": 98000 }, { "epoch": 0.4220453596532384, "grad_norm": 0.8830599188804626, "learning_rate": 0.00019824555757439274, "loss": 5.0358, "step": 98050 }, { "epoch": 0.422260579112521, "grad_norm": 1.0727115869522095, "learning_rate": 0.0001982437546299623, "loss": 5.2141, "step": 98100 }, { "epoch": 0.4224757985718037, "grad_norm": 0.6754670143127441, "learning_rate": 0.00019824195076782083, "loss": 5.5039, "step": 98150 }, { "epoch": 0.4226910180310863, "grad_norm": 1.0924664735794067, "learning_rate": 0.0001982401459879851, "loss": 5.433, "step": 98200 }, { "epoch": 0.42290623749036893, "grad_norm": 1.0812842845916748, "learning_rate": 0.0001982383402904721, "loss": 5.2831, "step": 98250 }, { "epoch": 0.4231214569496516, "grad_norm": 1.1327563524246216, "learning_rate": 0.0001982365336752986, "loss": 5.7427, "step": 98300 }, { "epoch": 0.4233366764089342, "grad_norm": 1.2452412843704224, "learning_rate": 0.0001982347261424815, "loss": 5.0053, "step": 98350 }, { "epoch": 0.42355189586821684, "grad_norm": 1.8447788953781128, "learning_rate": 0.0001982329176920377, "loss": 5.2795, "step": 98400 }, { "epoch": 0.42376711532749944, "grad_norm": 2.4793753623962402, "learning_rate": 0.0001982311083239841, "loss": 4.8085, "step": 98450 }, { "epoch": 0.4239823347867821, "grad_norm": 0.5339040756225586, "learning_rate": 0.00019822929803833757, "loss": 5.6025, "step": 98500 }, { "epoch": 0.4241975542460647, "grad_norm": 1.9252963066101074, "learning_rate": 0.000198227486835115, "loss": 5.3492, "step": 98550 }, { "epoch": 0.42441277370534736, "grad_norm": 0.2717931270599365, "learning_rate": 0.0001982256747143334, "loss": 5.0991, "step": 98600 }, { "epoch": 0.42462799316462996, "grad_norm": 1.3184789419174194, "learning_rate": 0.00019822386167600963, "loss": 5.7276, "step": 98650 }, { "epoch": 0.4248432126239126, "grad_norm": 0.9066381454467773, "learning_rate": 0.00019822204772016063, "loss": 5.0241, "step": 98700 }, { "epoch": 0.4250584320831952, "grad_norm": 2.8512184619903564, "learning_rate": 0.00019822023284680334, "loss": 5.1996, "step": 98750 }, { "epoch": 0.42527365154247787, "grad_norm": 0.36430537700653076, "learning_rate": 0.00019821841705595472, "loss": 4.9003, "step": 98800 }, { "epoch": 0.42548887100176047, "grad_norm": 2.1273550987243652, "learning_rate": 0.00019821660034763174, "loss": 5.4478, "step": 98850 }, { "epoch": 0.4257040904610431, "grad_norm": 0.8996185064315796, "learning_rate": 0.00019821478272185136, "loss": 5.7045, "step": 98900 }, { "epoch": 0.4259193099203258, "grad_norm": 1.8712717294692993, "learning_rate": 0.0001982129641786306, "loss": 5.0686, "step": 98950 }, { "epoch": 0.4261345293796084, "grad_norm": 2.27986741065979, "learning_rate": 0.00019821114471798635, "loss": 4.4387, "step": 99000 }, { "epoch": 0.4261345293796084, "eval_loss": 5.743788719177246, "eval_runtime": 49.7782, "eval_samples_per_second": 12.857, "eval_steps_per_second": 6.429, "eval_tts_loss": 6.302523045175333, "step": 99000 } ], "logging_steps": 50, "max_steps": 1626247, "num_input_tokens_seen": 0, "num_train_epochs": 7, "save_steps": 3000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.0280802130526208e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }